diff --git a/docs/dyn/bigquery_v2.datasets.html b/docs/dyn/bigquery_v2.datasets.html index 1a240d3c948..1a8fa857a10 100644 --- a/docs/dyn/bigquery_v2.datasets.html +++ b/docs/dyn/bigquery_v2.datasets.html @@ -181,6 +181,7 @@

Method Details

"maxTimeTravelHours": "A String", # [Optional] Number of hours for the max time travel for all tables in the dataset. "satisfiesPzs": True or False, # [Output-only] Reserved for future use. "selfLink": "A String", # [Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource. + "storageBillingModel": "A String", # [Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL. "tags": [ # [Optional]The tags associated with this dataset. Tag keys are globally unique. { "tagKey": "A String", # [Required] The namespaced friendly name of the tag key, e.g. "12345/environment" where 12345 is org id. @@ -254,6 +255,7 @@

Method Details

"maxTimeTravelHours": "A String", # [Optional] Number of hours for the max time travel for all tables in the dataset. "satisfiesPzs": True or False, # [Output-only] Reserved for future use. "selfLink": "A String", # [Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource. + "storageBillingModel": "A String", # [Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL. "tags": [ # [Optional]The tags associated with this dataset. Tag keys are globally unique. { "tagKey": "A String", # [Required] The namespaced friendly name of the tag key, e.g. "12345/environment" where 12345 is org id. @@ -321,6 +323,7 @@

Method Details

"maxTimeTravelHours": "A String", # [Optional] Number of hours for the max time travel for all tables in the dataset. "satisfiesPzs": True or False, # [Output-only] Reserved for future use. "selfLink": "A String", # [Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource. + "storageBillingModel": "A String", # [Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL. "tags": [ # [Optional]The tags associated with this dataset. Tag keys are globally unique. { "tagKey": "A String", # [Required] The namespaced friendly name of the tag key, e.g. "12345/environment" where 12345 is org id. @@ -445,6 +448,7 @@

Method Details

"maxTimeTravelHours": "A String", # [Optional] Number of hours for the max time travel for all tables in the dataset. "satisfiesPzs": True or False, # [Output-only] Reserved for future use. "selfLink": "A String", # [Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource. + "storageBillingModel": "A String", # [Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL. "tags": [ # [Optional]The tags associated with this dataset. Tag keys are globally unique. { "tagKey": "A String", # [Required] The namespaced friendly name of the tag key, e.g. "12345/environment" where 12345 is org id. @@ -512,6 +516,7 @@

Method Details

"maxTimeTravelHours": "A String", # [Optional] Number of hours for the max time travel for all tables in the dataset. "satisfiesPzs": True or False, # [Output-only] Reserved for future use. "selfLink": "A String", # [Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource. + "storageBillingModel": "A String", # [Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL. "tags": [ # [Optional]The tags associated with this dataset. Tag keys are globally unique. { "tagKey": "A String", # [Required] The namespaced friendly name of the tag key, e.g. "12345/environment" where 12345 is org id. @@ -586,6 +591,7 @@

Method Details

"maxTimeTravelHours": "A String", # [Optional] Number of hours for the max time travel for all tables in the dataset. "satisfiesPzs": True or False, # [Output-only] Reserved for future use. "selfLink": "A String", # [Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource. + "storageBillingModel": "A String", # [Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL. "tags": [ # [Optional]The tags associated with this dataset. Tag keys are globally unique. { "tagKey": "A String", # [Required] The namespaced friendly name of the tag key, e.g. "12345/environment" where 12345 is org id. @@ -653,6 +659,7 @@

Method Details

"maxTimeTravelHours": "A String", # [Optional] Number of hours for the max time travel for all tables in the dataset. "satisfiesPzs": True or False, # [Output-only] Reserved for future use. "selfLink": "A String", # [Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource. + "storageBillingModel": "A String", # [Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL. "tags": [ # [Optional]The tags associated with this dataset. Tag keys are globally unique. { "tagKey": "A String", # [Required] The namespaced friendly name of the tag key, e.g. "12345/environment" where 12345 is org id. diff --git a/docs/dyn/bigquery_v2.jobs.html b/docs/dyn/bigquery_v2.jobs.html index 350893b96fb..72968c4c869 100644 --- a/docs/dyn/bigquery_v2.jobs.html +++ b/docs/dyn/bigquery_v2.jobs.html @@ -409,6 +409,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -1102,6 +1104,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -1866,6 +1870,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -2534,6 +2540,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -3225,6 +3233,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. diff --git a/docs/dyn/bigquery_v2.tables.html b/docs/dyn/bigquery_v2.tables.html index 6a3079186cc..e5b84e565f2 100644 --- a/docs/dyn/bigquery_v2.tables.html +++ b/docs/dyn/bigquery_v2.tables.html @@ -218,6 +218,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -533,6 +535,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -789,6 +793,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -1128,6 +1134,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -1385,6 +1393,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -1756,6 +1766,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. @@ -2013,6 +2025,8 @@

Method Details

}, "ignoreUnknownValues": True or False, # [Optional] Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore backups: This setting is ignored. Avro: This setting is ignored. "maxBadRecords": 42, # [Optional] The maximum number of bad records that BigQuery can ignore when reading data. If the number of bad records exceeds this value, an invalid error is returned in the job result. This is only valid for CSV, JSON, and Google Sheets. The default value is 0, which requires that all records are valid. This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore backups and Avro formats. + "metadataCacheMode": "A String", # [Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source. + "objectMetadata": "A String", # ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type. "parquetOptions": { # Additional properties to set if sourceFormat is set to Parquet. "enableListInference": True or False, # [Optional] Indicates whether to use schema inference specifically for Parquet LIST logical type. "enumAsString": True or False, # [Optional] Indicates whether to infer Parquet ENUM logical type as STRING instead of BYTES by default. diff --git a/googleapiclient/discovery_cache/documents/bigquery.v2.json b/googleapiclient/discovery_cache/documents/bigquery.v2.json index c8547f02ffd..28a2f90b574 100644 --- a/googleapiclient/discovery_cache/documents/bigquery.v2.json +++ b/googleapiclient/discovery_cache/documents/bigquery.v2.json @@ -1710,7 +1710,7 @@ } } }, - "revision": "20220924", + "revision": "20221015", "rootUrl": "https://bigquery.googleapis.com/", "schemas": { "AggregateClassificationMetrics": { @@ -2856,6 +2856,10 @@ "description": "[Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource.", "type": "string" }, + "storageBillingModel": { + "description": "[Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL.", + "type": "string" + }, "tags": { "description": "[Optional]The tags associated with this dataset. Tag keys are globally unique.", "items": { @@ -3435,6 +3439,14 @@ "format": "int32", "type": "integer" }, + "metadataCacheMode": { + "description": "[Optional] Metadata Cache Mode for the table. Set this to enable caching of metadata from external data source.", + "type": "string" + }, + "objectMetadata": { + "description": "ObjectMetadata is used to create Object Tables. Object Tables contain a listing of objects (with their metadata) found at the source_uris. If ObjectMetadata is set, source_format should be omitted. Currently SIMPLE is the only supported Object Metadata type.", + "type": "string" + }, "parquetOptions": { "$ref": "ParquetOptions", "description": "Additional properties to set if sourceFormat is set to Parquet."