You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
#### dataplex:v1
The following keys were deleted:
- schemas.GoogleCloudDataplexV1SessionEvent.properties.warmPoolEnabled.type (Total Keys: 1)
The following keys were added:
- schemas.GoogleCloudDataplexV1SessionEvent.properties.fastStartupEnabled.type (Total Keys: 1)
Copy file name to clipboardexpand all lines: docs/dyn/dataplex_v1.projects.locations.lakes.tasks.html
+12-12
Original file line number
Diff line number
Diff line change
@@ -156,10 +156,10 @@ <h3>Method Details</h3>
156
156
},
157
157
"name": "A String", # Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
158
158
"notebook": { # Config for running scheduled notebooks. # Config related to running scheduled Notebooks.
159
-
"archiveUris": [ # Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
159
+
"archiveUris": [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
160
160
"A String",
161
161
],
162
-
"fileUris": [ # Optional. GCS URIs of files to be placed in the working directory of each executor.
162
+
"fileUris": [ # Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
163
163
"A String",
164
164
],
165
165
"infrastructureSpec": { # Configuration for the underlying infrastructure used to run workloads. # Optional. Infrastructure specification for the execution.
@@ -187,7 +187,7 @@ <h3>Method Details</h3>
187
187
"subNetwork": "A String", # Optional. The Cloud VPC sub-network in which the job is run.
188
188
},
189
189
},
190
-
"notebook": "A String", # Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
190
+
"notebook": "A String", # Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
191
191
},
192
192
"spark": { # User-specified config for running a Spark task. # Config related to running custom Spark tasks.
193
193
"archiveUris": [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -351,10 +351,10 @@ <h3>Method Details</h3>
351
351
},
352
352
"name": "A String", # Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
353
353
"notebook": { # Config for running scheduled notebooks. # Config related to running scheduled Notebooks.
354
-
"archiveUris": [ # Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
354
+
"archiveUris": [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
355
355
"A String",
356
356
],
357
-
"fileUris": [ # Optional. GCS URIs of files to be placed in the working directory of each executor.
357
+
"fileUris": [ # Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
358
358
"A String",
359
359
],
360
360
"infrastructureSpec": { # Configuration for the underlying infrastructure used to run workloads. # Optional. Infrastructure specification for the execution.
@@ -382,7 +382,7 @@ <h3>Method Details</h3>
382
382
"subNetwork": "A String", # Optional. The Cloud VPC sub-network in which the job is run.
383
383
},
384
384
},
385
-
"notebook": "A String", # Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
385
+
"notebook": "A String", # Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
386
386
},
387
387
"spark": { # User-specified config for running a Spark task. # Config related to running custom Spark tasks.
388
388
"archiveUris": [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -536,10 +536,10 @@ <h3>Method Details</h3>
536
536
},
537
537
"name": "A String", # Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
538
538
"notebook": { # Config for running scheduled notebooks. # Config related to running scheduled Notebooks.
539
-
"archiveUris": [ # Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
539
+
"archiveUris": [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
540
540
"A String",
541
541
],
542
-
"fileUris": [ # Optional. GCS URIs of files to be placed in the working directory of each executor.
542
+
"fileUris": [ # Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
543
543
"A String",
544
544
],
545
545
"infrastructureSpec": { # Configuration for the underlying infrastructure used to run workloads. # Optional. Infrastructure specification for the execution.
@@ -567,7 +567,7 @@ <h3>Method Details</h3>
567
567
"subNetwork": "A String", # Optional. The Cloud VPC sub-network in which the job is run.
568
568
},
569
569
},
570
-
"notebook": "A String", # Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
570
+
"notebook": "A String", # Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
571
571
},
572
572
"spark": { # User-specified config for running a Spark task. # Config related to running custom Spark tasks.
573
573
"archiveUris": [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -680,10 +680,10 @@ <h3>Method Details</h3>
680
680
},
681
681
"name": "A String", # Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
682
682
"notebook": { # Config for running scheduled notebooks. # Config related to running scheduled Notebooks.
683
-
"archiveUris": [ # Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
683
+
"archiveUris": [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
684
684
"A String",
685
685
],
686
-
"fileUris": [ # Optional. GCS URIs of files to be placed in the working directory of each executor.
686
+
"fileUris": [ # Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
687
687
"A String",
688
688
],
689
689
"infrastructureSpec": { # Configuration for the underlying infrastructure used to run workloads. # Optional. Infrastructure specification for the execution.
@@ -711,7 +711,7 @@ <h3>Method Details</h3>
711
711
"subNetwork": "A String", # Optional. The Cloud VPC sub-network in which the job is run.
712
712
},
713
713
},
714
-
"notebook": "A String", # Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
714
+
"notebook": "A String", # Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
715
715
},
716
716
"spark": { # User-specified config for running a Spark task. # Config related to running custom Spark tasks.
717
717
"archiveUris": [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
"description": "Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.",
5117
+
"description": "Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.",
5118
5118
"items": {
5119
5119
"type": "string"
5120
5120
},
5121
5121
"type": "array"
5122
5122
},
5123
5123
"fileUris": {
5124
-
"description": "Optional. GCS URIs of files to be placed in the working directory of each executor.",
5124
+
"description": "Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.",
5125
5125
"items": {
5126
5126
"type": "string"
5127
5127
},
@@ -5132,7 +5132,7 @@
5132
5132
"description": "Optional. Infrastructure specification for the execution."
5133
5133
},
5134
5134
"notebook": {
5135
-
"description": "Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).",
5135
+
"description": "Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).",
0 commit comments