Skip to content

Commit

Permalink
feat(dataplex): update the api
Browse files Browse the repository at this point in the history
#### dataplex:v1

The following keys were deleted:
- schemas.GoogleCloudDataplexV1SessionEvent.properties.warmPoolEnabled.type (Total Keys: 1)

The following keys were added:
- schemas.GoogleCloudDataplexV1SessionEvent.properties.fastStartupEnabled.type (Total Keys: 1)
  • Loading branch information
yoshi-automation committed Oct 25, 2022
1 parent 25af7f8 commit ca9cc14
Show file tree
Hide file tree
Showing 2 changed files with 20 additions and 20 deletions.
24 changes: 12 additions & 12 deletions docs/dyn/dataplex_v1.projects.locations.lakes.tasks.html
Expand Up @@ -156,10 +156,10 @@ <h3>Method Details</h3>
},
&quot;name&quot;: &quot;A String&quot;, # Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
&quot;notebook&quot;: { # Config for running scheduled notebooks. # Config related to running scheduled Notebooks.
&quot;archiveUris&quot;: [ # Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
&quot;archiveUris&quot;: [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
&quot;A String&quot;,
],
&quot;fileUris&quot;: [ # Optional. GCS URIs of files to be placed in the working directory of each executor.
&quot;fileUris&quot;: [ # Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
&quot;A String&quot;,
],
&quot;infrastructureSpec&quot;: { # Configuration for the underlying infrastructure used to run workloads. # Optional. Infrastructure specification for the execution.
Expand Down Expand Up @@ -187,7 +187,7 @@ <h3>Method Details</h3>
&quot;subNetwork&quot;: &quot;A String&quot;, # Optional. The Cloud VPC sub-network in which the job is run.
},
},
&quot;notebook&quot;: &quot;A String&quot;, # Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
&quot;notebook&quot;: &quot;A String&quot;, # Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
},
&quot;spark&quot;: { # User-specified config for running a Spark task. # Config related to running custom Spark tasks.
&quot;archiveUris&quot;: [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
Expand Down Expand Up @@ -351,10 +351,10 @@ <h3>Method Details</h3>
},
&quot;name&quot;: &quot;A String&quot;, # Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
&quot;notebook&quot;: { # Config for running scheduled notebooks. # Config related to running scheduled Notebooks.
&quot;archiveUris&quot;: [ # Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
&quot;archiveUris&quot;: [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
&quot;A String&quot;,
],
&quot;fileUris&quot;: [ # Optional. GCS URIs of files to be placed in the working directory of each executor.
&quot;fileUris&quot;: [ # Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
&quot;A String&quot;,
],
&quot;infrastructureSpec&quot;: { # Configuration for the underlying infrastructure used to run workloads. # Optional. Infrastructure specification for the execution.
Expand Down Expand Up @@ -382,7 +382,7 @@ <h3>Method Details</h3>
&quot;subNetwork&quot;: &quot;A String&quot;, # Optional. The Cloud VPC sub-network in which the job is run.
},
},
&quot;notebook&quot;: &quot;A String&quot;, # Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
&quot;notebook&quot;: &quot;A String&quot;, # Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
},
&quot;spark&quot;: { # User-specified config for running a Spark task. # Config related to running custom Spark tasks.
&quot;archiveUris&quot;: [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
Expand Down Expand Up @@ -536,10 +536,10 @@ <h3>Method Details</h3>
},
&quot;name&quot;: &quot;A String&quot;, # Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
&quot;notebook&quot;: { # Config for running scheduled notebooks. # Config related to running scheduled Notebooks.
&quot;archiveUris&quot;: [ # Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
&quot;archiveUris&quot;: [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
&quot;A String&quot;,
],
&quot;fileUris&quot;: [ # Optional. GCS URIs of files to be placed in the working directory of each executor.
&quot;fileUris&quot;: [ # Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
&quot;A String&quot;,
],
&quot;infrastructureSpec&quot;: { # Configuration for the underlying infrastructure used to run workloads. # Optional. Infrastructure specification for the execution.
Expand Down Expand Up @@ -567,7 +567,7 @@ <h3>Method Details</h3>
&quot;subNetwork&quot;: &quot;A String&quot;, # Optional. The Cloud VPC sub-network in which the job is run.
},
},
&quot;notebook&quot;: &quot;A String&quot;, # Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
&quot;notebook&quot;: &quot;A String&quot;, # Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
},
&quot;spark&quot;: { # User-specified config for running a Spark task. # Config related to running custom Spark tasks.
&quot;archiveUris&quot;: [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
Expand Down Expand Up @@ -680,10 +680,10 @@ <h3>Method Details</h3>
},
&quot;name&quot;: &quot;A String&quot;, # Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
&quot;notebook&quot;: { # Config for running scheduled notebooks. # Config related to running scheduled Notebooks.
&quot;archiveUris&quot;: [ # Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
&quot;archiveUris&quot;: [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
&quot;A String&quot;,
],
&quot;fileUris&quot;: [ # Optional. GCS URIs of files to be placed in the working directory of each executor.
&quot;fileUris&quot;: [ # Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
&quot;A String&quot;,
],
&quot;infrastructureSpec&quot;: { # Configuration for the underlying infrastructure used to run workloads. # Optional. Infrastructure specification for the execution.
Expand Down Expand Up @@ -711,7 +711,7 @@ <h3>Method Details</h3>
&quot;subNetwork&quot;: &quot;A String&quot;, # Optional. The Cloud VPC sub-network in which the job is run.
},
},
&quot;notebook&quot;: &quot;A String&quot;, # Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
&quot;notebook&quot;: &quot;A String&quot;, # Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).
},
&quot;spark&quot;: { # User-specified config for running a Spark task. # Config related to running custom Spark tasks.
&quot;archiveUris&quot;: [ # Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
Expand Down
16 changes: 8 additions & 8 deletions googleapiclient/discovery_cache/documents/dataplex.v1.json
Expand Up @@ -2764,7 +2764,7 @@
}
}
},
"revision": "20220928",
"revision": "20221015",
"rootUrl": "https://dataplex.googleapis.com/",
"schemas": {
"Empty": {
Expand Down Expand Up @@ -4693,6 +4693,10 @@
"description": "The status of the event.",
"type": "boolean"
},
"fastStartupEnabled": {
"description": "If the session is associated with an environment with fast startup enabled, and was created before being assigned to a user.",
"type": "boolean"
},
"message": {
"description": "The log message.",
"type": "string"
Expand Down Expand Up @@ -4731,10 +4735,6 @@
"userId": {
"description": "The information about the user that created the session. It will be the email address of the user.",
"type": "string"
},
"warmPoolEnabled": {
"description": "If the session is a warm pooled session.",
"type": "boolean"
}
},
"type": "object"
Expand Down Expand Up @@ -5114,14 +5114,14 @@
"id": "GoogleCloudDataplexV1TaskNotebookTaskConfig",
"properties": {
"archiveUris": {
"description": "Optional. GCS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.",
"description": "Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.",
"items": {
"type": "string"
},
"type": "array"
},
"fileUris": {
"description": "Optional. GCS URIs of files to be placed in the working directory of each executor.",
"description": "Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.",
"items": {
"type": "string"
},
Expand All @@ -5132,7 +5132,7 @@
"description": "Optional. Infrastructure specification for the execution."
},
"notebook": {
"description": "Required. Path to input notebook. This can be the GCS URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).",
"description": "Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (TASK_key=value).",
"type": "string"
}
},
Expand Down

0 comments on commit ca9cc14

Please sign in to comment.