Skip to content

Commit

Permalink
feat(dataproc): update the API
Browse files Browse the repository at this point in the history
#### dataproc:v1

The following keys were added:
- schemas.Session.properties.sparkConnectSession.$ref
- schemas.Session.properties.sparkConnectSession.description
- schemas.SessionTemplate.properties.sparkConnectSession.$ref
- schemas.SessionTemplate.properties.sparkConnectSession.description
- schemas.SparkConnectConfig.description
- schemas.SparkConnectConfig.id
- schemas.SparkConnectConfig.type

The following keys were changed:
- endpoints
- schemas.SoftwareConfig.properties.imageVersion.description
- schemas.SoftwareConfig.properties.optionalComponents.items.enumDescriptions
  • Loading branch information
yoshi-automation authored and sofisl committed Aug 30, 2024
1 parent 65db039 commit 0ce260e
Show file tree
Hide file tree
Showing 2 changed files with 35 additions and 4 deletions.
25 changes: 22 additions & 3 deletions discovery/dataproc-v1.json
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,11 @@
"endpointUrl": "https://dataproc.europe-west9.rep.googleapis.com/",
"location": "europe-west9"
},
{
"description": "Regional Endpoint",
"endpointUrl": "https://dataproc.us-central1.rep.googleapis.com/",
"location": "us-central1"
},
{
"description": "Regional Endpoint",
"endpointUrl": "https://dataproc.me-central2.rep.googleapis.com/",
Expand Down Expand Up @@ -3072,7 +3077,7 @@
}
}
},
"revision": "20240617",
"revision": "20240821",
"rootUrl": "https://dataproc.googleapis.com/",
"schemas": {
"AcceleratorConfig": {
Expand Down Expand Up @@ -6375,6 +6380,10 @@
"description": "Optional. The session template used by the session.Only resource names, including project ID and location, are valid.Example: * https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/sessionTemplates/[template_id] * projects/[project_id]/locations/[dataproc_region]/sessionTemplates/[template_id]The template must be in the same project and Dataproc region as the session.",
"type": "string"
},
"sparkConnectSession": {
"$ref": "SparkConnectConfig",
"description": "Optional. Spark connect session config."
},
"state": {
"description": "Output only. A state of the session.",
"enum": [
Expand Down Expand Up @@ -6567,6 +6576,10 @@
"$ref": "RuntimeConfig",
"description": "Optional. Runtime configuration for session execution."
},
"sparkConnectSession": {
"$ref": "SparkConnectConfig",
"description": "Optional. Spark connect session config."
},
"updateTime": {
"description": "Output only. The time the template was last updated.",
"format": "google-datetime",
Expand Down Expand Up @@ -6616,7 +6629,7 @@
"id": "SoftwareConfig",
"properties": {
"imageVersion": {
"description": "Optional. The version of software inside the cluster. It must be one of the supported Dataproc Versions (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_dataproc_versions), such as \"1.2\" (including a subminor version, such as \"1.2.29\"), or the \"preview\" version (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). If unspecified, it defaults to the latest Debian version.",
"description": "Optional. The version of software inside the cluster. It must be one of the supported Dataproc Versions (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported-dataproc-image-versions), such as \"1.2\" (including a subminor version, such as \"1.2.29\"), or the \"preview\" version (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). If unspecified, it defaults to the latest Debian version.",
"type": "string"
},
"optionalComponents": {
Expand All @@ -6641,7 +6654,7 @@
],
"enumDescriptions": [
"Unspecified component. Specifying this will cause Cluster creation to fail.",
"The Anaconda python distribution. The Anaconda component is not supported in the Dataproc 2.0 image. The 2.0 image is pre-installed with Miniconda.",
"The Anaconda component is no longer supported or applicable to supported Dataproc on Compute Engine image versions (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-version-clusters#supported-dataproc-image-versions). It cannot be activated on clusters created with supported Dataproc on Compute Engine image versions.",
"Docker",
"The Druid query engine. (alpha)",
"Flink",
Expand Down Expand Up @@ -6713,6 +6726,12 @@
},
"type": "object"
},
"SparkConnectConfig": {
"description": "Spark connect configuration for an interactive session.",
"id": "SparkConnectConfig",
"properties": {},
"type": "object"
},
"SparkHistoryServerConfig": {
"description": "Spark History Server configuration for the workload.",
"id": "SparkHistoryServerConfig",
Expand Down
14 changes: 13 additions & 1 deletion src/apis/dataproc/v1.ts
Original file line number Diff line number Diff line change
Expand Up @@ -2435,6 +2435,10 @@ export namespace dataproc_v1 {
* Optional. The session template used by the session.Only resource names, including project ID and location, are valid.Example: * https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/sessionTemplates/[template_id] * projects/[project_id]/locations/[dataproc_region]/sessionTemplates/[template_id]The template must be in the same project and Dataproc region as the session.
*/
sessionTemplate?: string | null;
/**
* Optional. Spark connect session config.
*/
sparkConnectSession?: Schema$SparkConnectConfig;
/**
* Output only. A state of the session.
*/
Expand Down Expand Up @@ -2550,6 +2554,10 @@ export namespace dataproc_v1 {
* Optional. Runtime configuration for session execution.
*/
runtimeConfig?: Schema$RuntimeConfig;
/**
* Optional. Spark connect session config.
*/
sparkConnectSession?: Schema$SparkConnectConfig;
/**
* Output only. The time the template was last updated.
*/
Expand Down Expand Up @@ -2590,7 +2598,7 @@ export namespace dataproc_v1 {
*/
export interface Schema$SoftwareConfig {
/**
* Optional. The version of software inside the cluster. It must be one of the supported Dataproc Versions (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_dataproc_versions), such as "1.2" (including a subminor version, such as "1.2.29"), or the "preview" version (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). If unspecified, it defaults to the latest Debian version.
* Optional. The version of software inside the cluster. It must be one of the supported Dataproc Versions (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported-dataproc-image-versions), such as "1.2" (including a subminor version, such as "1.2.29"), or the "preview" version (https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). If unspecified, it defaults to the latest Debian version.
*/
imageVersion?: string | null;
/**
Expand Down Expand Up @@ -2631,6 +2639,10 @@ export namespace dataproc_v1 {
*/
mainJarFileUri?: string | null;
}
/**
* Spark connect configuration for an interactive session.
*/
export interface Schema$SparkConnectConfig {}
/**
* Spark History Server configuration for the workload.
*/
Expand Down

0 comments on commit 0ce260e

Please sign in to comment.