Skip to content
This repository was archived by the owner on Nov 29, 2023. It is now read-only.

Commit f2e35d9

Browse files
feat: add Spark runtime versioning for Spark batches (#318)
- [ ] Regenerate this pull request now. Committer: @medb PiperOrigin-RevId: 420175701 Source-Link: googleapis/googleapis@1116092 Source-Link: https://github.com/googleapis/googleapis-gen/commit/72c0b7f4dcdf2cd4d4729769cbdeb2b228c504e1 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiNzJjMGI3ZjRkY2RmMmNkNGQ0NzI5NzY5Y2JkZWIyYjIyOGM1MDRlMSJ9 feat: custom image containers for Spark batches feat: auto-diagnostic of failed Spark batches feat: local SSD NVME interface support for GCE clusters
1 parent 0011b32 commit f2e35d9

File tree

9 files changed

+46
-20
lines changed

9 files changed

+46
-20
lines changed

google/cloud/dataproc_v1/services/batch_controller/async_client.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -208,7 +208,7 @@ async def create_batch(
208208
the final component of the batch's resource name.
209209
210210
This value must be 4-63 characters. Valid characters are
211-
``/[a-z][0-9]-/``.
211+
/[a-z][0-9]-/.
212212
213213
This corresponds to the ``batch_id`` field
214214
on the ``request`` instance; if ``request`` is provided, this

google/cloud/dataproc_v1/services/batch_controller/client.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -400,7 +400,7 @@ def create_batch(
400400
the final component of the batch's resource name.
401401
402402
This value must be 4-63 characters. Valid characters are
403-
``/[a-z][0-9]-/``.
403+
/[a-z][0-9]-/.
404404
405405
This corresponds to the ``batch_id`` field
406406
on the ``request`` instance; if ``request`` is provided, this

google/cloud/dataproc_v1/services/workflow_template_service/async_client.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -207,7 +207,7 @@ async def create_workflow_template(
207207
as described in
208208
https://cloud.google.com/apis/design/resource_names.
209209
210-
- For ``projects.regions.workflowTemplates,create``,
210+
- For ``projects.regions.workflowTemplates.create``,
211211
the resource name of the region has the following
212212
format: ``projects/{project_id}/regions/{region}``
213213

google/cloud/dataproc_v1/services/workflow_template_service/client.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -422,7 +422,7 @@ def create_workflow_template(
422422
as described in
423423
https://cloud.google.com/apis/design/resource_names.
424424
425-
- For ``projects.regions.workflowTemplates,create``,
425+
- For ``projects.regions.workflowTemplates.create``,
426426
the resource name of the region has the following
427427
format: ``projects/{project_id}/regions/{region}``
428428

google/cloud/dataproc_v1/types/batches.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ class CreateBatchRequest(proto.Message):
5050
final component of the batch's resource name.
5151
5252
This value must be 4-63 characters. Valid characters are
53-
``/[a-z][0-9]-/``.
53+
/[a-z][0-9]-/.
5454
request_id (str):
5555
Optional. A unique ID used to identify the request. If the
5656
service receives two

google/cloud/dataproc_v1/types/clusters.py

Lines changed: 10 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -75,9 +75,10 @@ class Cluster(proto.Message):
7575
within a project must be unique. Names of
7676
deleted clusters can be reused.
7777
config (google.cloud.dataproc_v1.types.ClusterConfig):
78-
Required. The cluster config. Note that
79-
Dataproc may set default values, and values may
80-
change when clusters are updated.
78+
Optional. The cluster config for a cluster of
79+
Compute Engine Instances. Note that Dataproc may
80+
set default values, and values may change when
81+
clusters are updated.
8182
labels (Sequence[google.cloud.dataproc_v1.types.Cluster.LabelsEntry]):
8283
Optional. The labels to associate with this cluster. Label
8384
**keys** must contain 1 to 63 characters, and must conform
@@ -682,11 +683,17 @@ class DiskConfig(proto.Message):
682683
data. If one or more SSDs are attached, this runtime bulk
683684
data is spread across them, and the boot disk contains only
684685
basic config and installed binaries.
686+
local_ssd_interface (str):
687+
Optional. Interface type of local SSDs (default is "scsi").
688+
Valid values: "scsi" (Small Computer System Interface),
689+
"nvme" (Non-Volatile Memory Express). See `SSD Interface
690+
types <https://cloud.google.com/compute/docs/disks/local-ssd#performance>`__.
685691
"""
686692

687693
boot_disk_type = proto.Field(proto.STRING, number=3,)
688694
boot_disk_size_gb = proto.Field(proto.INT32, number=1,)
689695
num_local_ssds = proto.Field(proto.INT32, number=2,)
696+
local_ssd_interface = proto.Field(proto.STRING, number=4,)
690697

691698

692699
class NodeInitializationAction(proto.Message):

google/cloud/dataproc_v1/types/jobs.py

Lines changed: 18 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -814,21 +814,28 @@ class JobScheduling(proto.Message):
814814
815815
Attributes:
816816
max_failures_per_hour (int):
817-
Optional. Maximum number of times per hour a
818-
driver may be restarted as a result of driver
819-
exiting with non-zero code before job is
820-
reported failed.
817+
Optional. Maximum number of times per hour a driver may be
818+
restarted as a result of driver exiting with non-zero code
819+
before job is reported failed.
821820
822-
A job may be reported as thrashing if driver
823-
exits with non-zero code 4 times within 10
824-
minute window.
821+
A job may be reported as thrashing if driver exits with
822+
non-zero code 4 times within 10 minute window.
825823
826824
Maximum value is 10.
825+
826+
**Note:** Currently, this restartable job option is not
827+
supported in Dataproc `workflow
828+
template <https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template>`__
829+
jobs.
827830
max_failures_total (int):
828-
Optional. Maximum number of times in total a
829-
driver may be restarted as a result of driver
830-
exiting with non-zero code before job is
831-
reported failed. Maximum value is 240.
831+
Optional. Maximum number of times in total a driver may be
832+
restarted as a result of driver exiting with non-zero code
833+
before job is reported failed. Maximum value is 240.
834+
835+
**Note:** Currently, this restartable job option is not
836+
supported in Dataproc `workflow
837+
template <https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template>`__
838+
jobs.
832839
"""
833840

834841
max_failures_per_hour = proto.Field(proto.INT32, number=1,)

google/cloud/dataproc_v1/types/shared.py

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -61,12 +61,20 @@ class RuntimeConfig(proto.Message):
6161
r"""Runtime configuration for a workload.
6262
6363
Attributes:
64+
version (str):
65+
Optional. Version of the batch runtime.
66+
container_image (str):
67+
Optional. Optional custom container image for
68+
the job runtime environment. If not specified, a
69+
default container image will be used.
6470
properties (Sequence[google.cloud.dataproc_v1.types.RuntimeConfig.PropertiesEntry]):
6571
Optional. A mapping of property names to
6672
values, which are used to configure workload
6773
execution.
6874
"""
6975

76+
version = proto.Field(proto.STRING, number=1,)
77+
container_image = proto.Field(proto.STRING, number=2,)
7078
properties = proto.MapField(proto.STRING, proto.STRING, number=3,)
7179

7280

@@ -174,10 +182,14 @@ class RuntimeInfo(proto.Message):
174182
output_uri (str):
175183
Output only. A URI pointing to the location
176184
of the stdout and stderr of the workload.
185+
diagnostic_output_uri (str):
186+
Output only. A URI pointing to the location
187+
of the diagnostics tarball.
177188
"""
178189

179190
endpoints = proto.MapField(proto.STRING, proto.STRING, number=1,)
180191
output_uri = proto.Field(proto.STRING, number=2,)
192+
diagnostic_output_uri = proto.Field(proto.STRING, number=3,)
181193

182194

183195
__all__ = tuple(sorted(__protobuf__.manifest))

google/cloud/dataproc_v1/types/workflow_templates.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -626,7 +626,7 @@ class CreateWorkflowTemplateRequest(proto.Message):
626626
described in
627627
https://cloud.google.com/apis/design/resource_names.
628628
629-
- For ``projects.regions.workflowTemplates,create``, the
629+
- For ``projects.regions.workflowTemplates.create``, the
630630
resource name of the region has the following format:
631631
``projects/{project_id}/regions/{region}``
632632

0 commit comments

Comments
 (0)