From 615854c0b2097c0437f63ce465ce350ca7948517 Mon Sep 17 00:00:00 2001 From: "gcf-owl-bot[bot]" <78513119+gcf-owl-bot[bot]@users.noreply.github.com> Date: Fri, 14 Jul 2023 10:17:00 -0400 Subject: [PATCH] chore: Update gapic-generator-python to v1.11.4 (#543) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * chore: Update gapic-generator-python to v1.11.4 PiperOrigin-RevId: 547897126 Source-Link: https://github.com/googleapis/googleapis/commit/c09c75e087d8f9a2d466b4aaad7dd2926b5ead5a Source-Link: https://github.com/googleapis/googleapis-gen/commit/45e0ec4343517cd0aa66b5ca64232a1802c2f945 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiNDVlMGVjNDM0MzUxN2NkMGFhNjZiNWNhNjQyMzJhMTgwMmMyZjk0NSJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md --------- Co-authored-by: Owl Bot --- google/cloud/dataproc_v1/types/clusters.py | 1 + google/cloud/dataproc_v1/types/jobs.py | 22 ++++++++++++------- .../dataproc_v1/types/workflow_templates.py | 1 + 3 files changed, 16 insertions(+), 8 deletions(-) diff --git a/google/cloud/dataproc_v1/types/clusters.py b/google/cloud/dataproc_v1/types/clusters.py index 267a6040..bfae67da 100644 --- a/google/cloud/dataproc_v1/types/clusters.py +++ b/google/cloud/dataproc_v1/types/clusters.py @@ -1285,6 +1285,7 @@ class KerberosConfig(proto.Message): enable_kerberos (bool): Optional. Flag to indicate whether to Kerberize the cluster (default: + false). Set this field to true to enable Kerberos on a cluster. root_principal_password_uri (str): diff --git a/google/cloud/dataproc_v1/types/jobs.py b/google/cloud/dataproc_v1/types/jobs.py index 7eecbffd..c666f5ac 100644 --- a/google/cloud/dataproc_v1/types/jobs.py +++ b/google/cloud/dataproc_v1/types/jobs.py @@ -64,6 +64,7 @@ class LoggingConfig(proto.Message): The per-package log levels for the driver. This may include "root" package name to configure rootLogger. Examples: + 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' """ @@ -129,6 +130,7 @@ class HadoopJob(proto.Message): main_jar_file_uri (str): The HCFS URI of the jar file containing the main class. Examples: + 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' 'hdfs:/tmp/test-samples/custom-wordcount.jar' 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar' @@ -156,8 +158,9 @@ class HadoopJob(proto.Message): archive_uris (MutableSequence[str]): Optional. HCFS URIs of archives to be extracted in the working directory of Hadoop - drivers and tasks. Supported file types: .jar, - .tar, .tar.gz, .tgz, or .zip. + drivers and tasks. Supported file types: + + .jar, .tar, .tar.gz, .tgz, or .zip. properties (MutableMapping[str, str]): Optional. A mapping of property names to values, used to configure Hadoop. Properties that conflict with values set @@ -245,8 +248,9 @@ class SparkJob(proto.Message): archive_uris (MutableSequence[str]): Optional. HCFS URIs of archives to be extracted into the working directory of each - executor. Supported file types: .jar, .tar, - .tar.gz, .tgz, and .zip. + executor. Supported file types: + + .jar, .tar, .tar.gz, .tgz, and .zip. properties (MutableMapping[str, str]): Optional. A mapping of property names to values, used to configure Spark. Properties that @@ -325,8 +329,9 @@ class PySparkJob(proto.Message): archive_uris (MutableSequence[str]): Optional. HCFS URIs of archives to be extracted into the working directory of each - executor. Supported file types: .jar, .tar, - .tar.gz, .tgz, and .zip. + executor. Supported file types: + + .jar, .tar, .tar.gz, .tgz, and .zip. properties (MutableMapping[str, str]): Optional. A mapping of property names to values, used to configure PySpark. Properties @@ -648,8 +653,9 @@ class SparkRJob(proto.Message): archive_uris (MutableSequence[str]): Optional. HCFS URIs of archives to be extracted into the working directory of each - executor. Supported file types: .jar, .tar, - .tar.gz, .tgz, and .zip. + executor. Supported file types: + + .jar, .tar, .tar.gz, .tgz, and .zip. properties (MutableMapping[str, str]): Optional. A mapping of property names to values, used to configure SparkR. Properties diff --git a/google/cloud/dataproc_v1/types/workflow_templates.py b/google/cloud/dataproc_v1/types/workflow_templates.py index d7292204..95341cb7 100644 --- a/google/cloud/dataproc_v1/types/workflow_templates.py +++ b/google/cloud/dataproc_v1/types/workflow_templates.py @@ -437,6 +437,7 @@ class OrderedJob(proto.Message): class TemplateParameter(proto.Message): r"""A configurable parameter that replaces one or more fields in the template. Parameterizable fields: + - Labels - File uris - Job properties