Skip to content
This repository was archived by the owner on Nov 29, 2023. It is now read-only.

Commit 615854c

Browse files
chore: Update gapic-generator-python to v1.11.4 (#543)
* chore: Update gapic-generator-python to v1.11.4 PiperOrigin-RevId: 547897126 Source-Link: googleapis/googleapis@c09c75e Source-Link: https://github.com/googleapis/googleapis-gen/commit/45e0ec4343517cd0aa66b5ca64232a1802c2f945 Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiNDVlMGVjNDM0MzUxN2NkMGFhNjZiNWNhNjQyMzJhMTgwMmMyZjk0NSJ9 * 🦉 Updates from OwlBot post-processor See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md --------- Co-authored-by: Owl Bot <gcf-owl-bot[bot]@users.noreply.github.com>
1 parent 098c954 commit 615854c

File tree

3 files changed

+16
-8
lines changed

3 files changed

+16
-8
lines changed

google/cloud/dataproc_v1/types/clusters.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1285,6 +1285,7 @@ class KerberosConfig(proto.Message):
12851285
enable_kerberos (bool):
12861286
Optional. Flag to indicate whether to
12871287
Kerberize the cluster (default:
1288+
12881289
false). Set this field to true to enable
12891290
Kerberos on a cluster.
12901291
root_principal_password_uri (str):

google/cloud/dataproc_v1/types/jobs.py

Lines changed: 14 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -64,6 +64,7 @@ class LoggingConfig(proto.Message):
6464
The per-package log levels for the driver.
6565
This may include "root" package name to
6666
configure rootLogger. Examples:
67+
6768
'com.google = FATAL', 'root = INFO',
6869
'org.apache = DEBUG'
6970
"""
@@ -129,6 +130,7 @@ class HadoopJob(proto.Message):
129130
main_jar_file_uri (str):
130131
The HCFS URI of the jar file containing the
131132
main class. Examples:
133+
132134
'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar'
133135
'hdfs:/tmp/test-samples/custom-wordcount.jar'
134136
'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar'
@@ -156,8 +158,9 @@ class HadoopJob(proto.Message):
156158
archive_uris (MutableSequence[str]):
157159
Optional. HCFS URIs of archives to be
158160
extracted in the working directory of Hadoop
159-
drivers and tasks. Supported file types: .jar,
160-
.tar, .tar.gz, .tgz, or .zip.
161+
drivers and tasks. Supported file types:
162+
163+
.jar, .tar, .tar.gz, .tgz, or .zip.
161164
properties (MutableMapping[str, str]):
162165
Optional. A mapping of property names to values, used to
163166
configure Hadoop. Properties that conflict with values set
@@ -245,8 +248,9 @@ class SparkJob(proto.Message):
245248
archive_uris (MutableSequence[str]):
246249
Optional. HCFS URIs of archives to be
247250
extracted into the working directory of each
248-
executor. Supported file types: .jar, .tar,
249-
.tar.gz, .tgz, and .zip.
251+
executor. Supported file types:
252+
253+
.jar, .tar, .tar.gz, .tgz, and .zip.
250254
properties (MutableMapping[str, str]):
251255
Optional. A mapping of property names to
252256
values, used to configure Spark. Properties that
@@ -325,8 +329,9 @@ class PySparkJob(proto.Message):
325329
archive_uris (MutableSequence[str]):
326330
Optional. HCFS URIs of archives to be
327331
extracted into the working directory of each
328-
executor. Supported file types: .jar, .tar,
329-
.tar.gz, .tgz, and .zip.
332+
executor. Supported file types:
333+
334+
.jar, .tar, .tar.gz, .tgz, and .zip.
330335
properties (MutableMapping[str, str]):
331336
Optional. A mapping of property names to
332337
values, used to configure PySpark. Properties
@@ -648,8 +653,9 @@ class SparkRJob(proto.Message):
648653
archive_uris (MutableSequence[str]):
649654
Optional. HCFS URIs of archives to be
650655
extracted into the working directory of each
651-
executor. Supported file types: .jar, .tar,
652-
.tar.gz, .tgz, and .zip.
656+
executor. Supported file types:
657+
658+
.jar, .tar, .tar.gz, .tgz, and .zip.
653659
properties (MutableMapping[str, str]):
654660
Optional. A mapping of property names to
655661
values, used to configure SparkR. Properties

google/cloud/dataproc_v1/types/workflow_templates.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -437,6 +437,7 @@ class OrderedJob(proto.Message):
437437
class TemplateParameter(proto.Message):
438438
r"""A configurable parameter that replaces one or more fields in
439439
the template. Parameterizable fields:
440+
440441
- Labels
441442
- File uris
442443
- Job properties

0 commit comments

Comments
 (0)
pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy