Skip to content

Commit e333534

Browse files
yoshi-automationtseaver
authored andcommitted
Copy lintified proto files (via synth). (googleapis#7465)
1 parent 88275c2 commit e333534

16 files changed

Lines changed: 247 additions & 179 deletions

dataproc/google/cloud/dataproc_v1/proto/clusters.proto

Lines changed: 54 additions & 37 deletions
Original file line numberDiff line numberDiff line change
@@ -29,28 +29,30 @@ option java_multiple_files = true;
2929
option java_outer_classname = "ClustersProto";
3030
option java_package = "com.google.cloud.dataproc.v1";
3131

32-
3332
// The ClusterControllerService provides methods to manage clusters
3433
// of Compute Engine instances.
3534
service ClusterController {
3635
// Creates a cluster in a project.
37-
rpc CreateCluster(CreateClusterRequest) returns (google.longrunning.Operation) {
36+
rpc CreateCluster(CreateClusterRequest)
37+
returns (google.longrunning.Operation) {
3838
option (google.api.http) = {
3939
post: "/v1/projects/{project_id}/regions/{region}/clusters"
4040
body: "cluster"
4141
};
4242
}
4343

4444
// Updates a cluster in a project.
45-
rpc UpdateCluster(UpdateClusterRequest) returns (google.longrunning.Operation) {
45+
rpc UpdateCluster(UpdateClusterRequest)
46+
returns (google.longrunning.Operation) {
4647
option (google.api.http) = {
4748
patch: "/v1/projects/{project_id}/regions/{region}/clusters/{cluster_name}"
4849
body: "cluster"
4950
};
5051
}
5152

5253
// Deletes a cluster in a project.
53-
rpc DeleteCluster(DeleteClusterRequest) returns (google.longrunning.Operation) {
54+
rpc DeleteCluster(DeleteClusterRequest)
55+
returns (google.longrunning.Operation) {
5456
option (google.api.http) = {
5557
delete: "/v1/projects/{project_id}/regions/{region}/clusters/{cluster_name}"
5658
};
@@ -73,7 +75,8 @@ service ClusterController {
7375
// Gets cluster diagnostic information.
7476
// After the operation completes, the Operation.response field
7577
// contains `DiagnoseClusterOutputLocation`.
76-
rpc DiagnoseCluster(DiagnoseClusterRequest) returns (google.longrunning.Operation) {
78+
rpc DiagnoseCluster(DiagnoseClusterRequest)
79+
returns (google.longrunning.Operation) {
7780
option (google.api.http) = {
7881
post: "/v1/projects/{project_id}/regions/{region}/clusters/{cluster_name}:diagnose"
7982
body: "*"
@@ -99,8 +102,9 @@ message Cluster {
99102
// Label **keys** must contain 1 to 63 characters, and must conform to
100103
// [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
101104
// Label **values** may be empty, but, if present, must contain 1 to 63
102-
// characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
103-
// No more than 32 labels can be associated with a cluster.
105+
// characters, and must conform to [RFC
106+
// 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be
107+
// associated with a cluster.
104108
map<string, string> labels = 8;
105109

106110
// Output only. Cluster status.
@@ -115,8 +119,8 @@ message Cluster {
115119

116120
// Contains cluster daemon metrics such as HDFS and YARN stats.
117121
//
118-
// **Beta Feature**: This report is available for testing purposes only. It may
119-
// be changed before final release.
122+
// **Beta Feature**: This report is available for testing purposes only. It
123+
// may be changed before final release.
120124
ClusterMetrics metrics = 9;
121125
}
122126

@@ -152,9 +156,11 @@ message ClusterConfig {
152156
// Optional. Commands to execute on each node after config is
153157
// completed. By default, executables are run on master and all worker nodes.
154158
// You can test a node's `role` metadata to run an executable on
155-
// a master or worker node, as shown below using `curl` (you can also use `wget`):
159+
// a master or worker node, as shown below using `curl` (you can also use
160+
// `wget`):
156161
//
157-
// ROLE=$(curl -H Metadata-Flavor:Google http://metadata/computeMetadata/v1/instance/attributes/dataproc-role)
162+
// ROLE=$(curl -H Metadata-Flavor:Google
163+
// http://metadata/computeMetadata/v1/instance/attributes/dataproc-role)
158164
// if [[ "${ROLE}" == 'Master' ]]; then
159165
// ... master specific actions ...
160166
// else
@@ -213,11 +219,11 @@ message GceClusterConfig {
213219
string subnetwork_uri = 6;
214220

215221
// Optional. If true, all instances in the cluster will only have internal IP
216-
// addresses. By default, clusters are not restricted to internal IP addresses,
217-
// and will have ephemeral external IP addresses assigned to each instance.
218-
// This `internal_ip_only` restriction can only be enabled for subnetwork
219-
// enabled networks, and all off-cluster dependencies must be configured to be
220-
// accessible without external IP addresses.
222+
// addresses. By default, clusters are not restricted to internal IP
223+
// addresses, and will have ephemeral external IP addresses assigned to each
224+
// instance. This `internal_ip_only` restriction can only be enabled for
225+
// subnetwork enabled networks, and all off-cluster dependencies must be
226+
// configured to be accessible without external IP addresses.
221227
bool internal_ip_only = 7;
222228

223229
// Optional. The service account of the instances. Defaults to the default
@@ -227,7 +233,8 @@ message GceClusterConfig {
227233
// * roles/logging.logWriter
228234
// * roles/storage.objectAdmin
229235
//
230-
// (see https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
236+
// (see
237+
// https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
231238
// for more information).
232239
// Example: `[account_id]@[project_id].iam.gserviceaccount.com`
233240
string service_account = 8;
@@ -253,7 +260,8 @@ message GceClusterConfig {
253260
repeated string tags = 4;
254261

255262
// The Compute Engine metadata entries to add to all instances (see
256-
// [Project and instance metadata](https://cloud.google.com/compute/docs/storing-retrieving-metadata#project_and_instance_metadata)).
263+
// [Project and instance
264+
// metadata](https://cloud.google.com/compute/docs/storing-retrieving-metadata#project_and_instance_metadata)).
257265
map<string, string> metadata = 5;
258266
}
259267

@@ -282,15 +290,17 @@ message InstanceGroupConfig {
282290
// * `n1-standard-2`
283291
//
284292
// **Auto Zone Exception**: If you are using the Cloud Dataproc
285-
// [Auto Zone Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
293+
// [Auto Zone
294+
// Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
286295
// feature, you must use the short name of the machine type
287296
// resource, for example, `n1-standard-2`.
288297
string machine_type_uri = 4;
289298

290299
// Optional. Disk option config settings.
291300
DiskConfig disk_config = 5;
292301

293-
// Optional. Specifies that this instance group contains preemptible instances.
302+
// Optional. Specifies that this instance group contains preemptible
303+
// instances.
294304
bool is_preemptible = 6;
295305

296306
// Output only. The config for Compute Engine Instance Group
@@ -321,7 +331,8 @@ message ManagedGroupConfig {
321331
message AcceleratorConfig {
322332
// Full URL, partial URI, or short name of the accelerator type resource to
323333
// expose to this instance. See
324-
// [Compute Engine AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes).
334+
// [Compute Engine
335+
// AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes).
325336
//
326337
// Examples:
327338
//
@@ -330,7 +341,8 @@ message AcceleratorConfig {
330341
// * `nvidia-tesla-k80`
331342
//
332343
// **Auto Zone Exception**: If you are using the Cloud Dataproc
333-
// [Auto Zone Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
344+
// [Auto Zone
345+
// Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
334346
// feature, you must use the short name of the accelerator type
335347
// resource, for example, `nvidia-tesla-k80`.
336348
string accelerator_type_uri = 1;
@@ -429,10 +441,12 @@ message ClusterStatus {
429441

430442
// Specifies the selection and config of software inside the cluster.
431443
message SoftwareConfig {
432-
// Optional. The version of software inside the cluster. It must be one of the supported
433-
// [Cloud Dataproc Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
444+
// Optional. The version of software inside the cluster. It must be one of the
445+
// supported [Cloud Dataproc
446+
// Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
434447
// such as "1.2" (including a subminor version, such as "1.2.29"), or the
435-
// ["preview" version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
448+
// ["preview"
449+
// version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
436450
// If unspecified, it defaults to the latest version.
437451
string image_version = 1;
438452

@@ -482,10 +496,11 @@ message CreateClusterRequest {
482496
Cluster cluster = 2;
483497

484498
// Optional. A unique id used to identify the request. If the server
485-
// receives two [CreateClusterRequest][google.cloud.dataproc.v1.CreateClusterRequest] requests with the same
486-
// id, then the second request will be ignored and the
487-
// first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the backend
488-
// is returned.
499+
// receives two
500+
// [CreateClusterRequest][google.cloud.dataproc.v1.CreateClusterRequest]
501+
// requests with the same id, then the second request will be ignored and the
502+
// first [google.longrunning.Operation][google.longrunning.Operation] created
503+
// and stored in the backend is returned.
489504
//
490505
// It is recommended to always set this value to a
491506
// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
@@ -570,10 +585,11 @@ message UpdateClusterRequest {
570585
google.protobuf.FieldMask update_mask = 4;
571586

572587
// Optional. A unique id used to identify the request. If the server
573-
// receives two [UpdateClusterRequest][google.cloud.dataproc.v1.UpdateClusterRequest] requests with the same
574-
// id, then the second request will be ignored and the
575-
// first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the
576-
// backend is returned.
588+
// receives two
589+
// [UpdateClusterRequest][google.cloud.dataproc.v1.UpdateClusterRequest]
590+
// requests with the same id, then the second request will be ignored and the
591+
// first [google.longrunning.Operation][google.longrunning.Operation] created
592+
// and stored in the backend is returned.
577593
//
578594
// It is recommended to always set this value to a
579595
// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).
@@ -600,10 +616,11 @@ message DeleteClusterRequest {
600616
string cluster_uuid = 4;
601617

602618
// Optional. A unique id used to identify the request. If the server
603-
// receives two [DeleteClusterRequest][google.cloud.dataproc.v1.DeleteClusterRequest] requests with the same
604-
// id, then the second request will be ignored and the
605-
// first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the
606-
// backend is returned.
619+
// receives two
620+
// [DeleteClusterRequest][google.cloud.dataproc.v1.DeleteClusterRequest]
621+
// requests with the same id, then the second request will be ignored and the
622+
// first [google.longrunning.Operation][google.longrunning.Operation] created
623+
// and stored in the backend is returned.
607624
//
608625
// It is recommended to always set this value to a
609626
// [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier).

dataproc/google/cloud/dataproc_v1/proto/clusters_pb2.py

Lines changed: 4 additions & 4 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

dataproc/google/cloud/dataproc_v1/proto/jobs.proto

Lines changed: 25 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,6 @@ option java_multiple_files = true;
2727
option java_outer_classname = "JobsProto";
2828
option java_package = "com.google.cloud.dataproc.v1";
2929

30-
3130
// The JobController provides methods to manage jobs.
3231
service JobController {
3332
// Submits a job to a cluster.
@@ -62,7 +61,8 @@ service JobController {
6261

6362
// Starts a job cancellation request. To access the job resource
6463
// after cancellation, call
65-
// [regions/{region}/jobs.list](/dataproc/docs/reference/rest/v1/projects.regions.jobs/list) or
64+
// [regions/{region}/jobs.list](/dataproc/docs/reference/rest/v1/projects.regions.jobs/list)
65+
// or
6666
// [regions/{region}/jobs.get](/dataproc/docs/reference/rest/v1/projects.regions.jobs/get).
6767
rpc CancelJob(CancelJobRequest) returns (Job) {
6868
option (google.api.http) = {
@@ -122,8 +122,10 @@ message LoggingConfig {
122122
}
123123

124124
// A Cloud Dataproc job for running
125-
// [Apache Hadoop MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html)
126-
// jobs on [Apache Hadoop YARN](https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html).
125+
// [Apache Hadoop
126+
// MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html)
127+
// jobs on [Apache Hadoop
128+
// YARN](https://hadoop.apache.org/docs/r2.7.1/hadoop-yarn/hadoop-yarn-site/YARN.html).
127129
message HadoopJob {
128130
// Required. Indicates the location of the driver's main class. Specify
129131
// either the jar file that contains the main class or the main class name.
@@ -143,8 +145,8 @@ message HadoopJob {
143145
}
144146

145147
// Optional. The arguments to pass to the driver. Do not
146-
// include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as job
147-
// properties, since a collision may occur that causes an incorrect job
148+
// include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as
149+
// job properties, since a collision may occur that causes an incorrect job
148150
// submission.
149151
repeated string args = 3;
150152

@@ -178,7 +180,8 @@ message SparkJob {
178180
// Required. The specification of the main method to call to drive the job.
179181
// Specify either the jar file that contains the main class or the main class
180182
// name. To pass both a main jar and a main class in that jar, add the jar to
181-
// `CommonJob.jar_file_uris`, and then specify the main class name in `main_class`.
183+
// `CommonJob.jar_file_uris`, and then specify the main class name in
184+
// `main_class`.
182185
oneof driver {
183186
// The HCFS URI of the jar file that contains the main class.
184187
string main_jar_file_uri = 1;
@@ -217,7 +220,8 @@ message SparkJob {
217220
}
218221

219222
// A Cloud Dataproc job for running
220-
// [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
223+
// [Apache
224+
// PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
221225
// applications on YARN.
222226
message PySparkJob {
223227
// Required. The HCFS URI of the main Python file to use as the driver. Must
@@ -288,8 +292,8 @@ message HiveJob {
288292
}
289293

290294
// Optional. Whether to continue executing queries if a query fails.
291-
// The default value is `false`. Setting to `true` can be useful when executing
292-
// independent parallel queries.
295+
// The default value is `false`. Setting to `true` can be useful when
296+
// executing independent parallel queries.
293297
bool continue_on_failure = 3;
294298

295299
// Optional. Mapping of query variable names to values (equivalent to the
@@ -308,8 +312,8 @@ message HiveJob {
308312
repeated string jar_file_uris = 6;
309313
}
310314

311-
// A Cloud Dataproc job for running [Apache Spark SQL](http://spark.apache.org/sql/)
312-
// queries.
315+
// A Cloud Dataproc job for running [Apache Spark
316+
// SQL](http://spark.apache.org/sql/) queries.
313317
message SparkSqlJob {
314318
// Required. The sequence of Spark SQL queries to execute, specified as
315319
// either an HCFS file URI or as a list of queries.
@@ -351,8 +355,8 @@ message PigJob {
351355
}
352356

353357
// Optional. Whether to continue executing queries if a query fails.
354-
// The default value is `false`. Setting to `true` can be useful when executing
355-
// independent parallel queries.
358+
// The default value is `false`. Setting to `true` can be useful when
359+
// executing independent parallel queries.
356360
bool continue_on_failure = 3;
357361

358362
// Optional. Mapping of query variable names to values (equivalent to the Pig
@@ -573,8 +577,8 @@ message Job {
573577

574578
// Output only. The collection of YARN applications spun up by this job.
575579
//
576-
// **Beta** Feature: This report is available for testing purposes only. It may
577-
// be changed before final release.
580+
// **Beta** Feature: This report is available for testing purposes only. It
581+
// may be changed before final release.
578582
repeated YarnApplication yarn_applications = 9;
579583

580584
// Output only. A URI pointing to the location of the stdout of the job's
@@ -590,8 +594,9 @@ message Job {
590594
// Label **keys** must contain 1 to 63 characters, and must conform to
591595
// [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
592596
// Label **values** may be empty, but, if present, must contain 1 to 63
593-
// characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt).
594-
// No more than 32 labels can be associated with a job.
597+
// characters, and must conform to [RFC
598+
// 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be
599+
// associated with a job.
595600
map<string, string> labels = 18;
596601

597602
// Optional. Job scheduling configuration.
@@ -629,8 +634,8 @@ message SubmitJobRequest {
629634
Job job = 2;
630635

631636
// Optional. A unique id used to identify the request. If the server
632-
// receives two [SubmitJobRequest][google.cloud.dataproc.v1.SubmitJobRequest] requests with the same
633-
// id, then the second request will be ignored and the
637+
// receives two [SubmitJobRequest][google.cloud.dataproc.v1.SubmitJobRequest]
638+
// requests with the same id, then the second request will be ignored and the
634639
// first [Job][google.cloud.dataproc.v1.Job] created and stored in the backend
635640
// is returned.
636641
//

dataproc/google/cloud/dataproc_v1/proto/jobs_pb2_grpc.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,8 @@ def UpdateJob(self, request, context):
8484
def CancelJob(self, request, context):
8585
"""Starts a job cancellation request. To access the job resource
8686
after cancellation, call
87-
[regions/{region}/jobs.list](/dataproc/docs/reference/rest/v1/projects.regions.jobs/list) or
87+
[regions/{region}/jobs.list](/dataproc/docs/reference/rest/v1/projects.regions.jobs/list)
88+
or
8889
[regions/{region}/jobs.get](/dataproc/docs/reference/rest/v1/projects.regions.jobs/get).
8990
"""
9091
context.set_code(grpc.StatusCode.UNIMPLEMENTED)

dataproc/google/cloud/dataproc_v1/proto/operations.proto

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,6 @@ option java_multiple_files = true;
2525
option java_outer_classname = "OperationsProto";
2626
option java_package = "com.google.cloud.dataproc.v1";
2727

28-
2928
// The status of the operation.
3029
message ClusterOperationStatus {
3130
// The operation state.

0 commit comments

Comments
 (0)