From f174b992fa65dbade37dc1df210b477d44af6758 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Bot Date: Thu, 19 Dec 2019 09:15:34 -0800 Subject: [PATCH] docs(regen): updated javadocs from protos (#40) --- .../dataproc/v1/ClusterControllerClient.java | 14 +- .../dataproc/v1/JobControllerClient.java | 12 +- .../v1/WorkflowTemplateServiceClient.java | 2 +- .../cloud/dataproc/v1/package-info.java | 3 +- .../v1/AutoscalingPolicyServiceGrpc.java | 10 +- .../v1/WorkflowTemplateServiceGrpc.java | 10 +- .../cloud/dataproc/v1/AcceleratorConfig.java | 14 +- .../v1/AcceleratorConfigOrBuilder.java | 4 +- .../cloud/dataproc/v1/AutoscalingConfig.java | 14 +- .../v1/AutoscalingConfigOrBuilder.java | 4 +- .../cloud/dataproc/v1/CancelJobRequest.java | 14 +- .../v1/CancelJobRequestOrBuilder.java | 4 +- .../com/google/cloud/dataproc/v1/Cluster.java | 114 +- .../cloud/dataproc/v1/ClusterConfig.java | 42 +- .../dataproc/v1/ClusterConfigOrBuilder.java | 12 +- .../cloud/dataproc/v1/ClusterOrBuilder.java | 30 +- .../cloud/dataproc/v1/ClusterStatus.java | 4 +- .../cloud/dataproc/v1/ClustersProto.java | 341 ++-- .../dataproc/v1/CreateClusterRequest.java | 14 +- .../v1/CreateClusterRequestOrBuilder.java | 4 +- .../dataproc/v1/DeleteClusterRequest.java | 14 +- .../v1/DeleteClusterRequestOrBuilder.java | 4 +- .../cloud/dataproc/v1/DeleteJobRequest.java | 14 +- .../v1/DeleteJobRequestOrBuilder.java | 4 +- .../dataproc/v1/DiagnoseClusterRequest.java | 14 +- .../v1/DiagnoseClusterRequestOrBuilder.java | 4 +- .../cloud/dataproc/v1/GceClusterConfig.java | 147 +- .../v1/GceClusterConfigOrBuilder.java | 42 +- .../cloud/dataproc/v1/GetClusterRequest.java | 14 +- .../v1/GetClusterRequestOrBuilder.java | 4 +- .../cloud/dataproc/v1/GetJobRequest.java | 14 +- .../dataproc/v1/GetJobRequestOrBuilder.java | 4 +- .../google/cloud/dataproc/v1/HadoopJob.java | 26 +- .../cloud/dataproc/v1/HadoopJobOrBuilder.java | 10 +- .../com/google/cloud/dataproc/v1/HiveJob.java | 26 +- .../cloud/dataproc/v1/HiveJobOrBuilder.java | 10 +- .../dataproc/v1/InstanceGroupConfig.java | 54 +- .../v1/InstanceGroupConfigOrBuilder.java | 16 +- .../com/google/cloud/dataproc/v1/Job.java | 4 +- .../cloud/dataproc/v1/JobPlacement.java | 18 +- .../dataproc/v1/JobPlacementOrBuilder.java | 4 +- .../google/cloud/dataproc/v1/JobStatus.java | 8 +- .../dataproc/v1/ListClustersRequest.java | 14 +- .../v1/ListClustersRequestOrBuilder.java | 4 +- .../cloud/dataproc/v1/ListJobsRequest.java | 14 +- .../dataproc/v1/ListJobsRequestOrBuilder.java | 4 +- .../com/google/cloud/dataproc/v1/PigJob.java | 26 +- .../cloud/dataproc/v1/PigJobOrBuilder.java | 10 +- .../google/cloud/dataproc/v1/PySparkJob.java | 26 +- .../dataproc/v1/PySparkJobOrBuilder.java | 10 +- .../cloud/dataproc/v1/SoftwareConfig.java | 14 +- .../dataproc/v1/SoftwareConfigOrBuilder.java | 4 +- .../google/cloud/dataproc/v1/SparkJob.java | 26 +- .../cloud/dataproc/v1/SparkJobOrBuilder.java | 10 +- .../google/cloud/dataproc/v1/SparkSqlJob.java | 26 +- .../dataproc/v1/SparkSqlJobOrBuilder.java | 10 +- .../cloud/dataproc/v1/SubmitJobRequest.java | 14 +- .../v1/SubmitJobRequestOrBuilder.java | 4 +- .../dataproc/v1/UpdateClusterRequest.java | 14 +- .../v1/UpdateClusterRequestOrBuilder.java | 4 +- .../cloud/dataproc/v1/UpdateJobRequest.java | 14 +- .../v1/UpdateJobRequestOrBuilder.java | 4 +- .../v1/UpdateWorkflowTemplateRequest.java | 24 +- ...pdateWorkflowTemplateRequestOrBuilder.java | 6 +- .../cloud/dataproc/v1/WorkflowMetadata.java | 4 +- .../cloud/dataproc/v1/WorkflowTemplate.java | 4 +- .../dataproc/v1/WorkflowTemplatesProto.java | 139 +- .../dataproc/v1/autoscaling_policies.proto | 2 +- .../google/cloud/dataproc/v1/clusters.proto | 67 +- .../proto/google/cloud/dataproc/v1/jobs.proto | 47 +- .../dataproc/v1/workflow_templates.proto | 11 +- .../cloud/dataproc/v1beta2/ClustersProto.java | 89 +- .../cloud/dataproc/v1beta2/JobsProto.java | 69 +- .../UpdateWorkflowTemplateRequest.java | 24 +- ...pdateWorkflowTemplateRequestOrBuilder.java | 6 +- .../v1beta2/WorkflowTemplatesProto.java | 145 +- .../cloud/dataproc/v1beta2/clusters.proto | 12 +- .../google/cloud/dataproc/v1beta2/jobs.proto | 12 +- .../dataproc/v1beta2/workflow_templates.proto | 5 +- synth.metadata | 1563 ++++++++++++++++- 80 files changed, 2592 insertions(+), 1010 deletions(-) diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java index 827c1e75..277f0cc8 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java @@ -195,7 +195,7 @@ public final OperationsClient getOperationsClient() { * * @param projectId Required. The ID of the Google Cloud Platform project that the cluster belongs * to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param cluster Required. The cluster to create. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ @@ -323,7 +323,7 @@ public final UnaryCallable createClusterCallabl * * * @param projectId Required. The ID of the Google Cloud Platform project the cluster belongs to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param clusterName Required. The cluster name. * @param cluster Required. The changes to the cluster. * @param updateMask Required. Specifies the path, relative to `Cluster`, of the field to update. @@ -486,7 +486,7 @@ public final UnaryCallable updateClusterCallabl * * @param projectId Required. The ID of the Google Cloud Platform project that the cluster belongs * to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param clusterName Required. The cluster name. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ @@ -611,7 +611,7 @@ public final UnaryCallable deleteClusterCallabl * * @param projectId Required. The ID of the Google Cloud Platform project that the cluster belongs * to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param clusterName Required. The cluster name. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ @@ -697,7 +697,7 @@ public final UnaryCallable getClusterCallable() { * * @param projectId Required. The ID of the Google Cloud Platform project that the cluster belongs * to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ public final ListClustersPagedResponse listClusters(String projectId, String region) { @@ -725,7 +725,7 @@ public final ListClustersPagedResponse listClusters(String projectId, String reg * * @param projectId Required. The ID of the Google Cloud Platform project that the cluster belongs * to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param filter Optional. A filter constraining the clusters to list. Filters are case-sensitive * and have the following syntax: *

field = value [AND [field = value]] ... @@ -862,7 +862,7 @@ public final UnaryCallable listCluste * * @param projectId Required. The ID of the Google Cloud Platform project that the cluster belongs * to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param clusterName Required. The cluster name. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java index d3dd62db..81fe3653 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java @@ -172,7 +172,7 @@ public JobControllerStub getStub() { * * * @param projectId Required. The ID of the Google Cloud Platform project that the job belongs to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param job Required. The job resource. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ @@ -252,7 +252,7 @@ public final UnaryCallable submitJobCallable() { * * * @param projectId Required. The ID of the Google Cloud Platform project that the job belongs to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param jobId Required. The job ID. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ @@ -337,7 +337,7 @@ public final UnaryCallable getJobCallable() { * * * @param projectId Required. The ID of the Google Cloud Platform project that the job belongs to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ public final ListJobsPagedResponse listJobs(String projectId, String region) { @@ -364,7 +364,7 @@ public final ListJobsPagedResponse listJobs(String projectId, String region) { * * * @param projectId Required. The ID of the Google Cloud Platform project that the job belongs to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param filter Optional. A filter constraining the jobs to list. Filters are case-sensitive and * have the following syntax: *

[field = value] AND [field [= value]] ... @@ -551,7 +551,7 @@ public final UnaryCallable updateJobCallable() { * * * @param projectId Required. The ID of the Google Cloud Platform project that the job belongs to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param jobId Required. The job ID. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ @@ -640,7 +640,7 @@ public final UnaryCallable cancelJobCallable() { * * * @param projectId Required. The ID of the Google Cloud Platform project that the job belongs to. - * @param region Required. The Cloud Dataproc region in which to handle the request. + * @param region Required. The Dataproc region in which to handle the request. * @param jobId Required. The job ID. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceClient.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceClient.java index 1ace8b7a..7c204be9 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceClient.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceClient.java @@ -41,7 +41,7 @@ // AUTO-GENERATED DOCUMENTATION AND SERVICE /** - * Service Description: The API interface for managing Workflow Templates in the Cloud Dataproc API. + * Service Description: The API interface for managing Workflow Templates in the Dataproc API. * *

This class provides the ability to make remote calls to the backing service through method * calls that map to API methods. Sample code to get started: diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/package-info.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/package-info.java index 2c253e93..0020602d 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/package-info.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/package-info.java @@ -56,8 +56,7 @@ * * ============================= WorkflowTemplateServiceClient ============================= * - *

Service Description: The API interface for managing Workflow Templates in the Cloud Dataproc - * API. + *

Service Description: The API interface for managing Workflow Templates in the Dataproc API. * *

Sample for WorkflowTemplateServiceClient: * diff --git a/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceGrpc.java b/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceGrpc.java index c00f9577..86e05c89 100644 --- a/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceGrpc.java +++ b/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceGrpc.java @@ -27,7 +27,7 @@ * *

  * The API interface for managing autoscaling policies in the
- * Google Cloud Dataproc API.
+ * Dataproc API.
  * 
*/ @javax.annotation.Generated( @@ -377,7 +377,7 @@ public static AutoscalingPolicyServiceFutureStub newFutureStub(io.grpc.Channel c * *
    * The API interface for managing autoscaling policies in the
-   * Google Cloud Dataproc API.
+   * Dataproc API.
    * 
*/ public abstract static class AutoscalingPolicyServiceImplBase implements io.grpc.BindableService { @@ -500,7 +500,7 @@ public final io.grpc.ServerServiceDefinition bindService() { * *
    * The API interface for managing autoscaling policies in the
-   * Google Cloud Dataproc API.
+   * Dataproc API.
    * 
*/ public static final class AutoscalingPolicyServiceStub @@ -612,7 +612,7 @@ public void deleteAutoscalingPolicy( * *
    * The API interface for managing autoscaling policies in the
-   * Google Cloud Dataproc API.
+   * Dataproc API.
    * 
*/ public static final class AutoscalingPolicyServiceBlockingStub @@ -706,7 +706,7 @@ public com.google.protobuf.Empty deleteAutoscalingPolicy( * *
    * The API interface for managing autoscaling policies in the
-   * Google Cloud Dataproc API.
+   * Dataproc API.
    * 
*/ public static final class AutoscalingPolicyServiceFutureStub diff --git a/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceGrpc.java b/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceGrpc.java index c8e9060d..47d8ee77 100644 --- a/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceGrpc.java +++ b/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceGrpc.java @@ -27,7 +27,7 @@ * *
  * The API interface for managing Workflow Templates in the
- * Cloud Dataproc API.
+ * Dataproc API.
  * 
*/ @javax.annotation.Generated( @@ -506,7 +506,7 @@ public static WorkflowTemplateServiceFutureStub newFutureStub(io.grpc.Channel ch * *
    * The API interface for managing Workflow Templates in the
-   * Cloud Dataproc API.
+   * Dataproc API.
    * 
*/ public abstract static class WorkflowTemplateServiceImplBase implements io.grpc.BindableService { @@ -703,7 +703,7 @@ public final io.grpc.ServerServiceDefinition bindService() { * *
    * The API interface for managing Workflow Templates in the
-   * Cloud Dataproc API.
+   * Dataproc API.
    * 
*/ public static final class WorkflowTemplateServiceStub @@ -881,7 +881,7 @@ public void deleteWorkflowTemplate( * *
    * The API interface for managing Workflow Templates in the
-   * Cloud Dataproc API.
+   * Dataproc API.
    * 
*/ public static final class WorkflowTemplateServiceBlockingStub @@ -1037,7 +1037,7 @@ public com.google.protobuf.Empty deleteWorkflowTemplate( * *
    * The API interface for managing Workflow Templates in the
-   * Cloud Dataproc API.
+   * Dataproc API.
    * 
*/ public static final class WorkflowTemplateServiceFutureStub diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java index cc0b2316..bd382ba7 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java @@ -131,7 +131,7 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type @@ -165,7 +165,7 @@ public java.lang.String getAcceleratorTypeUri() { * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type @@ -550,7 +550,7 @@ public Builder mergeFrom( * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type @@ -584,7 +584,7 @@ public java.lang.String getAcceleratorTypeUri() { * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type @@ -618,7 +618,7 @@ public com.google.protobuf.ByteString getAcceleratorTypeUriBytes() { * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type @@ -651,7 +651,7 @@ public Builder setAcceleratorTypeUri(java.lang.String value) { * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type @@ -680,7 +680,7 @@ public Builder clearAcceleratorTypeUri() { * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java index 46f8e7f2..f1e41845 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java @@ -35,7 +35,7 @@ public interface AcceleratorConfigOrBuilder * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type @@ -59,7 +59,7 @@ public interface AcceleratorConfigOrBuilder * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfig.java index bca3054c..6ecee42e 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfig.java @@ -122,7 +122,7 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -149,7 +149,7 @@ public java.lang.String getPolicyUri() { * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -494,7 +494,7 @@ public Builder mergeFrom( * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -521,7 +521,7 @@ public java.lang.String getPolicyUri() { * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -548,7 +548,7 @@ public com.google.protobuf.ByteString getPolicyUriBytes() { * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -574,7 +574,7 @@ public Builder setPolicyUri(java.lang.String value) { * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -596,7 +596,7 @@ public Builder clearPolicyUri() { * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfigOrBuilder.java index 22fb59af..7c393bc1 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfigOrBuilder.java @@ -32,7 +32,7 @@ public interface AutoscalingConfigOrBuilder * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -49,7 +49,7 @@ public interface AutoscalingConfigOrBuilder * Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` * * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]` - * Note that the policy must be in the same project and Cloud Dataproc region. + * Note that the policy must be in the same project and Dataproc region. * * * string policy_uri = 1 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequest.java index 807ad6f4..468402f8 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequest.java @@ -182,7 +182,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -204,7 +204,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -734,7 +734,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -756,7 +756,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -778,7 +778,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -799,7 +799,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -816,7 +816,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequestOrBuilder.java index 791ad4c6..6e3dddb8 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface CancelJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface CancelJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Cluster.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Cluster.java index e87b6cb3..129aafa4 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Cluster.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Cluster.java @@ -226,7 +226,7 @@ protected com.google.protobuf.MapField internalGetMapField(int number) { * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The projectId. */ @@ -248,7 +248,7 @@ public java.lang.String getProjectId() { * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for projectId. */ @@ -274,7 +274,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @return The clusterName. */ @@ -297,7 +297,7 @@ public java.lang.String getClusterName() { * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for clusterName. */ @@ -319,11 +319,13 @@ public com.google.protobuf.ByteString getClusterNameBytes() { * * *
-   * Required. The cluster config. Note that Cloud Dataproc may set
+   * Required. The cluster config. Note that Dataproc may set
    * default values, and values may change when clusters are updated.
    * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * * * @return Whether the config field is set. */ @@ -334,11 +336,13 @@ public boolean hasConfig() { * * *
-   * Required. The cluster config. Note that Cloud Dataproc may set
+   * Required. The cluster config. Note that Dataproc may set
    * default values, and values may change when clusters are updated.
    * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * * * @return The config. */ @@ -351,11 +355,13 @@ public com.google.cloud.dataproc.v1.ClusterConfig getConfig() { * * *
-   * Required. The cluster config. Note that Cloud Dataproc may set
+   * Required. The cluster config. Note that Dataproc may set
    * default values, and values may change when clusters are updated.
    * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ public com.google.cloud.dataproc.v1.ClusterConfigOrBuilder getConfigOrBuilder() { return getConfig(); @@ -609,7 +615,7 @@ public com.google.cloud.dataproc.v1.ClusterStatusOrBuilder getStatusHistoryOrBui * * *
-   * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+   * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
    * generates this value when it creates the cluster.
    * 
* @@ -632,7 +638,7 @@ public java.lang.String getClusterUuid() { * * *
-   * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+   * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
    * generates this value when it creates the cluster.
    * 
* @@ -1240,7 +1246,7 @@ public Builder mergeFrom( * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The projectId. */ @@ -1262,7 +1268,7 @@ public java.lang.String getProjectId() { * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for projectId. */ @@ -1284,7 +1290,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @param value The projectId to set. * @return This builder for chaining. @@ -1305,7 +1311,7 @@ public Builder setProjectId(java.lang.String value) { * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return This builder for chaining. */ @@ -1322,7 +1328,7 @@ public Builder clearProjectId() { * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @param value The bytes for projectId to set. * @return This builder for chaining. @@ -1347,7 +1353,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @return The clusterName. */ @@ -1370,7 +1376,7 @@ public java.lang.String getClusterName() { * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for clusterName. */ @@ -1393,7 +1399,7 @@ public com.google.protobuf.ByteString getClusterNameBytes() { * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @param value The clusterName to set. * @return This builder for chaining. @@ -1415,7 +1421,7 @@ public Builder setClusterName(java.lang.String value) { * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @return This builder for chaining. */ @@ -1433,7 +1439,7 @@ public Builder clearClusterName() { * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @param value The bytes for clusterName to set. * @return This builder for chaining. @@ -1459,11 +1465,13 @@ public Builder setClusterNameBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * * * @return Whether the config field is set. */ @@ -1474,11 +1482,13 @@ public boolean hasConfig() { * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * * * @return The config. */ @@ -1495,11 +1505,13 @@ public com.google.cloud.dataproc.v1.ClusterConfig getConfig() { * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ public Builder setConfig(com.google.cloud.dataproc.v1.ClusterConfig value) { if (configBuilder_ == null) { @@ -1518,11 +1530,13 @@ public Builder setConfig(com.google.cloud.dataproc.v1.ClusterConfig value) { * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ public Builder setConfig(com.google.cloud.dataproc.v1.ClusterConfig.Builder builderForValue) { if (configBuilder_ == null) { @@ -1538,11 +1552,13 @@ public Builder setConfig(com.google.cloud.dataproc.v1.ClusterConfig.Builder buil * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ public Builder mergeConfig(com.google.cloud.dataproc.v1.ClusterConfig value) { if (configBuilder_ == null) { @@ -1565,11 +1581,13 @@ public Builder mergeConfig(com.google.cloud.dataproc.v1.ClusterConfig value) { * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ public Builder clearConfig() { if (configBuilder_ == null) { @@ -1586,11 +1604,13 @@ public Builder clearConfig() { * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ public com.google.cloud.dataproc.v1.ClusterConfig.Builder getConfigBuilder() { @@ -1601,11 +1621,13 @@ public com.google.cloud.dataproc.v1.ClusterConfig.Builder getConfigBuilder() { * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ public com.google.cloud.dataproc.v1.ClusterConfigOrBuilder getConfigOrBuilder() { if (configBuilder_ != null) { @@ -1620,11 +1642,13 @@ public com.google.cloud.dataproc.v1.ClusterConfigOrBuilder getConfigOrBuilder() * * *
-     * Required. The cluster config. Note that Cloud Dataproc may set
+     * Required. The cluster config. Note that Dataproc may set
      * default values, and values may change when clusters are updated.
      * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1.ClusterConfig, @@ -2436,7 +2460,7 @@ public com.google.cloud.dataproc.v1.ClusterStatus.Builder addStatusHistoryBuilde * * *
-     * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+     * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
      * generates this value when it creates the cluster.
      * 
* @@ -2459,7 +2483,7 @@ public java.lang.String getClusterUuid() { * * *
-     * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+     * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
      * generates this value when it creates the cluster.
      * 
* @@ -2482,7 +2506,7 @@ public com.google.protobuf.ByteString getClusterUuidBytes() { * * *
-     * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+     * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
      * generates this value when it creates the cluster.
      * 
* @@ -2504,7 +2528,7 @@ public Builder setClusterUuid(java.lang.String value) { * * *
-     * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+     * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
      * generates this value when it creates the cluster.
      * 
* @@ -2522,7 +2546,7 @@ public Builder clearClusterUuid() { * * *
-     * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+     * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
      * generates this value when it creates the cluster.
      * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java index 1ef1ddab..3d970374 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java @@ -264,14 +264,14 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * * *
-   * Optional. A Google Cloud Storage bucket used to stage job
+   * Optional. A Cloud Storage bucket used to stage job
    * dependencies, config files, and job driver console output.
    * If you do not specify a staging bucket, Cloud
    * Dataproc will determine a Cloud Storage location (US,
-   * ASIA, or EU) for your cluster's staging bucket according to the Google
+   * ASIA, or EU) for your cluster's staging bucket according to the
    * Compute Engine zone where your cluster is deployed, and then create
    * and manage this project-level, per-location bucket (see
-   * [Cloud Dataproc staging
+   * [Dataproc staging
    * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
    * 
* @@ -294,14 +294,14 @@ public java.lang.String getConfigBucket() { * * *
-   * Optional. A Google Cloud Storage bucket used to stage job
+   * Optional. A Cloud Storage bucket used to stage job
    * dependencies, config files, and job driver console output.
    * If you do not specify a staging bucket, Cloud
    * Dataproc will determine a Cloud Storage location (US,
-   * ASIA, or EU) for your cluster's staging bucket according to the Google
+   * ASIA, or EU) for your cluster's staging bucket according to the
    * Compute Engine zone where your cluster is deployed, and then create
    * and manage this project-level, per-location bucket (see
-   * [Cloud Dataproc staging
+   * [Dataproc staging
    * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
    * 
* @@ -1483,14 +1483,14 @@ public Builder mergeFrom( * * *
-     * Optional. A Google Cloud Storage bucket used to stage job
+     * Optional. A Cloud Storage bucket used to stage job
      * dependencies, config files, and job driver console output.
      * If you do not specify a staging bucket, Cloud
      * Dataproc will determine a Cloud Storage location (US,
-     * ASIA, or EU) for your cluster's staging bucket according to the Google
+     * ASIA, or EU) for your cluster's staging bucket according to the
      * Compute Engine zone where your cluster is deployed, and then create
      * and manage this project-level, per-location bucket (see
-     * [Cloud Dataproc staging
+     * [Dataproc staging
      * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
      * 
* @@ -1513,14 +1513,14 @@ public java.lang.String getConfigBucket() { * * *
-     * Optional. A Google Cloud Storage bucket used to stage job
+     * Optional. A Cloud Storage bucket used to stage job
      * dependencies, config files, and job driver console output.
      * If you do not specify a staging bucket, Cloud
      * Dataproc will determine a Cloud Storage location (US,
-     * ASIA, or EU) for your cluster's staging bucket according to the Google
+     * ASIA, or EU) for your cluster's staging bucket according to the
      * Compute Engine zone where your cluster is deployed, and then create
      * and manage this project-level, per-location bucket (see
-     * [Cloud Dataproc staging
+     * [Dataproc staging
      * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
      * 
* @@ -1543,14 +1543,14 @@ public com.google.protobuf.ByteString getConfigBucketBytes() { * * *
-     * Optional. A Google Cloud Storage bucket used to stage job
+     * Optional. A Cloud Storage bucket used to stage job
      * dependencies, config files, and job driver console output.
      * If you do not specify a staging bucket, Cloud
      * Dataproc will determine a Cloud Storage location (US,
-     * ASIA, or EU) for your cluster's staging bucket according to the Google
+     * ASIA, or EU) for your cluster's staging bucket according to the
      * Compute Engine zone where your cluster is deployed, and then create
      * and manage this project-level, per-location bucket (see
-     * [Cloud Dataproc staging
+     * [Dataproc staging
      * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
      * 
* @@ -1572,14 +1572,14 @@ public Builder setConfigBucket(java.lang.String value) { * * *
-     * Optional. A Google Cloud Storage bucket used to stage job
+     * Optional. A Cloud Storage bucket used to stage job
      * dependencies, config files, and job driver console output.
      * If you do not specify a staging bucket, Cloud
      * Dataproc will determine a Cloud Storage location (US,
-     * ASIA, or EU) for your cluster's staging bucket according to the Google
+     * ASIA, or EU) for your cluster's staging bucket according to the
      * Compute Engine zone where your cluster is deployed, and then create
      * and manage this project-level, per-location bucket (see
-     * [Cloud Dataproc staging
+     * [Dataproc staging
      * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
      * 
* @@ -1597,14 +1597,14 @@ public Builder clearConfigBucket() { * * *
-     * Optional. A Google Cloud Storage bucket used to stage job
+     * Optional. A Cloud Storage bucket used to stage job
      * dependencies, config files, and job driver console output.
      * If you do not specify a staging bucket, Cloud
      * Dataproc will determine a Cloud Storage location (US,
-     * ASIA, or EU) for your cluster's staging bucket according to the Google
+     * ASIA, or EU) for your cluster's staging bucket according to the
      * Compute Engine zone where your cluster is deployed, and then create
      * and manage this project-level, per-location bucket (see
-     * [Cloud Dataproc staging
+     * [Dataproc staging
      * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
      * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java index 2bea5278..e7c8cbba 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java @@ -27,14 +27,14 @@ public interface ClusterConfigOrBuilder * * *
-   * Optional. A Google Cloud Storage bucket used to stage job
+   * Optional. A Cloud Storage bucket used to stage job
    * dependencies, config files, and job driver console output.
    * If you do not specify a staging bucket, Cloud
    * Dataproc will determine a Cloud Storage location (US,
-   * ASIA, or EU) for your cluster's staging bucket according to the Google
+   * ASIA, or EU) for your cluster's staging bucket according to the
    * Compute Engine zone where your cluster is deployed, and then create
    * and manage this project-level, per-location bucket (see
-   * [Cloud Dataproc staging
+   * [Dataproc staging
    * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
    * 
* @@ -47,14 +47,14 @@ public interface ClusterConfigOrBuilder * * *
-   * Optional. A Google Cloud Storage bucket used to stage job
+   * Optional. A Cloud Storage bucket used to stage job
    * dependencies, config files, and job driver console output.
    * If you do not specify a staging bucket, Cloud
    * Dataproc will determine a Cloud Storage location (US,
-   * ASIA, or EU) for your cluster's staging bucket according to the Google
+   * ASIA, or EU) for your cluster's staging bucket according to the
    * Compute Engine zone where your cluster is deployed, and then create
    * and manage this project-level, per-location bucket (see
-   * [Cloud Dataproc staging
+   * [Dataproc staging
    * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
    * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java index 81c9bffe..a2271e67 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java @@ -30,7 +30,7 @@ public interface ClusterOrBuilder * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The projectId. */ @@ -42,7 +42,7 @@ public interface ClusterOrBuilder * Required. The Google Cloud Platform project ID that the cluster belongs to. * * - * string project_id = 1; + * string project_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for projectId. */ @@ -56,7 +56,7 @@ public interface ClusterOrBuilder * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @return The clusterName. */ @@ -69,7 +69,7 @@ public interface ClusterOrBuilder * unique. Names of deleted clusters can be reused. * * - * string cluster_name = 2; + * string cluster_name = 2 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for clusterName. */ @@ -79,11 +79,13 @@ public interface ClusterOrBuilder * * *
-   * Required. The cluster config. Note that Cloud Dataproc may set
+   * Required. The cluster config. Note that Dataproc may set
    * default values, and values may change when clusters are updated.
    * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * * * @return Whether the config field is set. */ @@ -92,11 +94,13 @@ public interface ClusterOrBuilder * * *
-   * Required. The cluster config. Note that Cloud Dataproc may set
+   * Required. The cluster config. Note that Dataproc may set
    * default values, and values may change when clusters are updated.
    * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * * * @return The config. */ @@ -105,11 +109,13 @@ public interface ClusterOrBuilder * * *
-   * Required. The cluster config. Note that Cloud Dataproc may set
+   * Required. The cluster config. Note that Dataproc may set
    * default values, and values may change when clusters are updated.
    * 
* - * .google.cloud.dataproc.v1.ClusterConfig config = 3; + * + * .google.cloud.dataproc.v1.ClusterConfig config = 3 [(.google.api.field_behavior) = REQUIRED]; + * */ com.google.cloud.dataproc.v1.ClusterConfigOrBuilder getConfigOrBuilder(); @@ -304,7 +310,7 @@ public interface ClusterOrBuilder * * *
-   * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+   * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
    * generates this value when it creates the cluster.
    * 
* @@ -317,7 +323,7 @@ public interface ClusterOrBuilder * * *
-   * Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
+   * Output only. A cluster UUID (Unique Universal Identifier). Dataproc
    * generates this value when it creates the cluster.
    * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterStatus.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterStatus.java index 0c0a1a36..d8c828b2 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterStatus.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterStatus.java @@ -399,7 +399,7 @@ public enum Substate implements com.google.protobuf.ProtocolMessageEnum { * *
      * The agent-reported status is out of date (may occur if
-     * Cloud Dataproc loses communication with Agent).
+     * Dataproc loses communication with Agent).
      * Applies to RUNNING state.
      * 
* @@ -437,7 +437,7 @@ public enum Substate implements com.google.protobuf.ProtocolMessageEnum { * *
      * The agent-reported status is out of date (may occur if
-     * Cloud Dataproc loses communication with Agent).
+     * Dataproc loses communication with Agent).
      * Applies to RUNNING state.
      * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java index 3fd6046c..fe6c9fb6 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java @@ -157,177 +157,178 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "oogle/longrunning/operations.proto\032\036goog" + "le/protobuf/duration.proto\032 google/proto" + "buf/field_mask.proto\032\037google/protobuf/ti" - + "mestamp.proto\"\271\003\n\007Cluster\022\022\n\nproject_id\030" - + "\001 \001(\t\022\024\n\014cluster_name\030\002 \001(\t\0227\n\006config\030\003 " - + "\001(\0132\'.google.cloud.dataproc.v1.ClusterCo" - + "nfig\022B\n\006labels\030\010 \003(\0132-.google.cloud.data" - + "proc.v1.Cluster.LabelsEntryB\003\340A\001\022<\n\006stat" - + "us\030\004 \001(\0132\'.google.cloud.dataproc.v1.Clus" - + "terStatusB\003\340A\003\022D\n\016status_history\030\007 \003(\0132\'" - + ".google.cloud.dataproc.v1.ClusterStatusB" - + "\003\340A\003\022\031\n\014cluster_uuid\030\006 \001(\tB\003\340A\003\0229\n\007metri" - + "cs\030\t \001(\0132(.google.cloud.dataproc.v1.Clus" - + "terMetrics\032-\n\013LabelsEntry\022\013\n\003key\030\001 \001(\t\022\r" - + "\n\005value\030\002 \001(\t:\0028\001\"\346\005\n\rClusterConfig\022\032\n\rc" - + "onfig_bucket\030\001 \001(\tB\003\340A\001\022K\n\022gce_cluster_c" - + "onfig\030\010 \001(\0132*.google.cloud.dataproc.v1.G" - + "ceClusterConfigB\003\340A\001\022I\n\rmaster_config\030\t " - + "\001(\0132-.google.cloud.dataproc.v1.InstanceG" - + "roupConfigB\003\340A\001\022I\n\rworker_config\030\n \001(\0132-" - + ".google.cloud.dataproc.v1.InstanceGroupC" - + "onfigB\003\340A\001\022S\n\027secondary_worker_config\030\014 " - + "\001(\0132-.google.cloud.dataproc.v1.InstanceG" - + "roupConfigB\003\340A\001\022F\n\017software_config\030\r \001(\013" - + "2(.google.cloud.dataproc.v1.SoftwareConf" - + "igB\003\340A\001\022W\n\026initialization_actions\030\013 \003(\0132" - + "2.google.cloud.dataproc.v1.NodeInitializ" - + "ationActionB\003\340A\001\022J\n\021encryption_config\030\017 " - + "\001(\0132*.google.cloud.dataproc.v1.Encryptio" - + "nConfigB\003\340A\001\022L\n\022autoscaling_config\030\022 \001(\013" - + "2+.google.cloud.dataproc.v1.AutoscalingC" - + "onfigB\003\340A\001\022F\n\017security_config\030\020 \001(\0132(.go" - + "ogle.cloud.dataproc.v1.SecurityConfigB\003\340" - + "A\001\",\n\021AutoscalingConfig\022\027\n\npolicy_uri\030\001 " - + "\001(\tB\003\340A\001\"4\n\020EncryptionConfig\022 \n\023gce_pd_k" - + "ms_key_name\030\001 \001(\tB\003\340A\001\"\315\002\n\020GceClusterCon" - + "fig\022\025\n\010zone_uri\030\001 \001(\tB\003\340A\001\022\030\n\013network_ur" - + "i\030\002 \001(\tB\003\340A\001\022\033\n\016subnetwork_uri\030\006 \001(\tB\003\340A" - + "\001\022\035\n\020internal_ip_only\030\007 \001(\010B\003\340A\001\022\034\n\017serv" - + "ice_account\030\010 \001(\tB\003\340A\001\022#\n\026service_accoun" - + "t_scopes\030\003 \003(\tB\003\340A\001\022\014\n\004tags\030\004 \003(\t\022J\n\010met" - + "adata\030\005 \003(\01328.google.cloud.dataproc.v1.G" - + "ceClusterConfig.MetadataEntry\032/\n\rMetadat" - + "aEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\232" - + "\003\n\023InstanceGroupConfig\022\032\n\rnum_instances\030" - + "\001 \001(\005B\003\340A\001\022\033\n\016instance_names\030\002 \003(\tB\003\340A\003\022" - + "\026\n\timage_uri\030\003 \001(\tB\003\340A\001\022\035\n\020machine_type_" - + "uri\030\004 \001(\tB\003\340A\001\022>\n\013disk_config\030\005 \001(\0132$.go" - + "ogle.cloud.dataproc.v1.DiskConfigB\003\340A\001\022\033" - + "\n\016is_preemptible\030\006 \001(\010B\003\340A\001\022O\n\024managed_g" - + "roup_config\030\007 \001(\0132,.google.cloud.datapro" - + "c.v1.ManagedGroupConfigB\003\340A\003\022F\n\014accelera" - + "tors\030\010 \003(\0132+.google.cloud.dataproc.v1.Ac" - + "celeratorConfigB\003\340A\001\022\035\n\020min_cpu_platform" - + "\030\t \001(\tB\003\340A\001\"c\n\022ManagedGroupConfig\022#\n\026ins" - + "tance_template_name\030\001 \001(\tB\003\340A\003\022(\n\033instan" - + "ce_group_manager_name\030\002 \001(\tB\003\340A\003\"L\n\021Acce" - + "leratorConfig\022\034\n\024accelerator_type_uri\030\001 " - + "\001(\t\022\031\n\021accelerator_count\030\002 \001(\005\"f\n\nDiskCo" - + "nfig\022\033\n\016boot_disk_type\030\003 \001(\tB\003\340A\001\022\036\n\021boo" - + "t_disk_size_gb\030\001 \001(\005B\003\340A\001\022\033\n\016num_local_s" - + "sds\030\002 \001(\005B\003\340A\001\"s\n\030NodeInitializationActi" - + "on\022\034\n\017executable_file\030\001 \001(\tB\003\340A\002\0229\n\021exec" - + "ution_timeout\030\002 \001(\0132\031.google.protobuf.Du" - + "rationB\003\340A\001\"\204\003\n\rClusterStatus\022A\n\005state\030\001" - + " \001(\0162-.google.cloud.dataproc.v1.ClusterS" - + "tatus.StateB\003\340A\003\022\026\n\006detail\030\002 \001(\tB\006\340A\003\340A\001" - + "\0229\n\020state_start_time\030\003 \001(\0132\032.google.prot" - + "obuf.TimestampB\003\340A\003\022G\n\010substate\030\004 \001(\01620." - + "google.cloud.dataproc.v1.ClusterStatus.S" - + "ubstateB\003\340A\003\"V\n\005State\022\013\n\007UNKNOWN\020\000\022\014\n\010CR" - + "EATING\020\001\022\013\n\007RUNNING\020\002\022\t\n\005ERROR\020\003\022\014\n\010DELE" - + "TING\020\004\022\014\n\010UPDATING\020\005\"<\n\010Substate\022\017\n\013UNSP" - + "ECIFIED\020\000\022\r\n\tUNHEALTHY\020\001\022\020\n\014STALE_STATUS" - + "\020\002\"S\n\016SecurityConfig\022A\n\017kerberos_config\030" - + "\001 \001(\0132(.google.cloud.dataproc.v1.Kerbero" - + "sConfig\"\220\004\n\016KerberosConfig\022\034\n\017enable_ker" - + "beros\030\001 \001(\010B\003\340A\001\022(\n\033root_principal_passw" - + "ord_uri\030\002 \001(\tB\003\340A\002\022\030\n\013kms_key_uri\030\003 \001(\tB" - + "\003\340A\002\022\031\n\014keystore_uri\030\004 \001(\tB\003\340A\001\022\033\n\016trust" - + "store_uri\030\005 \001(\tB\003\340A\001\022\"\n\025keystore_passwor" - + "d_uri\030\006 \001(\tB\003\340A\001\022\035\n\020key_password_uri\030\007 \001" - + "(\tB\003\340A\001\022$\n\027truststore_password_uri\030\010 \001(\t" - + "B\003\340A\001\022$\n\027cross_realm_trust_realm\030\t \001(\tB\003" - + "\340A\001\022\"\n\025cross_realm_trust_kdc\030\n \001(\tB\003\340A\001\022" - + "+\n\036cross_realm_trust_admin_server\030\013 \001(\tB" - + "\003\340A\001\0222\n%cross_realm_trust_shared_passwor" - + "d_uri\030\014 \001(\tB\003\340A\001\022\033\n\016kdc_db_key_uri\030\r \001(\t" - + "B\003\340A\001\022\037\n\022tgt_lifetime_hours\030\016 \001(\005B\003\340A\001\022\022" - + "\n\005realm\030\017 \001(\tB\003\340A\001\"\371\001\n\016SoftwareConfig\022\032\n" - + "\rimage_version\030\001 \001(\tB\003\340A\001\022Q\n\nproperties\030" - + "\002 \003(\01328.google.cloud.dataproc.v1.Softwar" - + "eConfig.PropertiesEntryB\003\340A\001\022E\n\023optional" - + "_components\030\003 \003(\0162#.google.cloud.datapro" - + "c.v1.ComponentB\003\340A\001\0321\n\017PropertiesEntry\022\013" - + "\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\232\002\n\016Clust" - + "erMetrics\022O\n\014hdfs_metrics\030\001 \003(\01329.google" - + ".cloud.dataproc.v1.ClusterMetrics.HdfsMe" - + "tricsEntry\022O\n\014yarn_metrics\030\002 \003(\01329.googl" - + "e.cloud.dataproc.v1.ClusterMetrics.YarnM" - + "etricsEntry\0322\n\020HdfsMetricsEntry\022\013\n\003key\030\001" - + " \001(\t\022\r\n\005value\030\002 \001(\003:\0028\001\0322\n\020YarnMetricsEn" - + "try\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\003:\0028\001\"\226\001\n\024" - + "CreateClusterRequest\022\027\n\nproject_id\030\001 \001(\t" - + "B\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\0227\n\007cluster\030\002 " - + "\001(\0132!.google.cloud.dataproc.v1.ClusterB\003" - + "\340A\002\022\027\n\nrequest_id\030\004 \001(\tB\003\340A\001\"\256\002\n\024UpdateC" + + "mestamp.proto\"\310\003\n\007Cluster\022\027\n\nproject_id\030" + + "\001 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\022<\n" + + "\006config\030\003 \001(\0132\'.google.cloud.dataproc.v1" + + ".ClusterConfigB\003\340A\002\022B\n\006labels\030\010 \003(\0132-.go" + + "ogle.cloud.dataproc.v1.Cluster.LabelsEnt" + + "ryB\003\340A\001\022<\n\006status\030\004 \001(\0132\'.google.cloud.d" + + "ataproc.v1.ClusterStatusB\003\340A\003\022D\n\016status_" + + "history\030\007 \003(\0132\'.google.cloud.dataproc.v1" + + ".ClusterStatusB\003\340A\003\022\031\n\014cluster_uuid\030\006 \001(" + + "\tB\003\340A\003\0229\n\007metrics\030\t \001(\0132(.google.cloud.d" + + "ataproc.v1.ClusterMetrics\032-\n\013LabelsEntry" + + "\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\346\005\n\rClu" + + "sterConfig\022\032\n\rconfig_bucket\030\001 \001(\tB\003\340A\001\022K" + + "\n\022gce_cluster_config\030\010 \001(\0132*.google.clou" + + "d.dataproc.v1.GceClusterConfigB\003\340A\001\022I\n\rm" + + "aster_config\030\t \001(\0132-.google.cloud.datapr" + + "oc.v1.InstanceGroupConfigB\003\340A\001\022I\n\rworker" + + "_config\030\n \001(\0132-.google.cloud.dataproc.v1" + + ".InstanceGroupConfigB\003\340A\001\022S\n\027secondary_w" + + "orker_config\030\014 \001(\0132-.google.cloud.datapr" + + "oc.v1.InstanceGroupConfigB\003\340A\001\022F\n\017softwa" + + "re_config\030\r \001(\0132(.google.cloud.dataproc." + + "v1.SoftwareConfigB\003\340A\001\022W\n\026initialization" + + "_actions\030\013 \003(\01322.google.cloud.dataproc.v" + + "1.NodeInitializationActionB\003\340A\001\022J\n\021encry" + + "ption_config\030\017 \001(\0132*.google.cloud.datapr" + + "oc.v1.EncryptionConfigB\003\340A\001\022L\n\022autoscali" + + "ng_config\030\022 \001(\0132+.google.cloud.dataproc." + + "v1.AutoscalingConfigB\003\340A\001\022F\n\017security_co" + + "nfig\030\020 \001(\0132(.google.cloud.dataproc.v1.Se" + + "curityConfigB\003\340A\001\",\n\021AutoscalingConfig\022\027" + + "\n\npolicy_uri\030\001 \001(\tB\003\340A\001\"4\n\020EncryptionCon" + + "fig\022 \n\023gce_pd_kms_key_name\030\001 \001(\tB\003\340A\001\"\315\002" + + "\n\020GceClusterConfig\022\025\n\010zone_uri\030\001 \001(\tB\003\340A" + + "\001\022\030\n\013network_uri\030\002 \001(\tB\003\340A\001\022\033\n\016subnetwor" + + "k_uri\030\006 \001(\tB\003\340A\001\022\035\n\020internal_ip_only\030\007 \001" + + "(\010B\003\340A\001\022\034\n\017service_account\030\010 \001(\tB\003\340A\001\022#\n" + + "\026service_account_scopes\030\003 \003(\tB\003\340A\001\022\014\n\004ta" + + "gs\030\004 \003(\t\022J\n\010metadata\030\005 \003(\01328.google.clou" + + "d.dataproc.v1.GceClusterConfig.MetadataE" + + "ntry\032/\n\rMetadataEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005va" + + "lue\030\002 \001(\t:\0028\001\"\232\003\n\023InstanceGroupConfig\022\032\n" + + "\rnum_instances\030\001 \001(\005B\003\340A\001\022\033\n\016instance_na" + + "mes\030\002 \003(\tB\003\340A\003\022\026\n\timage_uri\030\003 \001(\tB\003\340A\001\022\035" + + "\n\020machine_type_uri\030\004 \001(\tB\003\340A\001\022>\n\013disk_co" + + "nfig\030\005 \001(\0132$.google.cloud.dataproc.v1.Di" + + "skConfigB\003\340A\001\022\033\n\016is_preemptible\030\006 \001(\010B\003\340" + + "A\001\022O\n\024managed_group_config\030\007 \001(\0132,.googl" + + "e.cloud.dataproc.v1.ManagedGroupConfigB\003" + + "\340A\003\022F\n\014accelerators\030\010 \003(\0132+.google.cloud" + + ".dataproc.v1.AcceleratorConfigB\003\340A\001\022\035\n\020m" + + "in_cpu_platform\030\t \001(\tB\003\340A\001\"c\n\022ManagedGro" + + "upConfig\022#\n\026instance_template_name\030\001 \001(\t" + + "B\003\340A\003\022(\n\033instance_group_manager_name\030\002 \001" + + "(\tB\003\340A\003\"L\n\021AcceleratorConfig\022\034\n\024accelera" + + "tor_type_uri\030\001 \001(\t\022\031\n\021accelerator_count\030" + + "\002 \001(\005\"f\n\nDiskConfig\022\033\n\016boot_disk_type\030\003 " + + "\001(\tB\003\340A\001\022\036\n\021boot_disk_size_gb\030\001 \001(\005B\003\340A\001" + + "\022\033\n\016num_local_ssds\030\002 \001(\005B\003\340A\001\"s\n\030NodeIni" + + "tializationAction\022\034\n\017executable_file\030\001 \001" + + "(\tB\003\340A\002\0229\n\021execution_timeout\030\002 \001(\0132\031.goo" + + "gle.protobuf.DurationB\003\340A\001\"\204\003\n\rClusterSt" + + "atus\022A\n\005state\030\001 \001(\0162-.google.cloud.datap" + + "roc.v1.ClusterStatus.StateB\003\340A\003\022\026\n\006detai" + + "l\030\002 \001(\tB\006\340A\003\340A\001\0229\n\020state_start_time\030\003 \001(" + + "\0132\032.google.protobuf.TimestampB\003\340A\003\022G\n\010su" + + "bstate\030\004 \001(\01620.google.cloud.dataproc.v1." + + "ClusterStatus.SubstateB\003\340A\003\"V\n\005State\022\013\n\007" + + "UNKNOWN\020\000\022\014\n\010CREATING\020\001\022\013\n\007RUNNING\020\002\022\t\n\005" + + "ERROR\020\003\022\014\n\010DELETING\020\004\022\014\n\010UPDATING\020\005\"<\n\010S" + + "ubstate\022\017\n\013UNSPECIFIED\020\000\022\r\n\tUNHEALTHY\020\001\022" + + "\020\n\014STALE_STATUS\020\002\"S\n\016SecurityConfig\022A\n\017k" + + "erberos_config\030\001 \001(\0132(.google.cloud.data" + + "proc.v1.KerberosConfig\"\220\004\n\016KerberosConfi" + + "g\022\034\n\017enable_kerberos\030\001 \001(\010B\003\340A\001\022(\n\033root_" + + "principal_password_uri\030\002 \001(\tB\003\340A\002\022\030\n\013kms" + + "_key_uri\030\003 \001(\tB\003\340A\002\022\031\n\014keystore_uri\030\004 \001(" + + "\tB\003\340A\001\022\033\n\016truststore_uri\030\005 \001(\tB\003\340A\001\022\"\n\025k" + + "eystore_password_uri\030\006 \001(\tB\003\340A\001\022\035\n\020key_p" + + "assword_uri\030\007 \001(\tB\003\340A\001\022$\n\027truststore_pas" + + "sword_uri\030\010 \001(\tB\003\340A\001\022$\n\027cross_realm_trus" + + "t_realm\030\t \001(\tB\003\340A\001\022\"\n\025cross_realm_trust_" + + "kdc\030\n \001(\tB\003\340A\001\022+\n\036cross_realm_trust_admi" + + "n_server\030\013 \001(\tB\003\340A\001\0222\n%cross_realm_trust" + + "_shared_password_uri\030\014 \001(\tB\003\340A\001\022\033\n\016kdc_d" + + "b_key_uri\030\r \001(\tB\003\340A\001\022\037\n\022tgt_lifetime_hou" + + "rs\030\016 \001(\005B\003\340A\001\022\022\n\005realm\030\017 \001(\tB\003\340A\001\"\371\001\n\016So" + + "ftwareConfig\022\032\n\rimage_version\030\001 \001(\tB\003\340A\001" + + "\022Q\n\nproperties\030\002 \003(\01328.google.cloud.data" + + "proc.v1.SoftwareConfig.PropertiesEntryB\003" + + "\340A\001\022E\n\023optional_components\030\003 \003(\0162#.googl" + + "e.cloud.dataproc.v1.ComponentB\003\340A\001\0321\n\017Pr" + + "opertiesEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(" + + "\t:\0028\001\"\232\002\n\016ClusterMetrics\022O\n\014hdfs_metrics" + + "\030\001 \003(\01329.google.cloud.dataproc.v1.Cluste" + + "rMetrics.HdfsMetricsEntry\022O\n\014yarn_metric" + + "s\030\002 \003(\01329.google.cloud.dataproc.v1.Clust" + + "erMetrics.YarnMetricsEntry\0322\n\020HdfsMetric" + + "sEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\003:\0028\001\0322" + + "\n\020YarnMetricsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value" + + "\030\002 \001(\003:\0028\001\"\226\001\n\024CreateClusterRequest\022\027\n\np" + + "roject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A" + + "\002\0227\n\007cluster\030\002 \001(\0132!.google.cloud.datapr" + + "oc.v1.ClusterB\003\340A\002\022\027\n\nrequest_id\030\004 \001(\tB\003" + + "\340A\001\"\256\002\n\024UpdateClusterRequest\022\027\n\nproject_" + + "id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\005 \001(\tB\003\340A\002\022\031\n\014cl" + + "uster_name\030\002 \001(\tB\003\340A\002\0227\n\007cluster\030\003 \001(\0132!" + + ".google.cloud.dataproc.v1.ClusterB\003\340A\002\022E" + + "\n\035graceful_decommission_timeout\030\006 \001(\0132\031." + + "google.protobuf.DurationB\003\340A\001\0224\n\013update_" + + "mask\030\004 \001(\0132\032.google.protobuf.FieldMaskB\003" + + "\340A\002\022\027\n\nrequest_id\030\007 \001(\tB\003\340A\001\"\223\001\n\024DeleteC" + "lusterRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023" - + "\n\006region\030\005 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(" - + "\tB\003\340A\002\0227\n\007cluster\030\003 \001(\0132!.google.cloud.d" - + "ataproc.v1.ClusterB\003\340A\002\022E\n\035graceful_deco" - + "mmission_timeout\030\006 \001(\0132\031.google.protobuf" - + ".DurationB\003\340A\001\0224\n\013update_mask\030\004 \001(\0132\032.go" - + "ogle.protobuf.FieldMaskB\003\340A\002\022\027\n\nrequest_" - + "id\030\007 \001(\tB\003\340A\001\"\223\001\n\024DeleteClusterRequest\022\027" - + "\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB" - + "\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\022\031\n\014clust" - + "er_uuid\030\004 \001(\tB\003\340A\001\022\027\n\nrequest_id\030\005 \001(\tB\003" - + "\340A\001\"\\\n\021GetClusterRequest\022\027\n\nproject_id\030\001" - + " \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cluste" - + "r_name\030\002 \001(\tB\003\340A\002\"\211\001\n\023ListClustersReques" - + "t\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\004 \001" - + "(\tB\003\340A\002\022\023\n\006filter\030\005 \001(\tB\003\340A\001\022\026\n\tpage_siz" - + "e\030\002 \001(\005B\003\340A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001\"n\n" - + "\024ListClustersResponse\0228\n\010clusters\030\001 \003(\0132" - + "!.google.cloud.dataproc.v1.ClusterB\003\340A\003\022" - + "\034\n\017next_page_token\030\002 \001(\tB\003\340A\003\"a\n\026Diagnos" - + "eClusterRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002" - + "\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 " - + "\001(\tB\003\340A\002\"1\n\026DiagnoseClusterResults\022\027\n\nou" - + "tput_uri\030\001 \001(\tB\003\340A\0032\256\014\n\021ClusterControlle" - + "r\022\200\002\n\rCreateCluster\022..google.cloud.datap" - + "roc.v1.CreateClusterRequest\032\035.google.lon" - + "grunning.Operation\"\237\001\202\323\344\223\002>\"3/v1/project" - + "s/{project_id}/regions/{region}/clusters" - + ":\007cluster\332A\031project_id,region,cluster\312A<" - + "\n\007Cluster\0221google.cloud.dataproc.v1.Clus" - + "terOperationMetadata\022\363\001\n\rUpdateCluster\022." - + ".google.cloud.dataproc.v1.UpdateClusterR" - + "equest\032\035.google.longrunning.Operation\"\222\001" - + "\202\323\344\223\002M2B/v1/projects/{project_id}/region" - + "s/{region}/clusters/{cluster_name}:\007clus" - + "ter\312A<\n\007Cluster\0221google.cloud.dataproc.v" - + "1.ClusterOperationMetadata\022\231\002\n\rDeleteClu" - + "ster\022..google.cloud.dataproc.v1.DeleteCl" - + "usterRequest\032\035.google.longrunning.Operat" - + "ion\"\270\001\202\323\344\223\002D*B/v1/projects/{project_id}/" - + "regions/{region}/clusters/{cluster_name}" - + "\332A\036project_id,region,cluster_name\312AJ\n\025go" - + "ogle.protobuf.Empty\0221google.cloud.datapr" - + "oc.v1.ClusterOperationMetadata\022\311\001\n\nGetCl" - + "uster\022+.google.cloud.dataproc.v1.GetClus" - + "terRequest\032!.google.cloud.dataproc.v1.Cl" - + "uster\"k\202\323\344\223\002D\022B/v1/projects/{project_id}" - + "/regions/{region}/clusters/{cluster_name" - + "}\332A\036project_id,region,cluster_name\022\331\001\n\014L" - + "istClusters\022-.google.cloud.dataproc.v1.L" - + "istClustersRequest\032..google.cloud.datapr" - + "oc.v1.ListClustersResponse\"j\202\323\344\223\0025\0223/v1/" - + "projects/{project_id}/regions/{region}/c" - + "lusters\332A\021project_id,region\332A\030project_id" - + ",region,filter\022\216\002\n\017DiagnoseCluster\0220.goo" - + "gle.cloud.dataproc.v1.DiagnoseClusterReq" - + "uest\032\035.google.longrunning.Operation\"\251\001\202\323" - + "\344\223\002P\"K/v1/projects/{project_id}/regions/" - + "{region}/clusters/{cluster_name}:diagnos" - + "e:\001*\332A\036project_id,region,cluster_name\312A/" - + "\n\025google.protobuf.Empty\022\026DiagnoseCluster" - + "Results\032K\312A\027dataproc.googleapis.com\322A.ht" - + "tps://www.googleapis.com/auth/cloud-plat" - + "formBq\n\034com.google.cloud.dataproc.v1B\rCl" - + "ustersProtoP\001Z@google.golang.org/genprot" - + "o/googleapis/cloud/dataproc/v1;dataprocb" - + "\006proto3" + + "\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(" + + "\tB\003\340A\002\022\031\n\014cluster_uuid\030\004 \001(\tB\003\340A\001\022\027\n\nreq" + + "uest_id\030\005 \001(\tB\003\340A\001\"\\\n\021GetClusterRequest\022" + + "\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\t" + + "B\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\"\211\001\n\023Lis" + + "tClustersRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A" + + "\002\022\023\n\006region\030\004 \001(\tB\003\340A\002\022\023\n\006filter\030\005 \001(\tB\003" + + "\340A\001\022\026\n\tpage_size\030\002 \001(\005B\003\340A\001\022\027\n\npage_toke" + + "n\030\003 \001(\tB\003\340A\001\"n\n\024ListClustersResponse\0228\n\010" + + "clusters\030\001 \003(\0132!.google.cloud.dataproc.v" + + "1.ClusterB\003\340A\003\022\034\n\017next_page_token\030\002 \001(\tB" + + "\003\340A\003\"a\n\026DiagnoseClusterRequest\022\027\n\nprojec" + + "t_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014" + + "cluster_name\030\002 \001(\tB\003\340A\002\"1\n\026DiagnoseClust" + + "erResults\022\027\n\noutput_uri\030\001 \001(\tB\003\340A\0032\343\014\n\021C" + + "lusterController\022\200\002\n\rCreateCluster\022..goo" + + "gle.cloud.dataproc.v1.CreateClusterReque" + + "st\032\035.google.longrunning.Operation\"\237\001\202\323\344\223" + + "\002>\"3/v1/projects/{project_id}/regions/{r" + + "egion}/clusters:\007cluster\332A\031project_id,re" + + "gion,cluster\312A<\n\007Cluster\0221google.cloud.d" + + "ataproc.v1.ClusterOperationMetadata\022\250\002\n\r" + + "UpdateCluster\022..google.cloud.dataproc.v1" + + ".UpdateClusterRequest\032\035.google.longrunni" + + "ng.Operation\"\307\001\202\323\344\223\002M2B/v1/projects/{pro" + + "ject_id}/regions/{region}/clusters/{clus" + + "ter_name}:\007cluster\312A<\n\007Cluster\0221google.c" + + "loud.dataproc.v1.ClusterOperationMetadat" + + "a\332A2project_id,region,cluster_name,clust" + + "er,update_mask\022\231\002\n\rDeleteCluster\022..googl" + + "e.cloud.dataproc.v1.DeleteClusterRequest" + + "\032\035.google.longrunning.Operation\"\270\001\202\323\344\223\002D" + + "*B/v1/projects/{project_id}/regions/{reg" + + "ion}/clusters/{cluster_name}\332A\036project_i" + + "d,region,cluster_name\312AJ\n\025google.protobu" + + "f.Empty\0221google.cloud.dataproc.v1.Cluste" + + "rOperationMetadata\022\311\001\n\nGetCluster\022+.goog" + + "le.cloud.dataproc.v1.GetClusterRequest\032!" + + ".google.cloud.dataproc.v1.Cluster\"k\202\323\344\223\002" + + "D\022B/v1/projects/{project_id}/regions/{re" + + "gion}/clusters/{cluster_name}\332A\036project_" + + "id,region,cluster_name\022\331\001\n\014ListClusters\022" + + "-.google.cloud.dataproc.v1.ListClustersR" + + "equest\032..google.cloud.dataproc.v1.ListCl" + + "ustersResponse\"j\202\323\344\223\0025\0223/v1/projects/{pr" + + "oject_id}/regions/{region}/clusters\332A\021pr" + + "oject_id,region\332A\030project_id,region,filt" + + "er\022\216\002\n\017DiagnoseCluster\0220.google.cloud.da" + + "taproc.v1.DiagnoseClusterRequest\032\035.googl" + + "e.longrunning.Operation\"\251\001\202\323\344\223\002P\"K/v1/pr" + + "ojects/{project_id}/regions/{region}/clu" + + "sters/{cluster_name}:diagnose:\001*\332A\036proje" + + "ct_id,region,cluster_name\312A/\n\025google.pro" + + "tobuf.Empty\022\026DiagnoseClusterResults\032K\312A\027" + + "dataproc.googleapis.com\322A.https://www.go" + + "ogleapis.com/auth/cloud-platformBq\n\034com." + + "google.cloud.dataproc.v1B\rClustersProtoP" + + "\001Z@google.golang.org/genproto/googleapis" + + "/cloud/dataproc/v1;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequest.java index 5daab432..5d33c27a 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequest.java @@ -198,7 +198,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -220,7 +220,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -842,7 +842,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -864,7 +864,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -886,7 +886,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -907,7 +907,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -924,7 +924,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequestOrBuilder.java index 7f24af8d..49482939 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface CreateClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface CreateClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequest.java index e5f7ab50..028c8667 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequest.java @@ -198,7 +198,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -220,7 +220,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -896,7 +896,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -918,7 +918,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -940,7 +940,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -961,7 +961,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -978,7 +978,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequestOrBuilder.java index 232c2f3f..bbf9c6a7 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface DeleteClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface DeleteClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequest.java index 390f6962..619b3fe2 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequest.java @@ -182,7 +182,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -204,7 +204,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -734,7 +734,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -756,7 +756,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -778,7 +778,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -799,7 +799,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -816,7 +816,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequestOrBuilder.java index fd7b266b..47009022 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface DeleteJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface DeleteJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequest.java index 427c2887..66cedbe2 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequest.java @@ -182,7 +182,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -204,7 +204,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -736,7 +736,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -758,7 +758,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -780,7 +780,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -801,7 +801,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -818,7 +818,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequestOrBuilder.java index a49b1053..b41a0b2c 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface DiagnoseClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface DiagnoseClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java index 5b178c6a..c6290b68 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java @@ -204,7 +204,7 @@ protected com.google.protobuf.MapField internalGetMapField(int number) { *
    * Optional. The zone where the Compute Engine cluster will be located.
    * On a create request, it is required in the "global" region. If omitted
-   * in a non-global Cloud Dataproc region, the service will pick a zone in the
+   * in a non-global Dataproc region, the service will pick a zone in the
    * corresponding Compute Engine region. On a get request, zone will
    * always be present.
    * A full URL, partial URI, or short name are valid. Examples:
@@ -234,7 +234,7 @@ public java.lang.String getZoneUri() {
    * 
    * Optional. The zone where the Compute Engine cluster will be located.
    * On a create request, it is required in the "global" region. If omitted
-   * in a non-global Cloud Dataproc region, the service will pick a zone in the
+   * in a non-global Dataproc region, the service will pick a zone in the
    * corresponding Compute Engine region. On a get request, zone will
    * always be present.
    * A full URL, partial URI, or short name are valid. Examples:
@@ -407,15 +407,16 @@ public boolean getInternalIpOnly() {
    *
    *
    * 
-   * Optional. The service account of the instances. Defaults to the default
-   * Compute Engine service account. Custom service accounts need
-   * permissions equivalent to the following IAM roles:
-   * * roles/logging.logWriter
-   * * roles/storage.objectAdmin
-   * (see
-   * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-   * for more information).
-   * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+   * Optional. The [Dataproc service
+   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * (also see [VM Data Plane
+   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * used by Dataproc cluster VM instances to access Google Cloud Platform
+   * services.
+   * If not specified, the
+   * [Compute Engine default service
+   * account](/compute/docs/access/service-accounts#default_service_account)
+   * is used.
    * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; @@ -437,15 +438,16 @@ public java.lang.String getServiceAccount() { * * *
-   * Optional. The service account of the instances. Defaults to the default
-   * Compute Engine service account. Custom service accounts need
-   * permissions equivalent to the following IAM roles:
-   * * roles/logging.logWriter
-   * * roles/storage.objectAdmin
-   * (see
-   * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-   * for more information).
-   * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+   * Optional. The [Dataproc service
+   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * (also see [VM Data Plane
+   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * used by Dataproc cluster VM instances to access Google Cloud Platform
+   * services.
+   * If not specified, the
+   * [Compute Engine default service
+   * account](/compute/docs/access/service-accounts#default_service_account)
+   * is used.
    * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; @@ -1234,7 +1236,7 @@ public Builder mergeFrom( *
      * Optional. The zone where the Compute Engine cluster will be located.
      * On a create request, it is required in the "global" region. If omitted
-     * in a non-global Cloud Dataproc region, the service will pick a zone in the
+     * in a non-global Dataproc region, the service will pick a zone in the
      * corresponding Compute Engine region. On a get request, zone will
      * always be present.
      * A full URL, partial URI, or short name are valid. Examples:
@@ -1264,7 +1266,7 @@ public java.lang.String getZoneUri() {
      * 
      * Optional. The zone where the Compute Engine cluster will be located.
      * On a create request, it is required in the "global" region. If omitted
-     * in a non-global Cloud Dataproc region, the service will pick a zone in the
+     * in a non-global Dataproc region, the service will pick a zone in the
      * corresponding Compute Engine region. On a get request, zone will
      * always be present.
      * A full URL, partial URI, or short name are valid. Examples:
@@ -1294,7 +1296,7 @@ public com.google.protobuf.ByteString getZoneUriBytes() {
      * 
      * Optional. The zone where the Compute Engine cluster will be located.
      * On a create request, it is required in the "global" region. If omitted
-     * in a non-global Cloud Dataproc region, the service will pick a zone in the
+     * in a non-global Dataproc region, the service will pick a zone in the
      * corresponding Compute Engine region. On a get request, zone will
      * always be present.
      * A full URL, partial URI, or short name are valid. Examples:
@@ -1323,7 +1325,7 @@ public Builder setZoneUri(java.lang.String value) {
      * 
      * Optional. The zone where the Compute Engine cluster will be located.
      * On a create request, it is required in the "global" region. If omitted
-     * in a non-global Cloud Dataproc region, the service will pick a zone in the
+     * in a non-global Dataproc region, the service will pick a zone in the
      * corresponding Compute Engine region. On a get request, zone will
      * always be present.
      * A full URL, partial URI, or short name are valid. Examples:
@@ -1348,7 +1350,7 @@ public Builder clearZoneUri() {
      * 
      * Optional. The zone where the Compute Engine cluster will be located.
      * On a create request, it is required in the "global" region. If omitted
-     * in a non-global Cloud Dataproc region, the service will pick a zone in the
+     * in a non-global Dataproc region, the service will pick a zone in the
      * corresponding Compute Engine region. On a get request, zone will
      * always be present.
      * A full URL, partial URI, or short name are valid. Examples:
@@ -1721,15 +1723,16 @@ public Builder clearInternalIpOnly() {
      *
      *
      * 
-     * Optional. The service account of the instances. Defaults to the default
-     * Compute Engine service account. Custom service accounts need
-     * permissions equivalent to the following IAM roles:
-     * * roles/logging.logWriter
-     * * roles/storage.objectAdmin
-     * (see
-     * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-     * for more information).
-     * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+     * Optional. The [Dataproc service
+     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * (also see [VM Data Plane
+     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * used by Dataproc cluster VM instances to access Google Cloud Platform
+     * services.
+     * If not specified, the
+     * [Compute Engine default service
+     * account](/compute/docs/access/service-accounts#default_service_account)
+     * is used.
      * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; @@ -1751,15 +1754,16 @@ public java.lang.String getServiceAccount() { * * *
-     * Optional. The service account of the instances. Defaults to the default
-     * Compute Engine service account. Custom service accounts need
-     * permissions equivalent to the following IAM roles:
-     * * roles/logging.logWriter
-     * * roles/storage.objectAdmin
-     * (see
-     * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-     * for more information).
-     * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+     * Optional. The [Dataproc service
+     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * (also see [VM Data Plane
+     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * used by Dataproc cluster VM instances to access Google Cloud Platform
+     * services.
+     * If not specified, the
+     * [Compute Engine default service
+     * account](/compute/docs/access/service-accounts#default_service_account)
+     * is used.
      * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; @@ -1781,15 +1785,16 @@ public com.google.protobuf.ByteString getServiceAccountBytes() { * * *
-     * Optional. The service account of the instances. Defaults to the default
-     * Compute Engine service account. Custom service accounts need
-     * permissions equivalent to the following IAM roles:
-     * * roles/logging.logWriter
-     * * roles/storage.objectAdmin
-     * (see
-     * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-     * for more information).
-     * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+     * Optional. The [Dataproc service
+     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * (also see [VM Data Plane
+     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * used by Dataproc cluster VM instances to access Google Cloud Platform
+     * services.
+     * If not specified, the
+     * [Compute Engine default service
+     * account](/compute/docs/access/service-accounts#default_service_account)
+     * is used.
      * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; @@ -1810,15 +1815,16 @@ public Builder setServiceAccount(java.lang.String value) { * * *
-     * Optional. The service account of the instances. Defaults to the default
-     * Compute Engine service account. Custom service accounts need
-     * permissions equivalent to the following IAM roles:
-     * * roles/logging.logWriter
-     * * roles/storage.objectAdmin
-     * (see
-     * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-     * for more information).
-     * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+     * Optional. The [Dataproc service
+     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * (also see [VM Data Plane
+     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * used by Dataproc cluster VM instances to access Google Cloud Platform
+     * services.
+     * If not specified, the
+     * [Compute Engine default service
+     * account](/compute/docs/access/service-accounts#default_service_account)
+     * is used.
      * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; @@ -1835,15 +1841,16 @@ public Builder clearServiceAccount() { * * *
-     * Optional. The service account of the instances. Defaults to the default
-     * Compute Engine service account. Custom service accounts need
-     * permissions equivalent to the following IAM roles:
-     * * roles/logging.logWriter
-     * * roles/storage.objectAdmin
-     * (see
-     * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-     * for more information).
-     * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+     * Optional. The [Dataproc service
+     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * (also see [VM Data Plane
+     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * used by Dataproc cluster VM instances to access Google Cloud Platform
+     * services.
+     * If not specified, the
+     * [Compute Engine default service
+     * account](/compute/docs/access/service-accounts#default_service_account)
+     * is used.
      * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java index 3b7a5a52..6e013d94 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java @@ -29,7 +29,7 @@ public interface GceClusterConfigOrBuilder *
    * Optional. The zone where the Compute Engine cluster will be located.
    * On a create request, it is required in the "global" region. If omitted
-   * in a non-global Cloud Dataproc region, the service will pick a zone in the
+   * in a non-global Dataproc region, the service will pick a zone in the
    * corresponding Compute Engine region. On a get request, zone will
    * always be present.
    * A full URL, partial URI, or short name are valid. Examples:
@@ -49,7 +49,7 @@ public interface GceClusterConfigOrBuilder
    * 
    * Optional. The zone where the Compute Engine cluster will be located.
    * On a create request, it is required in the "global" region. If omitted
-   * in a non-global Cloud Dataproc region, the service will pick a zone in the
+   * in a non-global Dataproc region, the service will pick a zone in the
    * corresponding Compute Engine region. On a get request, zone will
    * always be present.
    * A full URL, partial URI, or short name are valid. Examples:
@@ -162,15 +162,16 @@ public interface GceClusterConfigOrBuilder
    *
    *
    * 
-   * Optional. The service account of the instances. Defaults to the default
-   * Compute Engine service account. Custom service accounts need
-   * permissions equivalent to the following IAM roles:
-   * * roles/logging.logWriter
-   * * roles/storage.objectAdmin
-   * (see
-   * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-   * for more information).
-   * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+   * Optional. The [Dataproc service
+   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * (also see [VM Data Plane
+   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * used by Dataproc cluster VM instances to access Google Cloud Platform
+   * services.
+   * If not specified, the
+   * [Compute Engine default service
+   * account](/compute/docs/access/service-accounts#default_service_account)
+   * is used.
    * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; @@ -182,15 +183,16 @@ public interface GceClusterConfigOrBuilder * * *
-   * Optional. The service account of the instances. Defaults to the default
-   * Compute Engine service account. Custom service accounts need
-   * permissions equivalent to the following IAM roles:
-   * * roles/logging.logWriter
-   * * roles/storage.objectAdmin
-   * (see
-   * https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
-   * for more information).
-   * Example: `[account_id]@[project_id].iam.gserviceaccount.com`
+   * Optional. The [Dataproc service
+   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * (also see [VM Data Plane
+   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * used by Dataproc cluster VM instances to access Google Cloud Platform
+   * services.
+   * If not specified, the
+   * [Compute Engine default service
+   * account](/compute/docs/access/service-accounts#default_service_account)
+   * is used.
    * 
* * string service_account = 8 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequest.java index 1e5affdd..9ddb9a0e 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequest.java @@ -182,7 +182,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -204,7 +204,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -734,7 +734,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -756,7 +756,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -778,7 +778,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -799,7 +799,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -816,7 +816,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequestOrBuilder.java index da09621f..fe21af99 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface GetClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface GetClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequest.java index ade1a5cf..95a2029e 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequest.java @@ -182,7 +182,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -204,7 +204,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -734,7 +734,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -756,7 +756,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -778,7 +778,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -799,7 +799,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -816,7 +816,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequestOrBuilder.java index c31294b0..05b1e81b 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface GetJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface GetJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJob.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJob.java index 0f5246ad..cc1116fa 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJob.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJob.java @@ -22,7 +22,7 @@ * * *
- * A Cloud Dataproc job for running
+ * A Dataproc job for running
  * [Apache Hadoop
  * MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html)
  * jobs on [Apache Hadoop
@@ -695,7 +695,7 @@ public int getPropertiesCount() {
    *
    * 
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
@@ -719,7 +719,7 @@ public java.util.Map getProperties() { * *
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
@@ -735,7 +735,7 @@ public java.util.Map getPropertiesMap() { * *
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
@@ -756,7 +756,7 @@ public java.lang.String getPropertiesOrDefault( * *
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
@@ -1109,7 +1109,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc job for running
+   * A Dataproc job for running
    * [Apache Hadoop
    * MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html)
    * jobs on [Apache Hadoop
@@ -2447,7 +2447,7 @@ public int getPropertiesCount() {
      *
      * 
      * Optional. A mapping of property names to values, used to configure Hadoop.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
      * classes in user code.
      * 
@@ -2471,7 +2471,7 @@ public java.util.Map getProperties() { * *
      * Optional. A mapping of property names to values, used to configure Hadoop.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
      * classes in user code.
      * 
@@ -2487,7 +2487,7 @@ public java.util.Map getPropertiesMap() { * *
      * Optional. A mapping of property names to values, used to configure Hadoop.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
      * classes in user code.
      * 
@@ -2508,7 +2508,7 @@ public java.lang.String getPropertiesOrDefault( * *
      * Optional. A mapping of property names to values, used to configure Hadoop.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
      * classes in user code.
      * 
@@ -2536,7 +2536,7 @@ public Builder clearProperties() { * *
      * Optional. A mapping of property names to values, used to configure Hadoop.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
      * classes in user code.
      * 
@@ -2561,7 +2561,7 @@ public java.util.Map getMutableProperties() * *
      * Optional. A mapping of property names to values, used to configure Hadoop.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
      * classes in user code.
      * 
@@ -2584,7 +2584,7 @@ public Builder putProperties(java.lang.String key, java.lang.String value) { * *
      * Optional. A mapping of property names to values, used to configure Hadoop.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
      * classes in user code.
      * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJobOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJobOrBuilder.java index 0bac8db4..1c9fd6b5 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJobOrBuilder.java @@ -324,7 +324,7 @@ public interface HadoopJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
@@ -338,7 +338,7 @@ public interface HadoopJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
@@ -355,7 +355,7 @@ public interface HadoopJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
@@ -369,7 +369,7 @@ public interface HadoopJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
@@ -383,7 +383,7 @@ public interface HadoopJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Hadoop.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site and
    * classes in user code.
    * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJob.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJob.java index 16707404..ae47513b 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJob.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJob.java @@ -22,7 +22,7 @@ * * *
- * A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/)
+ * A Dataproc job for running [Apache Hive](https://hive.apache.org/)
  * queries on YARN.
  * 
* @@ -496,7 +496,7 @@ public int getPropertiesCount() { * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
@@ -520,7 +520,7 @@ public java.util.Map getProperties() { * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
@@ -536,7 +536,7 @@ public java.util.Map getPropertiesMap() { * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
@@ -557,7 +557,7 @@ public java.lang.String getPropertiesOrDefault( * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
@@ -893,7 +893,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/)
+   * A Dataproc job for running [Apache Hive](https://hive.apache.org/)
    * queries on YARN.
    * 
* @@ -1731,7 +1731,7 @@ public int getPropertiesCount() { * *
      * Optional. A mapping of property names and values, used to configure Hive.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/hive/conf/hive-site.xml, and classes in user code.
      * 
@@ -1755,7 +1755,7 @@ public java.util.Map getProperties() { * *
      * Optional. A mapping of property names and values, used to configure Hive.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/hive/conf/hive-site.xml, and classes in user code.
      * 
@@ -1771,7 +1771,7 @@ public java.util.Map getPropertiesMap() { * *
      * Optional. A mapping of property names and values, used to configure Hive.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/hive/conf/hive-site.xml, and classes in user code.
      * 
@@ -1792,7 +1792,7 @@ public java.lang.String getPropertiesOrDefault( * *
      * Optional. A mapping of property names and values, used to configure Hive.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/hive/conf/hive-site.xml, and classes in user code.
      * 
@@ -1820,7 +1820,7 @@ public Builder clearProperties() { * *
      * Optional. A mapping of property names and values, used to configure Hive.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/hive/conf/hive-site.xml, and classes in user code.
      * 
@@ -1845,7 +1845,7 @@ public java.util.Map getMutableProperties() * *
      * Optional. A mapping of property names and values, used to configure Hive.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/hive/conf/hive-site.xml, and classes in user code.
      * 
@@ -1868,7 +1868,7 @@ public Builder putProperties(java.lang.String key, java.lang.String value) { * *
      * Optional. A mapping of property names and values, used to configure Hive.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/hive/conf/hive-site.xml, and classes in user code.
      * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJobOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJobOrBuilder.java index 4810700e..88ed51f6 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJobOrBuilder.java @@ -167,7 +167,7 @@ public interface HiveJobOrBuilder * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
@@ -181,7 +181,7 @@ public interface HiveJobOrBuilder * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
@@ -198,7 +198,7 @@ public interface HiveJobOrBuilder * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
@@ -212,7 +212,7 @@ public interface HiveJobOrBuilder * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
@@ -226,7 +226,7 @@ public interface HiveJobOrBuilder * *
    * Optional. A mapping of property names and values, used to configure Hive.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/hive/conf/hive-site.xml, and classes in user code.
    * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java index de27e738..1bd6d2c6 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java @@ -225,7 +225,7 @@ public int getNumInstances() { * * *
-   * Output only. The list of instance names. Cloud Dataproc derives the names
+   * Output only. The list of instance names. Dataproc derives the names
    * from `cluster_name`, `num_instances`, and the instance group.
    * 
* @@ -240,7 +240,7 @@ public com.google.protobuf.ProtocolStringList getInstanceNamesList() { * * *
-   * Output only. The list of instance names. Cloud Dataproc derives the names
+   * Output only. The list of instance names. Dataproc derives the names
    * from `cluster_name`, `num_instances`, and the instance group.
    * 
* @@ -255,7 +255,7 @@ public int getInstanceNamesCount() { * * *
-   * Output only. The list of instance names. Cloud Dataproc derives the names
+   * Output only. The list of instance names. Dataproc derives the names
    * from `cluster_name`, `num_instances`, and the instance group.
    * 
* @@ -271,7 +271,7 @@ public java.lang.String getInstanceNames(int index) { * * *
-   * Output only. The list of instance names. Cloud Dataproc derives the names
+   * Output only. The list of instance names. Dataproc derives the names
    * from `cluster_name`, `num_instances`, and the instance group.
    * 
* @@ -346,7 +346,7 @@ public com.google.protobuf.ByteString getImageUriBytes() { * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -377,7 +377,7 @@ public java.lang.String getMachineTypeUri() { * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -613,7 +613,7 @@ public com.google.cloud.dataproc.v1.AcceleratorConfigOrBuilder getAcceleratorsOr * *
    * Optional. Specifies the minimum cpu platform for the Instance Group.
-   * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+   * See [Dataproc&rarr;Minimum CPU Platform]
    * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* @@ -637,7 +637,7 @@ public java.lang.String getMinCpuPlatform() { * *
    * Optional. Specifies the minimum cpu platform for the Instance Group.
-   * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+   * See [Dataproc&rarr;Minimum CPU Platform]
    * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* @@ -1251,7 +1251,7 @@ private void ensureInstanceNamesIsMutable() { * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1266,7 +1266,7 @@ public com.google.protobuf.ProtocolStringList getInstanceNamesList() { * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1281,7 +1281,7 @@ public int getInstanceNamesCount() { * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1297,7 +1297,7 @@ public java.lang.String getInstanceNames(int index) { * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1313,7 +1313,7 @@ public com.google.protobuf.ByteString getInstanceNamesBytes(int index) { * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1336,7 +1336,7 @@ public Builder setInstanceNames(int index, java.lang.String value) { * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1358,7 +1358,7 @@ public Builder addInstanceNames(java.lang.String value) { * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1377,7 +1377,7 @@ public Builder addAllInstanceNames(java.lang.Iterable values) * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1395,7 +1395,7 @@ public Builder clearInstanceNames() { * * *
-     * Output only. The list of instance names. Cloud Dataproc derives the names
+     * Output only. The list of instance names. Dataproc derives the names
      * from `cluster_name`, `num_instances`, and the instance group.
      * 
* @@ -1541,7 +1541,7 @@ public Builder setImageUriBytes(com.google.protobuf.ByteString value) { * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -1572,7 +1572,7 @@ public java.lang.String getMachineTypeUri() { * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -1603,7 +1603,7 @@ public com.google.protobuf.ByteString getMachineTypeUriBytes() { * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -1633,7 +1633,7 @@ public Builder setMachineTypeUri(java.lang.String value) { * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -1659,7 +1659,7 @@ public Builder clearMachineTypeUri() { * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -2578,7 +2578,7 @@ public com.google.cloud.dataproc.v1.AcceleratorConfig.Builder addAcceleratorsBui * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+     * See [Dataproc&rarr;Minimum CPU Platform]
      * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* @@ -2602,7 +2602,7 @@ public java.lang.String getMinCpuPlatform() { * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+     * See [Dataproc&rarr;Minimum CPU Platform]
      * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* @@ -2626,7 +2626,7 @@ public com.google.protobuf.ByteString getMinCpuPlatformBytes() { * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+     * See [Dataproc&rarr;Minimum CPU Platform]
      * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* @@ -2649,7 +2649,7 @@ public Builder setMinCpuPlatform(java.lang.String value) { * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+     * See [Dataproc&rarr;Minimum CPU Platform]
      * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* @@ -2668,7 +2668,7 @@ public Builder clearMinCpuPlatform() { * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+     * See [Dataproc&rarr;Minimum CPU Platform]
      * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java index 530d3fe2..f7168e00 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java @@ -41,7 +41,7 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Output only. The list of instance names. Cloud Dataproc derives the names
+   * Output only. The list of instance names. Dataproc derives the names
    * from `cluster_name`, `num_instances`, and the instance group.
    * 
* @@ -54,7 +54,7 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Output only. The list of instance names. Cloud Dataproc derives the names
+   * Output only. The list of instance names. Dataproc derives the names
    * from `cluster_name`, `num_instances`, and the instance group.
    * 
* @@ -67,7 +67,7 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Output only. The list of instance names. Cloud Dataproc derives the names
+   * Output only. The list of instance names. Dataproc derives the names
    * from `cluster_name`, `num_instances`, and the instance group.
    * 
* @@ -81,7 +81,7 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Output only. The list of instance names. Cloud Dataproc derives the names
+   * Output only. The list of instance names. Dataproc derives the names
    * from `cluster_name`, `num_instances`, and the instance group.
    * 
* @@ -130,7 +130,7 @@ public interface InstanceGroupConfigOrBuilder * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -151,7 +151,7 @@ public interface InstanceGroupConfigOrBuilder * * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` * * `n1-standard-2` - * **Auto Zone Exception**: If you are using the Cloud Dataproc + * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type @@ -338,7 +338,7 @@ public interface InstanceGroupConfigOrBuilder * *
    * Optional. Specifies the minimum cpu platform for the Instance Group.
-   * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+   * See [Dataproc&rarr;Minimum CPU Platform]
    * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* @@ -352,7 +352,7 @@ public interface InstanceGroupConfigOrBuilder * *
    * Optional. Specifies the minimum cpu platform for the Instance Group.
-   * See [Cloud Dataproc&rarr;Minimum CPU Platform]
+   * See [Dataproc&rarr;Minimum CPU Platform]
    * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java index e74438b7..c2b5eaaf 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java @@ -22,7 +22,7 @@ * * *
- * A Cloud Dataproc job resource.
+ * A Dataproc job resource.
  * 
* * Protobuf type {@code google.cloud.dataproc.v1.Job} @@ -1712,7 +1712,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc job resource.
+   * A Dataproc job resource.
    * 
* * Protobuf type {@code google.cloud.dataproc.v1.Job} diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacement.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacement.java index 8c1f3ce6..cb01f0ff 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacement.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacement.java @@ -22,7 +22,7 @@ * * *
- * Cloud Dataproc job config.
+ * Dataproc job config.
  * 
* * Protobuf type {@code google.cloud.dataproc.v1.JobPlacement} @@ -172,7 +172,7 @@ public com.google.protobuf.ByteString getClusterNameBytes() { * * *
-   * Output only. A cluster UUID generated by the Cloud Dataproc service when
+   * Output only. A cluster UUID generated by the Dataproc service when
    * the job is submitted.
    * 
* @@ -195,7 +195,7 @@ public java.lang.String getClusterUuid() { * * *
-   * Output only. A cluster UUID generated by the Cloud Dataproc service when
+   * Output only. A cluster UUID generated by the Dataproc service when
    * the job is submitted.
    * 
* @@ -387,7 +387,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * Cloud Dataproc job config.
+   * Dataproc job config.
    * 
* * Protobuf type {@code google.cloud.dataproc.v1.JobPlacement} @@ -658,7 +658,7 @@ public Builder setClusterNameBytes(com.google.protobuf.ByteString value) { * * *
-     * Output only. A cluster UUID generated by the Cloud Dataproc service when
+     * Output only. A cluster UUID generated by the Dataproc service when
      * the job is submitted.
      * 
* @@ -681,7 +681,7 @@ public java.lang.String getClusterUuid() { * * *
-     * Output only. A cluster UUID generated by the Cloud Dataproc service when
+     * Output only. A cluster UUID generated by the Dataproc service when
      * the job is submitted.
      * 
* @@ -704,7 +704,7 @@ public com.google.protobuf.ByteString getClusterUuidBytes() { * * *
-     * Output only. A cluster UUID generated by the Cloud Dataproc service when
+     * Output only. A cluster UUID generated by the Dataproc service when
      * the job is submitted.
      * 
* @@ -726,7 +726,7 @@ public Builder setClusterUuid(java.lang.String value) { * * *
-     * Output only. A cluster UUID generated by the Cloud Dataproc service when
+     * Output only. A cluster UUID generated by the Dataproc service when
      * the job is submitted.
      * 
* @@ -744,7 +744,7 @@ public Builder clearClusterUuid() { * * *
-     * Output only. A cluster UUID generated by the Cloud Dataproc service when
+     * Output only. A cluster UUID generated by the Dataproc service when
      * the job is submitted.
      * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacementOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacementOrBuilder.java index af633d1f..73e5b9f1 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacementOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacementOrBuilder.java @@ -52,7 +52,7 @@ public interface JobPlacementOrBuilder * * *
-   * Output only. A cluster UUID generated by the Cloud Dataproc service when
+   * Output only. A cluster UUID generated by the Dataproc service when
    * the job is submitted.
    * 
* @@ -65,7 +65,7 @@ public interface JobPlacementOrBuilder * * *
-   * Output only. A cluster UUID generated by the Cloud Dataproc service when
+   * Output only. A cluster UUID generated by the Dataproc service when
    * the job is submitted.
    * 
* diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobStatus.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobStatus.java index a4f01e88..f5f41643 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobStatus.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobStatus.java @@ -22,7 +22,7 @@ * * *
- * Cloud Dataproc job status.
+ * Dataproc job status.
  * 
* * Protobuf type {@code google.cloud.dataproc.v1.JobStatus} @@ -506,7 +506,7 @@ public enum Substate implements com.google.protobuf.ProtocolMessageEnum { * *
      * The agent-reported status is out of date, which may be caused by a
-     * loss of communication between the agent and Cloud Dataproc. If the
+     * loss of communication between the agent and Dataproc. If the
      * agent does not send a timely update, the job will fail.
      * Applies to RUNNING state.
      * 
@@ -556,7 +556,7 @@ public enum Substate implements com.google.protobuf.ProtocolMessageEnum { * *
      * The agent-reported status is out of date, which may be caused by a
-     * loss of communication between the agent and Cloud Dataproc. If the
+     * loss of communication between the agent and Dataproc. If the
      * agent does not send a timely update, the job will fail.
      * Applies to RUNNING state.
      * 
@@ -1022,7 +1022,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * Cloud Dataproc job status.
+   * Dataproc job status.
    * 
* * Protobuf type {@code google.cloud.dataproc.v1.JobStatus} diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequest.java index bcab2bfa..8467e8a2 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequest.java @@ -195,7 +195,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; @@ -217,7 +217,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; @@ -871,7 +871,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; @@ -893,7 +893,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; @@ -915,7 +915,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; @@ -936,7 +936,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; @@ -953,7 +953,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequestOrBuilder.java index 8c3bbf02..8c15274b 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface ListClustersRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface ListClustersRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 4 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequest.java index 84ee1bf3..4c03567b 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequest.java @@ -368,7 +368,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; @@ -390,7 +390,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; @@ -1161,7 +1161,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; @@ -1183,7 +1183,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; @@ -1205,7 +1205,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; @@ -1226,7 +1226,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; @@ -1243,7 +1243,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequestOrBuilder.java index 80bcea4a..ddcd2383 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface ListJobsRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface ListJobsRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 6 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJob.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJob.java index 3d3c70e7..f7914da3 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJob.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJob.java @@ -22,7 +22,7 @@ * * *
- * A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/)
+ * A Dataproc job for running [Apache Pig](https://pig.apache.org/)
  * queries on YARN.
  * 
* @@ -512,7 +512,7 @@ public int getPropertiesCount() { * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
@@ -536,7 +536,7 @@ public java.util.Map getProperties() { * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
@@ -552,7 +552,7 @@ public java.util.Map getPropertiesMap() { * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
@@ -573,7 +573,7 @@ public java.lang.String getPropertiesOrDefault( * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
@@ -970,7 +970,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/)
+   * A Dataproc job for running [Apache Pig](https://pig.apache.org/)
    * queries on YARN.
    * 
* @@ -1822,7 +1822,7 @@ public int getPropertiesCount() { * *
      * Optional. A mapping of property names to values, used to configure Pig.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/pig/conf/pig.properties, and classes in user code.
      * 
@@ -1846,7 +1846,7 @@ public java.util.Map getProperties() { * *
      * Optional. A mapping of property names to values, used to configure Pig.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/pig/conf/pig.properties, and classes in user code.
      * 
@@ -1862,7 +1862,7 @@ public java.util.Map getPropertiesMap() { * *
      * Optional. A mapping of property names to values, used to configure Pig.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/pig/conf/pig.properties, and classes in user code.
      * 
@@ -1883,7 +1883,7 @@ public java.lang.String getPropertiesOrDefault( * *
      * Optional. A mapping of property names to values, used to configure Pig.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/pig/conf/pig.properties, and classes in user code.
      * 
@@ -1911,7 +1911,7 @@ public Builder clearProperties() { * *
      * Optional. A mapping of property names to values, used to configure Pig.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/pig/conf/pig.properties, and classes in user code.
      * 
@@ -1936,7 +1936,7 @@ public java.util.Map getMutableProperties() * *
      * Optional. A mapping of property names to values, used to configure Pig.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/pig/conf/pig.properties, and classes in user code.
      * 
@@ -1959,7 +1959,7 @@ public Builder putProperties(java.lang.String key, java.lang.String value) { * *
      * Optional. A mapping of property names to values, used to configure Pig.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
      * /etc/pig/conf/pig.properties, and classes in user code.
      * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJobOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJobOrBuilder.java index 3d9517b6..66879a60 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJobOrBuilder.java @@ -167,7 +167,7 @@ public interface PigJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
@@ -181,7 +181,7 @@ public interface PigJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
@@ -198,7 +198,7 @@ public interface PigJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
@@ -212,7 +212,7 @@ public interface PigJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
@@ -226,7 +226,7 @@ public interface PigJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Pig.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
    * /etc/pig/conf/pig.properties, and classes in user code.
    * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJob.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJob.java index c503994f..250dff79 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJob.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJob.java @@ -22,7 +22,7 @@ * * *
- * A Cloud Dataproc job for running
+ * A Dataproc job for running
  * [Apache
  * PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
  * applications on YARN.
@@ -635,7 +635,7 @@ public int getPropertiesCount() {
    *
    * 
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -659,7 +659,7 @@ public java.util.Map getProperties() { * *
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -675,7 +675,7 @@ public java.util.Map getPropertiesMap() { * *
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -696,7 +696,7 @@ public java.lang.String getPropertiesOrDefault( * *
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -1039,7 +1039,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc job for running
+   * A Dataproc job for running
    * [Apache
    * PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
    * applications on YARN.
@@ -2359,7 +2359,7 @@ public int getPropertiesCount() {
      *
      * 
      * Optional. A mapping of property names to values, used to configure PySpark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2383,7 +2383,7 @@ public java.util.Map getProperties() { * *
      * Optional. A mapping of property names to values, used to configure PySpark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2399,7 +2399,7 @@ public java.util.Map getPropertiesMap() { * *
      * Optional. A mapping of property names to values, used to configure PySpark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2420,7 +2420,7 @@ public java.lang.String getPropertiesOrDefault( * *
      * Optional. A mapping of property names to values, used to configure PySpark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2448,7 +2448,7 @@ public Builder clearProperties() { * *
      * Optional. A mapping of property names to values, used to configure PySpark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2473,7 +2473,7 @@ public java.util.Map getMutableProperties() * *
      * Optional. A mapping of property names to values, used to configure PySpark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2496,7 +2496,7 @@ public Builder putProperties(java.lang.String key, java.lang.String value) { * *
      * Optional. A mapping of property names to values, used to configure PySpark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJobOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJobOrBuilder.java index e016d502..b9b55e8c 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJobOrBuilder.java @@ -334,7 +334,7 @@ public interface PySparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -348,7 +348,7 @@ public interface PySparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -365,7 +365,7 @@ public interface PySparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -379,7 +379,7 @@ public interface PySparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -393,7 +393,7 @@ public interface PySparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure PySpark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java index d9d054b6..d9f1bb03 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java @@ -174,7 +174,7 @@ protected com.google.protobuf.MapField internalGetMapField(int number) { * *
    * Optional. The version of software inside the cluster. It must be one of the
-   * supported [Cloud Dataproc
+   * supported [Dataproc
    * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
@@ -202,7 +202,7 @@ public java.lang.String getImageVersion() {
    *
    * 
    * Optional. The version of software inside the cluster. It must be one of the
-   * supported [Cloud Dataproc
+   * supported [Dataproc
    * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
@@ -890,7 +890,7 @@ public Builder mergeFrom(
      *
      * 
      * Optional. The version of software inside the cluster. It must be one of the
-     * supported [Cloud Dataproc
+     * supported [Dataproc
      * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
@@ -918,7 +918,7 @@ public java.lang.String getImageVersion() {
      *
      * 
      * Optional. The version of software inside the cluster. It must be one of the
-     * supported [Cloud Dataproc
+     * supported [Dataproc
      * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
@@ -946,7 +946,7 @@ public com.google.protobuf.ByteString getImageVersionBytes() {
      *
      * 
      * Optional. The version of software inside the cluster. It must be one of the
-     * supported [Cloud Dataproc
+     * supported [Dataproc
      * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
@@ -973,7 +973,7 @@ public Builder setImageVersion(java.lang.String value) {
      *
      * 
      * Optional. The version of software inside the cluster. It must be one of the
-     * supported [Cloud Dataproc
+     * supported [Dataproc
      * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
@@ -996,7 +996,7 @@ public Builder clearImageVersion() {
      *
      * 
      * Optional. The version of software inside the cluster. It must be one of the
-     * supported [Cloud Dataproc
+     * supported [Dataproc
      * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java
index 49903dfa..dd457410 100644
--- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java
+++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java
@@ -28,7 +28,7 @@ public interface SoftwareConfigOrBuilder
    *
    * 
    * Optional. The version of software inside the cluster. It must be one of the
-   * supported [Cloud Dataproc
+   * supported [Dataproc
    * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
@@ -46,7 +46,7 @@ public interface SoftwareConfigOrBuilder
    *
    * 
    * Optional. The version of software inside the cluster. It must be one of the
-   * supported [Cloud Dataproc
+   * supported [Dataproc
    * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJob.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJob.java
index bb939d70..f753fe43 100644
--- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJob.java
+++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJob.java
@@ -22,7 +22,7 @@
  *
  *
  * 
- * A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/)
+ * A Dataproc job for running [Apache Spark](http://spark.apache.org/)
  * applications on YARN.
  * 
* @@ -676,7 +676,7 @@ public int getPropertiesCount() { * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -700,7 +700,7 @@ public java.util.Map getProperties() { * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -716,7 +716,7 @@ public java.util.Map getPropertiesMap() { * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -737,7 +737,7 @@ public java.lang.String getPropertiesOrDefault( * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -1089,7 +1089,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/)
+   * A Dataproc job for running [Apache Spark](http://spark.apache.org/)
    * applications on YARN.
    * 
* @@ -2386,7 +2386,7 @@ public int getPropertiesCount() { * *
      * Optional. A mapping of property names to values, used to configure Spark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2410,7 +2410,7 @@ public java.util.Map getProperties() { * *
      * Optional. A mapping of property names to values, used to configure Spark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2426,7 +2426,7 @@ public java.util.Map getPropertiesMap() { * *
      * Optional. A mapping of property names to values, used to configure Spark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2447,7 +2447,7 @@ public java.lang.String getPropertiesOrDefault( * *
      * Optional. A mapping of property names to values, used to configure Spark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2475,7 +2475,7 @@ public Builder clearProperties() { * *
      * Optional. A mapping of property names to values, used to configure Spark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2500,7 +2500,7 @@ public java.util.Map getMutableProperties() * *
      * Optional. A mapping of property names to values, used to configure Spark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
@@ -2523,7 +2523,7 @@ public Builder putProperties(java.lang.String key, java.lang.String value) { * *
      * Optional. A mapping of property names to values, used to configure Spark.
-     * Properties that conflict with values set by the Cloud Dataproc API may be
+     * Properties that conflict with values set by the Dataproc API may be
      * overwritten. Can include properties set in
      * /etc/spark/conf/spark-defaults.conf and classes in user code.
      * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJobOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJobOrBuilder.java index a5112516..46f13101 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJobOrBuilder.java @@ -308,7 +308,7 @@ public interface SparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -322,7 +322,7 @@ public interface SparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -339,7 +339,7 @@ public interface SparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -353,7 +353,7 @@ public interface SparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
@@ -367,7 +367,7 @@ public interface SparkJobOrBuilder * *
    * Optional. A mapping of property names to values, used to configure Spark.
-   * Properties that conflict with values set by the Cloud Dataproc API may be
+   * Properties that conflict with values set by the Dataproc API may be
    * overwritten. Can include properties set in
    * /etc/spark/conf/spark-defaults.conf and classes in user code.
    * 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJob.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJob.java index a2eb9630..f8a69881 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJob.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJob.java @@ -22,7 +22,7 @@ * * *
- * A Cloud Dataproc job for running [Apache Spark
+ * A Dataproc job for running [Apache Spark
  * SQL](http://spark.apache.org/sql/) queries.
  * 
* @@ -489,7 +489,7 @@ public int getPropertiesCount() { *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -512,7 +512,7 @@ public java.util.Map getProperties() { *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -527,7 +527,7 @@ public java.util.Map getPropertiesMap() { *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -547,7 +547,7 @@ public java.lang.String getPropertiesOrDefault( *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -930,7 +930,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc job for running [Apache Spark
+   * A Dataproc job for running [Apache Spark
    * SQL](http://spark.apache.org/sql/) queries.
    * 
* @@ -1721,7 +1721,7 @@ public int getPropertiesCount() { *
      * Optional. A mapping of property names to values, used to configure
      * Spark SQL's SparkConf. Properties that conflict with values set by the
-     * Cloud Dataproc API may be overwritten.
+     * Dataproc API may be overwritten.
      * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -1744,7 +1744,7 @@ public java.util.Map getProperties() { *
      * Optional. A mapping of property names to values, used to configure
      * Spark SQL's SparkConf. Properties that conflict with values set by the
-     * Cloud Dataproc API may be overwritten.
+     * Dataproc API may be overwritten.
      * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -1759,7 +1759,7 @@ public java.util.Map getPropertiesMap() { *
      * Optional. A mapping of property names to values, used to configure
      * Spark SQL's SparkConf. Properties that conflict with values set by the
-     * Cloud Dataproc API may be overwritten.
+     * Dataproc API may be overwritten.
      * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -1779,7 +1779,7 @@ public java.lang.String getPropertiesOrDefault( *
      * Optional. A mapping of property names to values, used to configure
      * Spark SQL's SparkConf. Properties that conflict with values set by the
-     * Cloud Dataproc API may be overwritten.
+     * Dataproc API may be overwritten.
      * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -1806,7 +1806,7 @@ public Builder clearProperties() { *
      * Optional. A mapping of property names to values, used to configure
      * Spark SQL's SparkConf. Properties that conflict with values set by the
-     * Cloud Dataproc API may be overwritten.
+     * Dataproc API may be overwritten.
      * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -1830,7 +1830,7 @@ public java.util.Map getMutableProperties() *
      * Optional. A mapping of property names to values, used to configure
      * Spark SQL's SparkConf. Properties that conflict with values set by the
-     * Cloud Dataproc API may be overwritten.
+     * Dataproc API may be overwritten.
      * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -1852,7 +1852,7 @@ public Builder putProperties(java.lang.String key, java.lang.String value) { *
      * Optional. A mapping of property names to values, used to configure
      * Spark SQL's SparkConf. Properties that conflict with values set by the
-     * Cloud Dataproc API may be overwritten.
+     * Dataproc API may be overwritten.
      * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJobOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJobOrBuilder.java index 0d5a1878..05cce2f3 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJobOrBuilder.java @@ -153,7 +153,7 @@ public interface SparkSqlJobOrBuilder *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -166,7 +166,7 @@ public interface SparkSqlJobOrBuilder *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -182,7 +182,7 @@ public interface SparkSqlJobOrBuilder *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -195,7 +195,7 @@ public interface SparkSqlJobOrBuilder *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; @@ -208,7 +208,7 @@ public interface SparkSqlJobOrBuilder *
    * Optional. A mapping of property names to values, used to configure
    * Spark SQL's SparkConf. Properties that conflict with values set by the
-   * Cloud Dataproc API may be overwritten.
+   * Dataproc API may be overwritten.
    * 
* * map<string, string> properties = 4 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequest.java index 4b121373..39c69552 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequest.java @@ -197,7 +197,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -219,7 +219,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -836,7 +836,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -858,7 +858,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -880,7 +880,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -901,7 +901,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -918,7 +918,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequestOrBuilder.java index f11e53c7..10ba2896 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface SubmitJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface SubmitJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 3 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequest.java index 409a0746..1f35de61 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequest.java @@ -236,7 +236,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; @@ -258,7 +258,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; @@ -1258,7 +1258,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; @@ -1280,7 +1280,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; @@ -1302,7 +1302,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; @@ -1323,7 +1323,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; @@ -1340,7 +1340,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequestOrBuilder.java index d024b9d5..e700ab98 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface UpdateClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface UpdateClusterRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 5 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequest.java index 77e7f194..2725d809 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequest.java @@ -212,7 +212,7 @@ public com.google.protobuf.ByteString getProjectIdBytes() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; @@ -234,7 +234,7 @@ public java.lang.String getRegion() { * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; @@ -924,7 +924,7 @@ public Builder setProjectIdBytes(com.google.protobuf.ByteString value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; @@ -946,7 +946,7 @@ public java.lang.String getRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; @@ -968,7 +968,7 @@ public com.google.protobuf.ByteString getRegionBytes() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; @@ -989,7 +989,7 @@ public Builder setRegion(java.lang.String value) { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; @@ -1006,7 +1006,7 @@ public Builder clearRegion() { * * *
-     * Required. The Cloud Dataproc region in which to handle the request.
+     * Required. The Dataproc region in which to handle the request.
      * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequestOrBuilder.java index 86c23574..42517637 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequestOrBuilder.java @@ -54,7 +54,7 @@ public interface UpdateJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; @@ -66,7 +66,7 @@ public interface UpdateJobRequestOrBuilder * * *
-   * Required. The Cloud Dataproc region in which to handle the request.
+   * Required. The Dataproc region in which to handle the request.
    * 
* * string region = 2 [(.google.api.field_behavior) = REQUIRED]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequest.java index 85312213..313f7842 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequest.java @@ -129,7 +129,7 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return Whether the template field is set. @@ -146,7 +146,7 @@ public boolean hasTemplate() { *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return The template. @@ -165,7 +165,7 @@ public com.google.cloud.dataproc.v1.WorkflowTemplate getTemplate() { *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public com.google.cloud.dataproc.v1.WorkflowTemplateOrBuilder getTemplateOrBuilder() { @@ -518,7 +518,7 @@ public Builder mergeFrom( *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return Whether the template field is set. @@ -535,7 +535,7 @@ public boolean hasTemplate() { *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return The template. @@ -558,7 +558,7 @@ public com.google.cloud.dataproc.v1.WorkflowTemplate getTemplate() { *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public Builder setTemplate(com.google.cloud.dataproc.v1.WorkflowTemplate value) { @@ -583,7 +583,7 @@ public Builder setTemplate(com.google.cloud.dataproc.v1.WorkflowTemplate value) *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public Builder setTemplate( @@ -606,7 +606,7 @@ public Builder setTemplate( *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public Builder mergeTemplate(com.google.cloud.dataproc.v1.WorkflowTemplate value) { @@ -635,7 +635,7 @@ public Builder mergeTemplate(com.google.cloud.dataproc.v1.WorkflowTemplate value *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public Builder clearTemplate() { @@ -658,7 +658,7 @@ public Builder clearTemplate() { *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public com.google.cloud.dataproc.v1.WorkflowTemplate.Builder getTemplateBuilder() { @@ -675,7 +675,7 @@ public com.google.cloud.dataproc.v1.WorkflowTemplate.Builder getTemplateBuilder( *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public com.google.cloud.dataproc.v1.WorkflowTemplateOrBuilder getTemplateOrBuilder() { @@ -696,7 +696,7 @@ public com.google.cloud.dataproc.v1.WorkflowTemplateOrBuilder getTemplateOrBuild *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ private com.google.protobuf.SingleFieldBuilderV3< diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequestOrBuilder.java index 80e4045a..56064397 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequestOrBuilder.java @@ -32,7 +32,7 @@ public interface UpdateWorkflowTemplateRequestOrBuilder *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return Whether the template field is set. @@ -47,7 +47,7 @@ public interface UpdateWorkflowTemplateRequestOrBuilder *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return The template. @@ -62,7 +62,7 @@ public interface UpdateWorkflowTemplateRequestOrBuilder *
* * - * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ com.google.cloud.dataproc.v1.WorkflowTemplateOrBuilder getTemplateOrBuilder(); diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowMetadata.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowMetadata.java index 9af90628..e379e7a7 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowMetadata.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowMetadata.java @@ -22,7 +22,7 @@ * * *
- * A Cloud Dataproc workflow template resource.
+ * A Dataproc workflow template resource.
  * 
* * Protobuf type {@code google.cloud.dataproc.v1.WorkflowMetadata} @@ -1258,7 +1258,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc workflow template resource.
+   * A Dataproc workflow template resource.
    * 
* * Protobuf type {@code google.cloud.dataproc.v1.WorkflowMetadata} diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplate.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplate.java index f53d44b8..d4f31881 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplate.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplate.java @@ -22,7 +22,7 @@ * * *
- * A Cloud Dataproc workflow template resource.
+ * A Dataproc workflow template resource.
  * 
* * Protobuf type {@code google.cloud.dataproc.v1.WorkflowTemplate} @@ -1035,7 +1035,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Cloud Dataproc workflow template resource.
+   * A Dataproc workflow template resource.
    * 
* * Protobuf type {@code google.cloud.dataproc.v1.WorkflowTemplate} diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java index de205997..b77bb4c2 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java @@ -260,77 +260,76 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "oc.googleapis.com/WorkflowTemplate\022A\n\010te" + "mplate\030\002 \001(\0132*.google.cloud.dataproc.v1." + "WorkflowTemplateB\003\340A\002\022\027\n\nrequest_id\030\003 \001(" - + "\tB\003\340A\001\"\217\001\n\035UpdateWorkflowTemplateRequest" - + "\022n\n\010template\030\001 \001(\0132*.google.cloud.datapr" - + "oc.v1.WorkflowTemplateB0\340A\002\372A*\n(dataproc" - + ".googleapis.com/WorkflowTemplate\"\221\001\n\034Lis" - + "tWorkflowTemplatesRequest\022@\n\006parent\030\001 \001(" - + "\tB0\340A\002\372A*\022(dataproc.googleapis.com/Workf" - + "lowTemplate\022\026\n\tpage_size\030\002 \001(\005B\003\340A\001\022\027\n\np" - + "age_token\030\003 \001(\tB\003\340A\001\"\201\001\n\035ListWorkflowTem" - + "platesResponse\022B\n\ttemplates\030\001 \003(\0132*.goog" - + "le.cloud.dataproc.v1.WorkflowTemplateB\003\340" - + "A\003\022\034\n\017next_page_token\030\002 \001(\tB\003\340A\003\"u\n\035Dele" - + "teWorkflowTemplateRequest\022>\n\004name\030\001 \001(\tB" - + "0\340A\002\372A*\n(dataproc.googleapis.com/Workflo" - + "wTemplate\022\024\n\007version\030\002 \001(\005B\003\340A\0012\346\020\n\027Work" - + "flowTemplateService\022\233\002\n\026CreateWorkflowTe" - + "mplate\0227.google.cloud.dataproc.v1.Create" - + "WorkflowTemplateRequest\032*.google.cloud.d" - + "ataproc.v1.WorkflowTemplate\"\233\001\202\323\344\223\002\202\001\"5/" + + "\tB\003\340A\001\"b\n\035UpdateWorkflowTemplateRequest\022" + + "A\n\010template\030\001 \001(\0132*.google.cloud.datapro" + + "c.v1.WorkflowTemplateB\003\340A\002\"\221\001\n\034ListWorkf" + + "lowTemplatesRequest\022@\n\006parent\030\001 \001(\tB0\340A\002" + + "\372A*\022(dataproc.googleapis.com/WorkflowTem" + + "plate\022\026\n\tpage_size\030\002 \001(\005B\003\340A\001\022\027\n\npage_to" + + "ken\030\003 \001(\tB\003\340A\001\"\201\001\n\035ListWorkflowTemplates" + + "Response\022B\n\ttemplates\030\001 \003(\0132*.google.clo" + + "ud.dataproc.v1.WorkflowTemplateB\003\340A\003\022\034\n\017" + + "next_page_token\030\002 \001(\tB\003\340A\003\"u\n\035DeleteWork" + + "flowTemplateRequest\022>\n\004name\030\001 \001(\tB0\340A\002\372A" + + "*\n(dataproc.googleapis.com/WorkflowTempl" + + "ate\022\024\n\007version\030\002 \001(\005B\003\340A\0012\346\020\n\027WorkflowTe" + + "mplateService\022\233\002\n\026CreateWorkflowTemplate" + + "\0227.google.cloud.dataproc.v1.CreateWorkfl" + + "owTemplateRequest\032*.google.cloud.datapro" + + "c.v1.WorkflowTemplate\"\233\001\202\323\344\223\002\202\001\"5/v1/{pa" + + "rent=projects/*/locations/*}/workflowTem" + + "plates:\010templateZ?\"3/v1/{parent=projects" + + "/*/regions/*}/workflowTemplates:\010templat" + + "e\332A\017parent,template\022\364\001\n\023GetWorkflowTempl" + + "ate\0224.google.cloud.dataproc.v1.GetWorkfl" + + "owTemplateRequest\032*.google.cloud.datapro" + + "c.v1.WorkflowTemplate\"{\202\323\344\223\002n\0225/v1/{name" + + "=projects/*/locations/*/workflowTemplate" + + "s/*}Z5\0223/v1/{name=projects/*/regions/*/w" + + "orkflowTemplates/*}\332A\004name\022\325\002\n\033Instantia" + + "teWorkflowTemplate\022<.google.cloud.datapr" + + "oc.v1.InstantiateWorkflowTemplateRequest" + + "\032\035.google.longrunning.Operation\"\330\001\202\323\344\223\002\214" + + "\001\"A/v1/{name=projects/*/locations/*/work" + + "flowTemplates/*}:instantiate:\001*ZD\"?/v1/{" + + "name=projects/*/regions/*/workflowTempla" + + "tes/*}:instantiate:\001*\332A\004name\332A\017name,para" + + "meters\312A)\n\025google.protobuf.Empty\022\020Workfl" + + "owMetadata\022\364\002\n!InstantiateInlineWorkflow" + + "Template\022B.google.cloud.dataproc.v1.Inst" + + "antiateInlineWorkflowTemplateRequest\032\035.g" + + "oogle.longrunning.Operation\"\353\001\202\323\344\223\002\246\001\"G/" + "v1/{parent=projects/*/locations/*}/workf" - + "lowTemplates:\010templateZ?\"3/v1/{parent=pr" - + "ojects/*/regions/*}/workflowTemplates:\010t" - + "emplate\332A\017parent,template\022\364\001\n\023GetWorkflo" - + "wTemplate\0224.google.cloud.dataproc.v1.Get" - + "WorkflowTemplateRequest\032*.google.cloud.d" - + "ataproc.v1.WorkflowTemplate\"{\202\323\344\223\002n\0225/v1" - + "/{name=projects/*/locations/*/workflowTe" - + "mplates/*}Z5\0223/v1/{name=projects/*/regio" - + "ns/*/workflowTemplates/*}\332A\004name\022\325\002\n\033Ins" - + "tantiateWorkflowTemplate\022<.google.cloud." - + "dataproc.v1.InstantiateWorkflowTemplateR" - + "equest\032\035.google.longrunning.Operation\"\330\001" - + "\202\323\344\223\002\214\001\"A/v1/{name=projects/*/locations/" - + "*/workflowTemplates/*}:instantiate:\001*ZD\"" - + "?/v1/{name=projects/*/regions/*/workflow" - + "Templates/*}:instantiate:\001*\332A\004name\332A\017nam" - + "e,parameters\312A)\n\025google.protobuf.Empty\022\020" - + "WorkflowMetadata\022\364\002\n!InstantiateInlineWo" - + "rkflowTemplate\022B.google.cloud.dataproc.v" - + "1.InstantiateInlineWorkflowTemplateReque" - + "st\032\035.google.longrunning.Operation\"\353\001\202\323\344\223" - + "\002\246\001\"G/v1/{parent=projects/*/locations/*}" - + "/workflowTemplates:instantiateInline:\010te" - + "mplateZQ\"E/v1/{parent=projects/*/regions" - + "/*}/workflowTemplates:instantiateInline:" - + "\010template\332A\017parent,template\312A)\n\025google.p" - + "rotobuf.Empty\022\020WorkflowMetadata\022\246\002\n\026Upda" - + "teWorkflowTemplate\0227.google.cloud.datapr" - + "oc.v1.UpdateWorkflowTemplateRequest\032*.go" - + "ogle.cloud.dataproc.v1.WorkflowTemplate\"" - + "\246\001\202\323\344\223\002\224\001\032>/v1/{template.name=projects/*" - + "/locations/*/workflowTemplates/*}:\010templ" - + "ateZH\032/v1/{template.name=projects/*/locat" + + "ions/*/workflowTemplates/*}:\010templateZH\032" + + " driver_log_levels = 2; } -// A Cloud Dataproc job for running +// A Dataproc job for running // [Apache Hadoop // MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html) // jobs on [Apache Hadoop @@ -176,7 +175,7 @@ message HadoopJob { repeated string archive_uris = 6 [(google.api.field_behavior) = OPTIONAL]; // Optional. A mapping of property names to values, used to configure Hadoop. - // Properties that conflict with values set by the Cloud Dataproc API may be + // Properties that conflict with values set by the Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site and // classes in user code. map properties = 7 [(google.api.field_behavior) = OPTIONAL]; @@ -185,7 +184,7 @@ message HadoopJob { LoggingConfig logging_config = 8 [(google.api.field_behavior) = OPTIONAL]; } -// A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/) +// A Dataproc job for running [Apache Spark](http://spark.apache.org/) // applications on YARN. message SparkJob { // Required. The specification of the main method to call to drive the job. @@ -221,7 +220,7 @@ message SparkJob { repeated string archive_uris = 6 [(google.api.field_behavior) = OPTIONAL]; // Optional. A mapping of property names to values, used to configure Spark. - // Properties that conflict with values set by the Cloud Dataproc API may be + // Properties that conflict with values set by the Dataproc API may be // overwritten. Can include properties set in // /etc/spark/conf/spark-defaults.conf and classes in user code. map properties = 7 [(google.api.field_behavior) = OPTIONAL]; @@ -230,7 +229,7 @@ message SparkJob { LoggingConfig logging_config = 8 [(google.api.field_behavior) = OPTIONAL]; } -// A Cloud Dataproc job for running +// A Dataproc job for running // [Apache // PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) // applications on YARN. @@ -261,7 +260,7 @@ message PySparkJob { repeated string archive_uris = 6 [(google.api.field_behavior) = OPTIONAL]; // Optional. A mapping of property names to values, used to configure PySpark. - // Properties that conflict with values set by the Cloud Dataproc API may be + // Properties that conflict with values set by the Dataproc API may be // overwritten. Can include properties set in // /etc/spark/conf/spark-defaults.conf and classes in user code. map properties = 7 [(google.api.field_behavior) = OPTIONAL]; @@ -289,7 +288,7 @@ message QueryList { repeated string queries = 1 [(google.api.field_behavior) = REQUIRED]; } -// A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/) +// A Dataproc job for running [Apache Hive](https://hive.apache.org/) // queries on YARN. message HiveJob { // Required. The sequence of Hive queries to execute, specified as either @@ -312,7 +311,7 @@ message HiveJob { map script_variables = 4 [(google.api.field_behavior) = OPTIONAL]; // Optional. A mapping of property names and values, used to configure Hive. - // Properties that conflict with values set by the Cloud Dataproc API may be + // Properties that conflict with values set by the Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, // /etc/hive/conf/hive-site.xml, and classes in user code. map properties = 5 [(google.api.field_behavior) = OPTIONAL]; @@ -323,7 +322,7 @@ message HiveJob { repeated string jar_file_uris = 6 [(google.api.field_behavior) = OPTIONAL]; } -// A Cloud Dataproc job for running [Apache Spark +// A Dataproc job for running [Apache Spark // SQL](http://spark.apache.org/sql/) queries. message SparkSqlJob { // Required. The sequence of Spark SQL queries to execute, specified as @@ -342,7 +341,7 @@ message SparkSqlJob { // Optional. A mapping of property names to values, used to configure // Spark SQL's SparkConf. Properties that conflict with values set by the - // Cloud Dataproc API may be overwritten. + // Dataproc API may be overwritten. map properties = 4 [(google.api.field_behavior) = OPTIONAL]; // Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH. @@ -352,7 +351,7 @@ message SparkSqlJob { LoggingConfig logging_config = 6 [(google.api.field_behavior) = OPTIONAL]; } -// A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/) +// A Dataproc job for running [Apache Pig](https://pig.apache.org/) // queries on YARN. message PigJob { // Required. The sequence of Pig queries to execute, specified as an HCFS @@ -375,7 +374,7 @@ message PigJob { map script_variables = 4 [(google.api.field_behavior) = OPTIONAL]; // Optional. A mapping of property names to values, used to configure Pig. - // Properties that conflict with values set by the Cloud Dataproc API may be + // Properties that conflict with values set by the Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, // /etc/pig/conf/pig.properties, and classes in user code. map properties = 5 [(google.api.field_behavior) = OPTIONAL]; @@ -388,17 +387,17 @@ message PigJob { LoggingConfig logging_config = 7 [(google.api.field_behavior) = OPTIONAL]; } -// Cloud Dataproc job config. +// Dataproc job config. message JobPlacement { // Required. The name of the cluster where the job will be submitted. string cluster_name = 1 [(google.api.field_behavior) = REQUIRED]; - // Output only. A cluster UUID generated by the Cloud Dataproc service when + // Output only. A cluster UUID generated by the Dataproc service when // the job is submitted. string cluster_uuid = 2 [(google.api.field_behavior) = OUTPUT_ONLY]; } -// Cloud Dataproc job status. +// Dataproc job status. message JobStatus { // The job state. enum State { @@ -456,7 +455,7 @@ message JobStatus { QUEUED = 2; // The agent-reported status is out of date, which may be caused by a - // loss of communication between the agent and Cloud Dataproc. If the + // loss of communication between the agent and Dataproc. If the // agent does not send a timely update, the job will fail. // // Applies to RUNNING state. @@ -549,7 +548,7 @@ message YarnApplication { string tracking_url = 4 [(google.api.field_behavior) = OPTIONAL]; } -// A Cloud Dataproc job resource. +// A Dataproc job resource. message Job { // Optional. The fully qualified reference to the job, which can be used to // obtain the equivalent REST path of the job resource. If this property @@ -642,7 +641,7 @@ message SubmitJobRequest { // belongs to. string project_id = 1 [(google.api.field_behavior) = REQUIRED]; - // Required. The Cloud Dataproc region in which to handle the request. + // Required. The Dataproc region in which to handle the request. string region = 3 [(google.api.field_behavior) = REQUIRED]; // Required. The job resource. @@ -668,7 +667,7 @@ message GetJobRequest { // belongs to. string project_id = 1 [(google.api.field_behavior) = REQUIRED]; - // Required. The Cloud Dataproc region in which to handle the request. + // Required. The Dataproc region in which to handle the request. string region = 3 [(google.api.field_behavior) = REQUIRED]; // Required. The job ID. @@ -694,7 +693,7 @@ message ListJobsRequest { // belongs to. string project_id = 1 [(google.api.field_behavior) = REQUIRED]; - // Required. The Cloud Dataproc region in which to handle the request. + // Required. The Dataproc region in which to handle the request. string region = 6 [(google.api.field_behavior) = REQUIRED]; // Optional. The number of results to return in each response. @@ -737,7 +736,7 @@ message UpdateJobRequest { // belongs to. string project_id = 1 [(google.api.field_behavior) = REQUIRED]; - // Required. The Cloud Dataproc region in which to handle the request. + // Required. The Dataproc region in which to handle the request. string region = 2 [(google.api.field_behavior) = REQUIRED]; // Required. The job ID. @@ -772,7 +771,7 @@ message CancelJobRequest { // belongs to. string project_id = 1 [(google.api.field_behavior) = REQUIRED]; - // Required. The Cloud Dataproc region in which to handle the request. + // Required. The Dataproc region in which to handle the request. string region = 3 [(google.api.field_behavior) = REQUIRED]; // Required. The job ID. @@ -785,7 +784,7 @@ message DeleteJobRequest { // belongs to. string project_id = 1 [(google.api.field_behavior) = REQUIRED]; - // Required. The Cloud Dataproc region in which to handle the request. + // Required. The Dataproc region in which to handle the request. string region = 3 [(google.api.field_behavior) = REQUIRED]; // Required. The job ID. diff --git a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto index 8976c42e..30b5ced4 100644 --- a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto +++ b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto @@ -33,7 +33,7 @@ option java_outer_classname = "WorkflowTemplatesProto"; option java_package = "com.google.cloud.dataproc.v1"; // The API interface for managing Workflow Templates in the -// Cloud Dataproc API. +// Dataproc API. service WorkflowTemplateService { option (google.api.default_host) = "dataproc.googleapis.com"; option (google.api.oauth_scopes) = "https://www.googleapis.com/auth/cloud-platform"; @@ -179,7 +179,7 @@ service WorkflowTemplateService { } } -// A Cloud Dataproc workflow template resource. +// A Dataproc workflow template resource. message WorkflowTemplate { option (google.api.resource) = { type: "dataproc.googleapis.com/WorkflowTemplate" @@ -458,7 +458,7 @@ message ValueValidation { repeated string values = 1 [(google.api.field_behavior) = REQUIRED]; } -// A Cloud Dataproc workflow template resource. +// A Dataproc workflow template resource. message WorkflowMetadata { // The operation state. enum State { @@ -709,10 +709,7 @@ message UpdateWorkflowTemplateRequest { // // The `template.version` field must match the current version. WorkflowTemplate template = 1 [ - (google.api.field_behavior) = REQUIRED, - (google.api.resource_reference) = { - type: "dataproc.googleapis.com/WorkflowTemplate" - } + (google.api.field_behavior) = REQUIRED ]; } diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java index 931de286..76b2d9a0 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java @@ -328,55 +328,54 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "\001(\tB\003\340A\001\022\023\n\006values\030\003 \003(\tB\003\340A\001\"_\n\004Type\022\024\n" + "\020TYPE_UNSPECIFIED\020\000\022\022\n\016NO_RESERVATION\020\001\022" + "\023\n\017ANY_RESERVATION\020\002\022\030\n\024SPECIFIC_RESERVA" - + "TION\020\0032\207\016\n\021ClusterController\022\225\002\n\rCreateC" + + "TION\020\0032\347\r\n\021ClusterController\022\221\002\n\rCreateC" + "luster\0223.google.cloud.dataproc.v1beta2.C" + "reateClusterRequest\032\035.google.longrunning" - + ".Operation\"\257\001\202\323\344\223\002G\"\022 * * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return Whether the template field is set. @@ -147,7 +147,7 @@ public boolean hasTemplate() { *
* * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return The template. @@ -166,7 +166,7 @@ public com.google.cloud.dataproc.v1beta2.WorkflowTemplate getTemplate() { *
* * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public com.google.cloud.dataproc.v1beta2.WorkflowTemplateOrBuilder getTemplateOrBuilder() { @@ -523,7 +523,7 @@ public Builder mergeFrom( *
* * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return Whether the template field is set. @@ -540,7 +540,7 @@ public boolean hasTemplate() { *
* * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return The template. @@ -563,7 +563,7 @@ public com.google.cloud.dataproc.v1beta2.WorkflowTemplate getTemplate() { *
* * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public Builder setTemplate(com.google.cloud.dataproc.v1beta2.WorkflowTemplate value) { @@ -588,7 +588,7 @@ public Builder setTemplate(com.google.cloud.dataproc.v1beta2.WorkflowTemplate va *
* * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public Builder setTemplate( @@ -611,7 +611,7 @@ public Builder setTemplate( *
* * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public Builder mergeTemplate(com.google.cloud.dataproc.v1beta2.WorkflowTemplate value) { @@ -640,7 +640,7 @@ public Builder mergeTemplate(com.google.cloud.dataproc.v1beta2.WorkflowTemplate * * * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public Builder clearTemplate() { @@ -663,7 +663,7 @@ public Builder clearTemplate() { * * * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public com.google.cloud.dataproc.v1beta2.WorkflowTemplate.Builder getTemplateBuilder() { @@ -680,7 +680,7 @@ public com.google.cloud.dataproc.v1beta2.WorkflowTemplate.Builder getTemplateBui * * * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ public com.google.cloud.dataproc.v1beta2.WorkflowTemplateOrBuilder getTemplateOrBuilder() { @@ -701,7 +701,7 @@ public com.google.cloud.dataproc.v1beta2.WorkflowTemplateOrBuilder getTemplateOr * * * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ private com.google.protobuf.SingleFieldBuilderV3< diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateWorkflowTemplateRequestOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateWorkflowTemplateRequestOrBuilder.java index 64f701b1..45a16446 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateWorkflowTemplateRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateWorkflowTemplateRequestOrBuilder.java @@ -32,7 +32,7 @@ public interface UpdateWorkflowTemplateRequestOrBuilder * * * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return Whether the template field is set. @@ -47,7 +47,7 @@ public interface UpdateWorkflowTemplateRequestOrBuilder * * * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * * * @return The template. @@ -62,7 +62,7 @@ public interface UpdateWorkflowTemplateRequestOrBuilder * * * - * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... } + * .google.cloud.dataproc.v1beta2.WorkflowTemplate template = 1 [(.google.api.field_behavior) = REQUIRED]; * */ com.google.cloud.dataproc.v1beta2.WorkflowTemplateOrBuilder getTemplateOrBuilder(); diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java index 0b751251..b6fe4025 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java @@ -262,80 +262,79 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "dataproc.googleapis.com/WorkflowTemplate" + "\022F\n\010template\030\002 \001(\0132/.google.cloud.datapr" + "oc.v1beta2.WorkflowTemplateB\003\340A\002\022\023\n\013inst" - + "ance_id\030\003 \001(\t\022\022\n\nrequest_id\030\004 \001(\t\"\224\001\n\035Up" - + "dateWorkflowTemplateRequest\022s\n\010template\030" - + "\001 \001(\0132/.google.cloud.dataproc.v1beta2.Wo" - + "rkflowTemplateB0\340A\002\372A*\n(dataproc.googlea" - + "pis.com/WorkflowTemplate\"Z\n\034ListWorkflow" - + "TemplatesRequest\022\023\n\006parent\030\001 \001(\tB\003\340A\002\022\021\n" - + "\tpage_size\030\002 \001(\005\022\022\n\npage_token\030\003 \001(\t\"\206\001\n" - + "\035ListWorkflowTemplatesResponse\022G\n\ttempla" - + "tes\030\001 \003(\0132/.google.cloud.dataproc.v1beta" - + "2.WorkflowTemplateB\003\340A\003\022\034\n\017next_page_tok" - + "en\030\002 \001(\tB\003\340A\003\"p\n\035DeleteWorkflowTemplateR" - + "equest\022>\n\004name\030\001 \001(\tB0\340A\002\372A*\n(dataproc.g" - + "oogleapis.com/WorkflowTemplate\022\017\n\007versio" - + "n\030\002 \001(\0052\351\021\n\027WorkflowTemplateService\022\260\002\n\026" - + "CreateWorkflowTemplate\022<.google.cloud.da" - + "taproc.v1beta2.CreateWorkflowTemplateReq" + + "ance_id\030\003 \001(\t\022\022\n\nrequest_id\030\004 \001(\t\"g\n\035Upd" + + "ateWorkflowTemplateRequest\022F\n\010template\030\001" + + " \001(\0132/.google.cloud.dataproc.v1beta2.Wor" + + "kflowTemplateB\003\340A\002\"Z\n\034ListWorkflowTempla" + + "tesRequest\022\023\n\006parent\030\001 \001(\tB\003\340A\002\022\021\n\tpage_" + + "size\030\002 \001(\005\022\022\n\npage_token\030\003 \001(\t\"\206\001\n\035ListW" + + "orkflowTemplatesResponse\022G\n\ttemplates\030\001 " + + "\003(\0132/.google.cloud.dataproc.v1beta2.Work" + + "flowTemplateB\003\340A\003\022\034\n\017next_page_token\030\002 \001" + + "(\tB\003\340A\003\"p\n\035DeleteWorkflowTemplateRequest" + + "\022>\n\004name\030\001 \001(\tB0\340A\002\372A*\n(dataproc.googlea" + + "pis.com/WorkflowTemplate\022\017\n\007version\030\002 \001(" + + "\0052\351\021\n\027WorkflowTemplateService\022\260\002\n\026Create" + + "WorkflowTemplate\022<.google.cloud.dataproc" + + ".v1beta2.CreateWorkflowTemplateRequest\032/" + + ".google.cloud.dataproc.v1beta2.WorkflowT" + + "emplate\"\246\001\202\323\344\223\002\214\001\"8/v1beta2/{parent=proj" + + "ects/*/regions/*}/workflowTemplates:\010tem" + + "plateZF\":/v1beta2/{parent=projects/*/loc" + + "ations/*}/workflowTemplates:\010template\332A\020" + + "parent, template\022\211\002\n\023GetWorkflowTemplate" + + "\0229.google.cloud.dataproc.v1beta2.GetWork" + + "flowTemplateRequest\032/.google.cloud.datap" + + "roc.v1beta2.WorkflowTemplate\"\205\001\202\323\344\223\002x\0228/" + + "v1beta2/{name=projects/*/regions/*/workf" + + "lowTemplates/*}Z<\022:/v1beta2/{name=projec" + + "ts/*/locations/*/workflowTemplates/*}\332A\004" + + "name\022\345\002\n\033InstantiateWorkflowTemplate\022A.g" + + "oogle.cloud.dataproc.v1beta2.Instantiate" + + "WorkflowTemplateRequest\032\035.google.longrun" + + "ning.Operation\"\343\001\202\323\344\223\002\226\001\"D/v1beta2/{name" + + "=projects/*/regions/*/workflowTemplates/" + + "*}:instantiate:\001*ZK\"F/v1beta2/{name=proj" + + "ects/*/locations/*/workflowTemplates/*}:" + + "instantiate:\001*\332A\004name\332A\020name, parameters" + + "\312A)\n\025google.protobuf.Empty\022\020WorkflowMeta" + + "data\022\204\003\n!InstantiateInlineWorkflowTempla" + + "te\022G.google.cloud.dataproc.v1beta2.Insta" + + "ntiateInlineWorkflowTemplateRequest\032\035.go" + + "ogle.longrunning.Operation\"\366\001\202\323\344\223\002\260\001\"L/v" + + "1beta2/{parent=projects/*/locations/*}/w" + + "orkflowTemplates:instantiateInline:\010temp" + + "lateZV\"J/v1beta2/{parent=projects/*/regi" + + "ons/*}/workflowTemplates:instantiateInli" + + "ne:\010template\332A\020parent, template\312A)\n\025goog" + + "le.protobuf.Empty\022\020WorkflowMetadata\022\272\002\n\026" + + "UpdateWorkflowTemplate\022<.google.cloud.da" + + "taproc.v1beta2.UpdateWorkflowTemplateReq" + "uest\032/.google.cloud.dataproc.v1beta2.Wor" - + "kflowTemplate\"\246\001\202\323\344\223\002\214\001\"8/v1beta2/{paren" - + "t=projects/*/regions/*}/workflowTemplate" - + "s:\010templateZF\":/v1beta2/{parent=projects" - + "/*/locations/*}/workflowTemplates:\010templ" - + "ate\332A\020parent, template\022\211\002\n\023GetWorkflowTe" - + "mplate\0229.google.cloud.dataproc.v1beta2.G" - + "etWorkflowTemplateRequest\032/.google.cloud" - + ".dataproc.v1beta2.WorkflowTemplate\"\205\001\202\323\344" - + "\223\002x\0228/v1beta2/{name=projects/*/regions/*" - + "/workflowTemplates/*}Z<\022:/v1beta2/{name=" - + "projects/*/locations/*/workflowTemplates" - + "/*}\332A\004name\022\345\002\n\033InstantiateWorkflowTempla" - + "te\022A.google.cloud.dataproc.v1beta2.Insta" - + "ntiateWorkflowTemplateRequest\032\035.google.l" - + "ongrunning.Operation\"\343\001\202\323\344\223\002\226\001\"D/v1beta2" - + "/{name=projects/*/regions/*/workflowTemp" - + "lates/*}:instantiate:\001*ZK\"F/v1beta2/{nam" - + "e=projects/*/locations/*/workflowTemplat" - + "es/*}:instantiate:\001*\332A\004name\332A\020name, para" - + "meters\312A)\n\025google.protobuf.Empty\022\020Workfl" - + "owMetadata\022\204\003\n!InstantiateInlineWorkflow" - + "Template\022G.google.cloud.dataproc.v1beta2" - + ".InstantiateInlineWorkflowTemplateReques" - + "t\032\035.google.longrunning.Operation\"\366\001\202\323\344\223\002" - + "\260\001\"L/v1beta2/{parent=projects/*/location" - + "s/*}/workflowTemplates:instantiateInline" - + ":\010templateZV\"J/v1beta2/{parent=projects/" - + "*/regions/*}/workflowTemplates:instantia" - + "teInline:\010template\332A\020parent, template\312A)" - + "\n\025google.protobuf.Empty\022\020WorkflowMetadat" - + "a\022\272\002\n\026UpdateWorkflowTemplate\022<.google.cl" - + "oud.dataproc.v1beta2.UpdateWorkflowTempl" - + "ateRequest\032/.google.cloud.dataproc.v1bet" - + "a2.WorkflowTemplate\"\260\001\202\323\344\223\002\236\001\032A/v1beta2/" - + "{template.name=projects/*/regions/*/work" - + "flowTemplates/*}:\010templateZO\032C/v1beta2/{" - + "template.name=projects/*/locations/*/wor" - + "kflowTemplates/*}:\010template\332A\010template\022\234" - + "\002\n\025ListWorkflowTemplates\022;.google.cloud." - + "dataproc.v1beta2.ListWorkflowTemplatesRe" - + "quest\032<.google.cloud.dataproc.v1beta2.Li" - + "stWorkflowTemplatesResponse\"\207\001\202\323\344\223\002x\0228/v" - + "1beta2/{parent=projects/*/regions/*}/wor" - + "kflowTemplatesZ<\022:/v1beta2/{parent=proje" - + "cts/*/locations/*}/workflowTemplates\332A\006p" - + "arent\022\366\001\n\026DeleteWorkflowTemplate\022<.googl" - + "e.cloud.dataproc.v1beta2.DeleteWorkflowT" - + "emplateRequest\032\026.google.protobuf.Empty\"\205" - + "\001\202\323\344\223\002x*8/v1beta2/{name=projects/*/regio" - + "ns/*/workflowTemplates/*}Z<*:/v1beta2/{n" - + "ame=projects/*/locations/*/workflowTempl" - + "ates/*}\332A\004name\032K\312A\027dataproc.googleapis.c" - + "om\322A.https://www.googleapis.com/auth/clo" - + "ud-platformB\204\001\n!com.google.cloud.datapro" - + "c.v1beta2B\026WorkflowTemplatesProtoP\001ZEgoo" - + "gle.golang.org/genproto/googleapis/cloud" - + "/dataproc/v1beta2;dataprocb\006proto3" + + "kflowTemplate\"\260\001\202\323\344\223\002\236\001\032A/v1beta2/{templ" + + "ate.name=projects/*/regions/*/workflowTe" + + "mplates/*}:\010templateZO\032C/v1beta2/{templa" + + "te.name=projects/*/locations/*/workflowT" + + "emplates/*}:\010template\332A\010template\022\234\002\n\025Lis" + + "tWorkflowTemplates\022;.google.cloud.datapr" + + "oc.v1beta2.ListWorkflowTemplatesRequest\032" + + "<.google.cloud.dataproc.v1beta2.ListWork" + + "flowTemplatesResponse\"\207\001\202\323\344\223\002x\0228/v1beta2" + + "/{parent=projects/*/regions/*}/workflowT" + + "emplatesZ<\022:/v1beta2/{parent=projects/*/" + + "locations/*}/workflowTemplates\332A\006parent\022" + + "\366\001\n\026DeleteWorkflowTemplate\022<.google.clou" + + "d.dataproc.v1beta2.DeleteWorkflowTemplat" + + "eRequest\032\026.google.protobuf.Empty\"\205\001\202\323\344\223\002" + + "x*8/v1beta2/{name=projects/*/regions/*/w" + + "orkflowTemplates/*}Z<*:/v1beta2/{name=pr" + + "ojects/*/locations/*/workflowTemplates/*" + + "}\332A\004name\032K\312A\027dataproc.googleapis.com\322A.h" + + "ttps://www.googleapis.com/auth/cloud-pla" + + "tformB\204\001\n!com.google.cloud.dataproc.v1be" + + "ta2B\026WorkflowTemplatesProtoP\001ZEgoogle.go" + + "lang.org/genproto/googleapis/cloud/datap" + + "roc/v1beta2;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto index 62543505..4b2ee649 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto +++ b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto @@ -43,7 +43,7 @@ service ClusterController { // [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata). rpc CreateCluster(CreateClusterRequest) returns (google.longrunning.Operation) { option (google.api.http) = { - post: "/v1beta2/projects/{project_id=*}/regions/{region=*}/clusters" + post: "/v1beta2/projects/{project_id}/regions/{region}/clusters" body: "cluster" }; option (google.api.method_signature) = "project_id, region, cluster"; @@ -58,7 +58,7 @@ service ClusterController { // [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata). rpc UpdateCluster(UpdateClusterRequest) returns (google.longrunning.Operation) { option (google.api.http) = { - patch: "/v1beta2/projects/{project_id=*}/regions/{region=*}/clusters/{cluster_name=*}" + patch: "/v1beta2/projects/{project_id}/regions/{region}/clusters/{cluster_name}" body: "cluster" }; option (google.api.method_signature) = "project_id, region, cluster_name, cluster, update_mask"; @@ -73,7 +73,7 @@ service ClusterController { // [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata). rpc DeleteCluster(DeleteClusterRequest) returns (google.longrunning.Operation) { option (google.api.http) = { - delete: "/v1beta2/projects/{project_id=*}/regions/{region=*}/clusters/{cluster_name=*}" + delete: "/v1beta2/projects/{project_id}/regions/{region}/clusters/{cluster_name}" }; option (google.api.method_signature) = "project_id, region, cluster_name"; option (google.longrunning.operation_info) = { @@ -85,7 +85,7 @@ service ClusterController { // Gets the resource representation for a cluster in a project. rpc GetCluster(GetClusterRequest) returns (Cluster) { option (google.api.http) = { - get: "/v1beta2/projects/{project_id=*}/regions/{region=*}/clusters/{cluster_name=*}" + get: "/v1beta2/projects/{project_id}/regions/{region}/clusters/{cluster_name}" }; option (google.api.method_signature) = "project_id, region, cluster_name"; } @@ -93,7 +93,7 @@ service ClusterController { // Lists all regions/{region}/clusters in a project. rpc ListClusters(ListClustersRequest) returns (ListClustersResponse) { option (google.api.http) = { - get: "/v1beta2/projects/{project_id=*}/regions/{region=*}/clusters" + get: "/v1beta2/projects/{project_id}/regions/{region}/clusters" }; option (google.api.method_signature) = "project_id, region"; option (google.api.method_signature) = "project_id, region, filter"; @@ -108,7 +108,7 @@ service ClusterController { // [Empty](google.protobuf.Empty). rpc DiagnoseCluster(DiagnoseClusterRequest) returns (google.longrunning.Operation) { option (google.api.http) = { - post: "/v1beta2/projects/{project_id=*}/regions/{region=*}/clusters/{cluster_name=*}:diagnose" + post: "/v1beta2/projects/{project_id}/regions/{region}/clusters/{cluster_name}:diagnose" body: "*" }; option (google.api.method_signature) = "project_id, region, cluster_name"; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto index 02d44719..c1e643c9 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto +++ b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto @@ -37,7 +37,7 @@ service JobController { // Submits a job to a cluster. rpc SubmitJob(SubmitJobRequest) returns (Job) { option (google.api.http) = { - post: "/v1beta2/projects/{project_id=*}/regions/{region=*}/jobs:submit" + post: "/v1beta2/projects/{project_id}/regions/{region}/jobs:submit" body: "*" }; option (google.api.method_signature) = "project_id, region, job"; @@ -46,7 +46,7 @@ service JobController { // Gets the resource representation for a job in a project. rpc GetJob(GetJobRequest) returns (Job) { option (google.api.http) = { - get: "/v1beta2/projects/{project_id=*}/regions/{region=*}/jobs/{job_id=*}" + get: "/v1beta2/projects/{project_id}/regions/{region}/jobs/{job_id}" }; option (google.api.method_signature) = "project_id, region, job_id"; } @@ -54,7 +54,7 @@ service JobController { // Lists regions/{region}/jobs in a project. rpc ListJobs(ListJobsRequest) returns (ListJobsResponse) { option (google.api.http) = { - get: "/v1beta2/projects/{project_id=*}/regions/{region=*}/jobs" + get: "/v1beta2/projects/{project_id}/regions/{region}/jobs" }; option (google.api.method_signature) = "project_id, region"; option (google.api.method_signature) = "project_id, region, filter"; @@ -63,7 +63,7 @@ service JobController { // Updates a job in a project. rpc UpdateJob(UpdateJobRequest) returns (Job) { option (google.api.http) = { - patch: "/v1beta2/projects/{project_id=*}/regions/{region=*}/jobs/{job_id=*}" + patch: "/v1beta2/projects/{project_id}/regions/{region}/jobs/{job_id}" body: "job" }; } @@ -75,7 +75,7 @@ service JobController { // [regions/{region}/jobs.get](/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/get). rpc CancelJob(CancelJobRequest) returns (Job) { option (google.api.http) = { - post: "/v1beta2/projects/{project_id=*}/regions/{region=*}/jobs/{job_id=*}:cancel" + post: "/v1beta2/projects/{project_id}/regions/{region}/jobs/{job_id}:cancel" body: "*" }; option (google.api.method_signature) = "project_id, region, job_id"; @@ -85,7 +85,7 @@ service JobController { // and the response returns `FAILED_PRECONDITION`. rpc DeleteJob(DeleteJobRequest) returns (google.protobuf.Empty) { option (google.api.http) = { - delete: "/v1beta2/projects/{project_id=*}/regions/{region=*}/jobs/{job_id=*}" + delete: "/v1beta2/projects/{project_id}/regions/{region}/jobs/{job_id}" }; option (google.api.method_signature) = "project_id, region, job_id"; } diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto index 6fd30974..2979593d 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto +++ b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto @@ -722,10 +722,7 @@ message UpdateWorkflowTemplateRequest { // // The `template.version` field must match the current version. WorkflowTemplate template = 1 [ - (google.api.field_behavior) = REQUIRED, - (google.api.resource_reference) = { - type: "dataproc.googleapis.com/WorkflowTemplate" - } + (google.api.field_behavior) = REQUIRED ]; } diff --git a/synth.metadata b/synth.metadata index 2578f6e8..65bef1bc 100644 --- a/synth.metadata +++ b/synth.metadata @@ -1,19 +1,19 @@ { - "updateTime": "2019-12-12T08:40:24.848514Z", + "updateTime": "2019-12-19T08:42:14.120575Z", "sources": [ { "generator": { "name": "artman", - "version": "0.42.2", - "dockerImage": "googleapis/artman@sha256:a86051ce32e2454ca2558ec9e437d46c2184b6dd291f57b0686676d77242658e" + "version": "0.42.3", + "dockerImage": "googleapis/artman@sha256:feed210b5723c6f524b52ef6d7740a030f2d1a8f7c29a71c5e5b4481ceaad7f5" } }, { "git": { "name": "googleapis", "remote": "https://github.com/googleapis/googleapis.git", - "sha": "19c4589a3cb44b3679f7b3fba88365b3d055d5f8", - "internalRef": "285013366" + "sha": "d400cb8d45df5b2ae796b909f098a215b2275c1d", + "internalRef": "286312673" } }, { @@ -45,5 +45,1558 @@ "config": "google/cloud/dataproc/artman_dataproc_v1beta2.yaml" } } + ], + "newFiles": [ + { + "path": ".gitignore" + }, + { + "path": "README.md" + }, + { + "path": "renovate.json" + }, + { + "path": "license-checks.xml" + }, + { + "path": "CONTRIBUTING.md" + }, + { + "path": "synth.py" + }, + { + "path": "synth.metadata" + }, + { + "path": "java.header" + }, + { + "path": "pom.xml" + }, + { + "path": "CODE_OF_CONDUCT.md" + }, + { + "path": "versions.txt" + }, + { + "path": "CHANGELOG.md" + }, + { + "path": "LICENSE" + }, + { + "path": ".repo-metadata.json" + }, + { + "path": "codecov.yaml" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/pom.xml" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/autoscaling_policies.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/shared.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/operations.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListClustersRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/LoggingConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/LoggingConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListWorkflowTemplatesResponseOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobScheduling.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterOperationStatusOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobPlacementOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DeleteWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListAutoscalingPoliciesResponseOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SecurityConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SparkJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterOperationStatus.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SubmitJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PySparkJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListAutoscalingPoliciesRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstantiateWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobReferenceOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/BasicAutoscalingAlgorithmOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplate.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GetWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SparkSqlJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListAutoscalingPoliciesResponse.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GetAutoscalingPolicyRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SparkRJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/HiveJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterOperationMetadata.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Cluster.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobsProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/CancelJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/CreateAutoscalingPolicyRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupAutoscalingPolicyConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ManagedCluster.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListWorkflowTemplatesResponse.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/EndpointConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobStatusOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/CancelJobRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListJobsResponse.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ManagedGroupConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatePlacementOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/QueryList.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DeleteJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterStatusOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DiagnoseClusterResultsOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterSelectorOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateAutoscalingPolicyRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingPolicyName.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstantiateInlineWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListJobsRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingPoliciesProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowMetadata.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ManagedGroupConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListJobsRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/BasicYarnAutoscalingConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DeleteJobRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListClustersResponseOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupAutoscalingPolicyConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DeleteAutoscalingPolicyRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameter.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SharedProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateName.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Component.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowMetadataOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DiskConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SparkJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/RegexValidation.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/NodeInitializationAction.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/EncryptionConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterSelector.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/LifecycleConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DeleteAutoscalingPolicyRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowNode.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/HadoopJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowGraphOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GetAutoscalingPolicyRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SparkRJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PySparkJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListWorkflowTemplatesRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ParameterValidation.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobStatus.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/CreateAutoscalingPolicyRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/RegexValidationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobPlacement.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListClustersRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SubmitJobRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GetClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ManagedClusterOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/CreateClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DiagnoseClusterResults.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/CreateClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/EndpointConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListClustersResponse.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OperationsProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterMetricsOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameterOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DiagnoseClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowNodeOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateAutoscalingPolicyRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PigJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GetJobRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Job.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingPolicy.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GetJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/NodeInitializationActionOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SecurityConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/EncryptionConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DiskConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/YarnApplication.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterStatus.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstantiateInlineWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListAutoscalingPoliciesRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/YarnApplicationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobSchedulingOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/CreateWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ValueValidation.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/LifecycleConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PigJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingPolicyOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DeleteClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterOperationMetadataOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterMetrics.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterOperation.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListWorkflowTemplatesRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GetWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SparkSqlJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatePlacement.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ReservationAffinityOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DeleteClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/BasicYarnAutoscalingConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/RegionName.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ValueValidationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GetClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/HadoopJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/QueryListOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/HiveJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DeleteWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/CreateWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/BasicAutoscalingAlgorithm.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/DiagnoseClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterOperationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstantiateWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ListJobsResponseOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ParameterValidationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ReservationAffinity.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowGraph.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobReference.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/UpdateJobRequestOrBuilder.java" + }, + { + "path": "google-cloud-dataproc/pom.xml" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceClientTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockJobControllerImpl.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockWorkflowTemplateServiceImpl.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockClusterControllerImpl.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/ClusterControllerSmokeTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/ClusterControllerClientTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockJobController.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/JobControllerClientTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockClusterController.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockWorkflowTemplateService.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/it/ITSystemTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateServiceClientTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockJobControllerImpl.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockWorkflowTemplateServiceImpl.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockClusterControllerImpl.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/ClusterControllerSmokeTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/ClusterControllerClientTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockJobController.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/JobControllerClientTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockClusterController.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/AutoscalingPolicyServiceClientTest.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockAutoscalingPolicyServiceImpl.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockWorkflowTemplateService.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockAutoscalingPolicyService.java" + }, + { + "path": "google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/it/ITSystemTest.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceClient.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/package-info.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcWorkflowTemplateServiceStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcClusterControllerCallableFactory.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcClusterControllerStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/ClusterControllerStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcJobControllerStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcJobControllerCallableFactory.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/ClusterControllerStubSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcWorkflowTemplateServiceCallableFactory.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/WorkflowTemplateServiceStubSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/WorkflowTemplateServiceStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStubSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingPolicyServiceClient.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateServiceClient.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingPolicyServiceSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateServiceSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerClient.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerClient.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/package-info.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcWorkflowTemplateServiceStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcClusterControllerCallableFactory.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcClusterControllerStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/ClusterControllerStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcJobControllerStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcJobControllerCallableFactory.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/AutoscalingPolicyServiceStubSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcAutoscalingPolicyServiceCallableFactory.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/ClusterControllerStubSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcAutoscalingPolicyServiceStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/AutoscalingPolicyServiceStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcWorkflowTemplateServiceCallableFactory.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/WorkflowTemplateServiceStubSettings.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/WorkflowTemplateServiceStub.java" + }, + { + "path": "google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStubSettings.java" + }, + { + "path": ".github/release-please.yml" + }, + { + "path": ".github/PULL_REQUEST_TEMPLATE.md" + }, + { + "path": ".github/ISSUE_TEMPLATE/support_request.md" + }, + { + "path": ".github/ISSUE_TEMPLATE/feature_request.md" + }, + { + "path": ".github/ISSUE_TEMPLATE/bug_report.md" + }, + { + "path": ".kokoro/trampoline.sh" + }, + { + "path": ".kokoro/build.sh" + }, + { + "path": ".kokoro/build.bat" + }, + { + "path": ".kokoro/linkage-monitor.sh" + }, + { + "path": ".kokoro/dependencies.sh" + }, + { + "path": ".kokoro/common.cfg" + }, + { + "path": ".kokoro/coerce_logs.sh" + }, + { + "path": ".kokoro/release/common.sh" + }, + { + "path": ".kokoro/release/publish_javadoc.cfg" + }, + { + "path": ".kokoro/release/stage.cfg" + }, + { + "path": ".kokoro/release/publish_javadoc.sh" + }, + { + "path": ".kokoro/release/bump_snapshot.cfg" + }, + { + "path": ".kokoro/release/drop.cfg" + }, + { + "path": ".kokoro/release/snapshot.cfg" + }, + { + "path": ".kokoro/release/promote.sh" + }, + { + "path": ".kokoro/release/stage.sh" + }, + { + "path": ".kokoro/release/bump_snapshot.sh" + }, + { + "path": ".kokoro/release/promote.cfg" + }, + { + "path": ".kokoro/release/drop.sh" + }, + { + "path": ".kokoro/release/common.cfg" + }, + { + "path": ".kokoro/release/snapshot.sh" + }, + { + "path": ".kokoro/nightly/java11.cfg" + }, + { + "path": ".kokoro/nightly/java8-win.cfg" + }, + { + "path": ".kokoro/nightly/lint.cfg" + }, + { + "path": ".kokoro/nightly/java8-osx.cfg" + }, + { + "path": ".kokoro/nightly/integration.cfg" + }, + { + "path": ".kokoro/nightly/java7.cfg" + }, + { + "path": ".kokoro/nightly/java8.cfg" + }, + { + "path": ".kokoro/nightly/dependencies.cfg" + }, + { + "path": ".kokoro/nightly/common.cfg" + }, + { + "path": ".kokoro/presubmit/java11.cfg" + }, + { + "path": ".kokoro/presubmit/java8-win.cfg" + }, + { + "path": ".kokoro/presubmit/lint.cfg" + }, + { + "path": ".kokoro/presubmit/clirr.cfg" + }, + { + "path": ".kokoro/presubmit/java8-osx.cfg" + }, + { + "path": ".kokoro/presubmit/integration.cfg" + }, + { + "path": ".kokoro/presubmit/java7.cfg" + }, + { + "path": ".kokoro/presubmit/java8.cfg" + }, + { + "path": ".kokoro/presubmit/dependencies.cfg" + }, + { + "path": ".kokoro/presubmit/linkage-monitor.cfg" + }, + { + "path": ".kokoro/presubmit/common.cfg" + }, + { + "path": ".kokoro/continuous/propose_release.cfg" + }, + { + "path": ".kokoro/continuous/java11.cfg" + }, + { + "path": ".kokoro/continuous/java8-win.cfg" + }, + { + "path": ".kokoro/continuous/lint.cfg" + }, + { + "path": ".kokoro/continuous/java8-osx.cfg" + }, + { + "path": ".kokoro/continuous/propose_release.sh" + }, + { + "path": ".kokoro/continuous/integration.cfg" + }, + { + "path": ".kokoro/continuous/java7.cfg" + }, + { + "path": ".kokoro/continuous/java8.cfg" + }, + { + "path": ".kokoro/continuous/dependencies.cfg" + }, + { + "path": ".kokoro/continuous/common.cfg" + }, + { + "path": "grpc-google-cloud-dataproc-v1beta2/pom.xml" + }, + { + "path": "grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerGrpc.java" + }, + { + "path": "grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerGrpc.java" + }, + { + "path": "grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateServiceGrpc.java" + }, + { + "path": "grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AutoscalingPolicyServiceGrpc.java" + }, + { + "path": "google-cloud-dataproc-bom/pom.xml" + }, + { + "path": "__pycache__/synth.cpython-36.pyc" + }, + { + "path": "grpc-google-cloud-dataproc-v1/pom.xml" + }, + { + "path": "grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobControllerGrpc.java" + }, + { + "path": "grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerGrpc.java" + }, + { + "path": "grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateServiceGrpc.java" + }, + { + "path": "grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceGrpc.java" + }, + { + "path": ".git/shallow" + }, + { + "path": ".git/index" + }, + { + "path": ".git/packed-refs" + }, + { + "path": ".git/HEAD" + }, + { + "path": ".git/config" + }, + { + "path": ".git/logs/HEAD" + }, + { + "path": ".git/logs/refs/heads/autosynth" + }, + { + "path": ".git/logs/refs/heads/master" + }, + { + "path": ".git/logs/refs/remotes/origin/HEAD" + }, + { + "path": ".git/objects/pack/pack-0e1ebd057cb815de98ae44e95065b4dd13d2fc96.idx" + }, + { + "path": ".git/objects/pack/pack-0e1ebd057cb815de98ae44e95065b4dd13d2fc96.pack" + }, + { + "path": ".git/refs/heads/autosynth" + }, + { + "path": ".git/refs/heads/master" + }, + { + "path": ".git/refs/remotes/origin/HEAD" + }, + { + "path": "proto-google-cloud-dataproc-v1/pom.xml" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/autoscaling_policies.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/shared.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/clusters.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/jobs.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/operations.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/LoggingConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/LoggingConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListWorkflowTemplatesResponseOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobScheduling.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOperationStatusOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacementOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListAutoscalingPoliciesResponseOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SecurityConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOperationStatus.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListAutoscalingPoliciesRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstantiateWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobReferenceOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/BasicAutoscalingAlgorithmOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplate.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListAutoscalingPoliciesResponse.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetAutoscalingPolicyRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOperationMetadata.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Cluster.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobsProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupAutoscalingPolicyConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ManagedCluster.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListWorkflowTemplatesResponse.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobStatusOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CancelJobRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsResponse.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ManagedGroupConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatePlacementOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/QueryList.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterStatusOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterResultsOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterSelectorOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateAutoscalingPolicyRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstantiateInlineWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPoliciesProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowMetadata.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ManagedGroupConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/BasicYarnAutoscalingConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteJobRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersResponseOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupAutoscalingPolicyConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteAutoscalingPolicyRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/TemplateParameter.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SharedProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplateName.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Component.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowMetadataOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiskConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/RegexValidation.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/NodeInitializationAction.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/EncryptionConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterSelector.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteAutoscalingPolicyRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowNode.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowGraphOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetAutoscalingPolicyRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PySparkJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListWorkflowTemplatesRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ParameterValidation.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobStatus.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/RegexValidationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobPlacement.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SubmitJobRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ManagedClusterOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterResults.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListClustersResponse.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OperationsProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterMetricsOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/TemplateParameterOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowNodeOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateAutoscalingPolicyRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicy.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetJobRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/NodeInitializationActionOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SecurityConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/EncryptionConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiskConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/YarnApplication.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterStatus.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstantiateInlineWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListAutoscalingPoliciesRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/YarnApplicationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobSchedulingOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ValueValidation.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PigJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOperationMetadataOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterMetrics.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOperation.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListWorkflowTemplatesRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SparkSqlJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatePlacement.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/BasicYarnAutoscalingConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/RegionName.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ValueValidationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GetClusterRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HadoopJobOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/QueryListOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/HiveJob.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DeleteWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateWorkflowTemplateRequestOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/BasicAutoscalingAlgorithm.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/DiagnoseClusterRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOperationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstantiateWorkflowTemplateRequest.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ListJobsResponseOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ParameterValidationOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowGraph.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobReference.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java" + }, + { + "path": "proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/UpdateJobRequestOrBuilder.java" + } ] } \ No newline at end of file