diff --git a/google-cloud-dataproc-bom/pom.xml b/google-cloud-dataproc-bom/pom.xml index f6fb48c4..67b1ce76 100644 --- a/google-cloud-dataproc-bom/pom.xml +++ b/google-cloud-dataproc-bom/pom.xml @@ -11,7 +11,7 @@ 0.5.0 - Google Cloud dataproc BOM + Google Cloud Dataproc BOM https://github.com/googleapis/java-dataproc BOM for Google Cloud Dataproc diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceClient.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceClient.java index bc2cb235..198afa73 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceClient.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPolicyServiceClient.java @@ -251,7 +251,7 @@ public final AutoscalingPolicy updateAutoscalingPolicy(UpdateAutoscalingPolicyRe * has the following format: `projects/{project_id}/regions/{region}` *

* For `projects.locations.autoscalingPolicies.create`, the resource name of the * location has the following format: `projects/{project_id}/locations/{location}` - * @param policy The autoscaling policy to create. + * @param policy Required. The autoscaling policy to create. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ public final AutoscalingPolicy createAutoscalingPolicy( @@ -284,7 +284,7 @@ public final AutoscalingPolicy createAutoscalingPolicy( * has the following format: `projects/{project_id}/regions/{region}` *

* For `projects.locations.autoscalingPolicies.create`, the resource name of the * location has the following format: `projects/{project_id}/locations/{location}` - * @param policy The autoscaling policy to create. + * @param policy Required. The autoscaling policy to create. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ public final AutoscalingPolicy createAutoscalingPolicy( @@ -317,7 +317,7 @@ public final AutoscalingPolicy createAutoscalingPolicy( * has the following format: `projects/{project_id}/regions/{region}` *

* For `projects.locations.autoscalingPolicies.create`, the resource name of the * location has the following format: `projects/{project_id}/locations/{location}` - * @param policy The autoscaling policy to create. + * @param policy Required. The autoscaling policy to create. * @throws com.google.api.gax.rpc.ApiException if the remote call fails */ public final AutoscalingPolicy createAutoscalingPolicy(String parent, AutoscalingPolicy policy) { @@ -335,8 +335,10 @@ public final AutoscalingPolicy createAutoscalingPolicy(String parent, Autoscalin *


    * try (AutoscalingPolicyServiceClient autoscalingPolicyServiceClient = AutoscalingPolicyServiceClient.create()) {
    *   RegionName parent = RegionName.of("[PROJECT]", "[REGION]");
+   *   AutoscalingPolicy policy = AutoscalingPolicy.newBuilder().build();
    *   CreateAutoscalingPolicyRequest request = CreateAutoscalingPolicyRequest.newBuilder()
    *     .setParent(parent.toString())
+   *     .setPolicy(policy)
    *     .build();
    *   AutoscalingPolicy response = autoscalingPolicyServiceClient.createAutoscalingPolicy(request);
    * }
@@ -358,8 +360,10 @@ public final AutoscalingPolicy createAutoscalingPolicy(CreateAutoscalingPolicyRe
    * 

    * try (AutoscalingPolicyServiceClient autoscalingPolicyServiceClient = AutoscalingPolicyServiceClient.create()) {
    *   RegionName parent = RegionName.of("[PROJECT]", "[REGION]");
+   *   AutoscalingPolicy policy = AutoscalingPolicy.newBuilder().build();
    *   CreateAutoscalingPolicyRequest request = CreateAutoscalingPolicyRequest.newBuilder()
    *     .setParent(parent.toString())
+   *     .setPolicy(policy)
    *     .build();
    *   ApiFuture<AutoscalingPolicy> future = autoscalingPolicyServiceClient.createAutoscalingPolicyCallable().futureCall(request);
    *   // Do something
diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java
index 9da0987a..bc3c3a40 100644
--- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java
+++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerClient.java
@@ -677,7 +677,7 @@ public final UnaryCallable getClusterCallable() {
 
   // AUTO-GENERATED DOCUMENTATION AND METHOD
   /**
-   * Lists all regions/{region}/clusters in a project.
+   * Lists all regions/{region}/clusters in a project alphabetically.
    *
    * 

Sample code: * @@ -704,7 +704,7 @@ public final ListClustersPagedResponse listClusters(String projectId, String reg // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * @@ -751,7 +751,7 @@ public final ListClustersPagedResponse listClusters( // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * @@ -778,7 +778,7 @@ public final ListClustersPagedResponse listClusters(ListClustersRequest request) // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * @@ -805,7 +805,7 @@ public final ListClustersPagedResponse listClusters(ListClustersRequest request) // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java index ad10a4f5..7b7c5377 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerClient.java @@ -20,14 +20,18 @@ import com.google.api.core.ApiFutures; import com.google.api.core.BetaApi; import com.google.api.gax.core.BackgroundResource; +import com.google.api.gax.longrunning.OperationFuture; import com.google.api.gax.paging.AbstractFixedSizeCollection; import com.google.api.gax.paging.AbstractPage; import com.google.api.gax.paging.AbstractPagedListResponse; +import com.google.api.gax.rpc.OperationCallable; import com.google.api.gax.rpc.PageContext; import com.google.api.gax.rpc.UnaryCallable; import com.google.cloud.dataproc.v1.stub.JobControllerStub; import com.google.cloud.dataproc.v1.stub.JobControllerStubSettings; import com.google.common.util.concurrent.MoreExecutors; +import com.google.longrunning.Operation; +import com.google.longrunning.OperationsClient; import com.google.protobuf.Empty; import java.io.IOException; import java.util.List; @@ -107,6 +111,7 @@ public class JobControllerClient implements BackgroundResource { private final JobControllerSettings settings; private final JobControllerStub stub; + private final OperationsClient operationsClient; /** Constructs an instance of JobControllerClient with default settings. */ public static final JobControllerClient create() throws IOException { @@ -139,12 +144,14 @@ public static final JobControllerClient create(JobControllerStub stub) { protected JobControllerClient(JobControllerSettings settings) throws IOException { this.settings = settings; this.stub = ((JobControllerStubSettings) settings.getStubSettings()).createStub(); + this.operationsClient = OperationsClient.create(this.stub.getOperationsStub()); } @BetaApi("A restructuring of stub classes is planned, so this may break in the future") protected JobControllerClient(JobControllerStub stub) { this.settings = null; this.stub = stub; + this.operationsClient = OperationsClient.create(this.stub.getOperationsStub()); } public final JobControllerSettings getSettings() { @@ -156,6 +163,16 @@ public JobControllerStub getStub() { return stub; } + /** + * Returns the OperationsClient that can be used to query the status of a long-running operation + * returned by another API method call. + */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public final OperationsClient getOperationsClient() { + return operationsClient; + } + // AUTO-GENERATED DOCUMENTATION AND METHOD /** * Submits a job to a cluster. @@ -709,6 +726,119 @@ public final UnaryCallable deleteJobCallable() { return stub.deleteJobCallable(); } + // AUTO-GENERATED DOCUMENTATION AND METHOD + /** + * Submits job to a cluster. + * + *

Sample code: + * + *


+   * try (JobControllerClient jobControllerClient = JobControllerClient.create()) {
+   *   String projectId = "";
+   *   String region = "";
+   *   Job job = Job.newBuilder().build();
+   *   Job response = jobControllerClient.submitJobAsOperationAsync(projectId, region, job).get();
+   * }
+   * 
+ * + * @param projectId Required. The ID of the Google Cloud Platform project that the job belongs to. + * @param region Required. The Dataproc region in which to handle the request. + * @param job Required. The job resource. + * @throws com.google.api.gax.rpc.ApiException if the remote call fails + */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public final OperationFuture submitJobAsOperationAsync( + String projectId, String region, Job job) { + SubmitJobRequest request = + SubmitJobRequest.newBuilder().setProjectId(projectId).setRegion(region).setJob(job).build(); + return submitJobAsOperationAsync(request); + } + + // AUTO-GENERATED DOCUMENTATION AND METHOD + /** + * Submits job to a cluster. + * + *

Sample code: + * + *


+   * try (JobControllerClient jobControllerClient = JobControllerClient.create()) {
+   *   String projectId = "";
+   *   String region = "";
+   *   Job job = Job.newBuilder().build();
+   *   SubmitJobRequest request = SubmitJobRequest.newBuilder()
+   *     .setProjectId(projectId)
+   *     .setRegion(region)
+   *     .setJob(job)
+   *     .build();
+   *   Job response = jobControllerClient.submitJobAsOperationAsync(request).get();
+   * }
+   * 
+ * + * @param request The request object containing all of the parameters for the API call. + * @throws com.google.api.gax.rpc.ApiException if the remote call fails + */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public final OperationFuture submitJobAsOperationAsync( + SubmitJobRequest request) { + return submitJobAsOperationOperationCallable().futureCall(request); + } + + // AUTO-GENERATED DOCUMENTATION AND METHOD + /** + * Submits job to a cluster. + * + *

Sample code: + * + *


+   * try (JobControllerClient jobControllerClient = JobControllerClient.create()) {
+   *   String projectId = "";
+   *   String region = "";
+   *   Job job = Job.newBuilder().build();
+   *   SubmitJobRequest request = SubmitJobRequest.newBuilder()
+   *     .setProjectId(projectId)
+   *     .setRegion(region)
+   *     .setJob(job)
+   *     .build();
+   *   OperationFuture<Job, JobMetadata> future = jobControllerClient.submitJobAsOperationOperationCallable().futureCall(request);
+   *   // Do something
+   *   Job response = future.get();
+   * }
+   * 
+ */ + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public final OperationCallable + submitJobAsOperationOperationCallable() { + return stub.submitJobAsOperationOperationCallable(); + } + + // AUTO-GENERATED DOCUMENTATION AND METHOD + /** + * Submits job to a cluster. + * + *

Sample code: + * + *


+   * try (JobControllerClient jobControllerClient = JobControllerClient.create()) {
+   *   String projectId = "";
+   *   String region = "";
+   *   Job job = Job.newBuilder().build();
+   *   SubmitJobRequest request = SubmitJobRequest.newBuilder()
+   *     .setProjectId(projectId)
+   *     .setRegion(region)
+   *     .setJob(job)
+   *     .build();
+   *   ApiFuture<Operation> future = jobControllerClient.submitJobAsOperationCallable().futureCall(request);
+   *   // Do something
+   *   Operation response = future.get();
+   * }
+   * 
+ */ + public final UnaryCallable submitJobAsOperationCallable() { + return stub.submitJobAsOperationCallable(); + } + @Override public final void close() { stub.close(); diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerSettings.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerSettings.java index 0bf1ecf0..c4b3fe2b 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerSettings.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/JobControllerSettings.java @@ -25,10 +25,12 @@ import com.google.api.gax.rpc.ApiClientHeaderProvider; import com.google.api.gax.rpc.ClientContext; import com.google.api.gax.rpc.ClientSettings; +import com.google.api.gax.rpc.OperationCallSettings; import com.google.api.gax.rpc.PagedCallSettings; import com.google.api.gax.rpc.TransportChannelProvider; import com.google.api.gax.rpc.UnaryCallSettings; import com.google.cloud.dataproc.v1.stub.JobControllerStubSettings; +import com.google.longrunning.Operation; import com.google.protobuf.Empty; import java.io.IOException; import java.util.List; @@ -99,6 +101,19 @@ public UnaryCallSettings deleteJobSettings() { return ((JobControllerStubSettings) getStubSettings()).deleteJobSettings(); } + /** Returns the object with the settings used for calls to submitJobAsOperation. */ + public UnaryCallSettings submitJobAsOperationSettings() { + return ((JobControllerStubSettings) getStubSettings()).submitJobAsOperationSettings(); + } + + /** Returns the object with the settings used for calls to submitJobAsOperation. */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public OperationCallSettings + submitJobAsOperationOperationSettings() { + return ((JobControllerStubSettings) getStubSettings()).submitJobAsOperationOperationSettings(); + } + public static final JobControllerSettings create(JobControllerStubSettings stub) throws IOException { return new JobControllerSettings.Builder(stub.toBuilder()).build(); @@ -227,6 +242,19 @@ public UnaryCallSettings.Builder deleteJobSettings() { return getStubSettingsBuilder().deleteJobSettings(); } + /** Returns the builder for the settings used for calls to submitJobAsOperation. */ + public UnaryCallSettings.Builder submitJobAsOperationSettings() { + return getStubSettingsBuilder().submitJobAsOperationSettings(); + } + + /** Returns the builder for the settings used for calls to submitJobAsOperation. */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public OperationCallSettings.Builder + submitJobAsOperationOperationSettings() { + return getStubSettingsBuilder().submitJobAsOperationOperationSettings(); + } + @Override public JobControllerSettings build() throws IOException { return new JobControllerSettings(this); diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcJobControllerStub.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcJobControllerStub.java index 7e8e54df..b12016cf 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcJobControllerStub.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/GrpcJobControllerStub.java @@ -23,15 +23,19 @@ import com.google.api.gax.grpc.GrpcCallSettings; import com.google.api.gax.grpc.GrpcStubCallableFactory; import com.google.api.gax.rpc.ClientContext; +import com.google.api.gax.rpc.OperationCallable; import com.google.api.gax.rpc.UnaryCallable; import com.google.cloud.dataproc.v1.CancelJobRequest; import com.google.cloud.dataproc.v1.DeleteJobRequest; import com.google.cloud.dataproc.v1.GetJobRequest; import com.google.cloud.dataproc.v1.Job; +import com.google.cloud.dataproc.v1.JobMetadata; import com.google.cloud.dataproc.v1.ListJobsRequest; import com.google.cloud.dataproc.v1.ListJobsResponse; import com.google.cloud.dataproc.v1.SubmitJobRequest; import com.google.cloud.dataproc.v1.UpdateJobRequest; +import com.google.longrunning.Operation; +import com.google.longrunning.stub.GrpcOperationsStub; import com.google.protobuf.Empty; import io.grpc.MethodDescriptor; import io.grpc.protobuf.ProtoUtils; @@ -92,8 +96,17 @@ public class GrpcJobControllerStub extends JobControllerStub { .setRequestMarshaller(ProtoUtils.marshaller(DeleteJobRequest.getDefaultInstance())) .setResponseMarshaller(ProtoUtils.marshaller(Empty.getDefaultInstance())) .build(); + private static final MethodDescriptor + submitJobAsOperationMethodDescriptor = + MethodDescriptor.newBuilder() + .setType(MethodDescriptor.MethodType.UNARY) + .setFullMethodName("google.cloud.dataproc.v1.JobController/SubmitJobAsOperation") + .setRequestMarshaller(ProtoUtils.marshaller(SubmitJobRequest.getDefaultInstance())) + .setResponseMarshaller(ProtoUtils.marshaller(Operation.getDefaultInstance())) + .build(); private final BackgroundResource backgroundResources; + private final GrpcOperationsStub operationsStub; private final UnaryCallable submitJobCallable; private final UnaryCallable getJobCallable; @@ -102,6 +115,9 @@ public class GrpcJobControllerStub extends JobControllerStub { private final UnaryCallable updateJobCallable; private final UnaryCallable cancelJobCallable; private final UnaryCallable deleteJobCallable; + private final UnaryCallable submitJobAsOperationCallable; + private final OperationCallable + submitJobAsOperationOperationCallable; private final GrpcStubCallableFactory callableFactory; @@ -141,6 +157,7 @@ protected GrpcJobControllerStub( GrpcStubCallableFactory callableFactory) throws IOException { this.callableFactory = callableFactory; + this.operationsStub = GrpcOperationsStub.create(clientContext, callableFactory); GrpcCallSettings submitJobTransportSettings = GrpcCallSettings.newBuilder() @@ -166,6 +183,10 @@ protected GrpcJobControllerStub( GrpcCallSettings.newBuilder() .setMethodDescriptor(deleteJobMethodDescriptor) .build(); + GrpcCallSettings submitJobAsOperationTransportSettings = + GrpcCallSettings.newBuilder() + .setMethodDescriptor(submitJobAsOperationMethodDescriptor) + .build(); this.submitJobCallable = callableFactory.createUnaryCallable( @@ -188,10 +209,26 @@ protected GrpcJobControllerStub( this.deleteJobCallable = callableFactory.createUnaryCallable( deleteJobTransportSettings, settings.deleteJobSettings(), clientContext); + this.submitJobAsOperationCallable = + callableFactory.createUnaryCallable( + submitJobAsOperationTransportSettings, + settings.submitJobAsOperationSettings(), + clientContext); + this.submitJobAsOperationOperationCallable = + callableFactory.createOperationCallable( + submitJobAsOperationTransportSettings, + settings.submitJobAsOperationOperationSettings(), + clientContext, + this.operationsStub); backgroundResources = new BackgroundResourceAggregation(clientContext.getBackgroundResources()); } + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public GrpcOperationsStub getOperationsStub() { + return operationsStub; + } + public UnaryCallable submitJobCallable() { return submitJobCallable; } @@ -220,6 +257,16 @@ public UnaryCallable deleteJobCallable() { return deleteJobCallable; } + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public OperationCallable + submitJobAsOperationOperationCallable() { + return submitJobAsOperationOperationCallable; + } + + public UnaryCallable submitJobAsOperationCallable() { + return submitJobAsOperationCallable; + } + @Override public final void close() { shutdown(); diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStub.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStub.java index 2892482c..169133c3 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStub.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStub.java @@ -19,15 +19,19 @@ import com.google.api.core.BetaApi; import com.google.api.gax.core.BackgroundResource; +import com.google.api.gax.rpc.OperationCallable; import com.google.api.gax.rpc.UnaryCallable; import com.google.cloud.dataproc.v1.CancelJobRequest; import com.google.cloud.dataproc.v1.DeleteJobRequest; import com.google.cloud.dataproc.v1.GetJobRequest; import com.google.cloud.dataproc.v1.Job; +import com.google.cloud.dataproc.v1.JobMetadata; import com.google.cloud.dataproc.v1.ListJobsRequest; import com.google.cloud.dataproc.v1.ListJobsResponse; import com.google.cloud.dataproc.v1.SubmitJobRequest; import com.google.cloud.dataproc.v1.UpdateJobRequest; +import com.google.longrunning.Operation; +import com.google.longrunning.stub.OperationsStub; import com.google.protobuf.Empty; import javax.annotation.Generated; @@ -41,6 +45,11 @@ @BetaApi("A restructuring of stub classes is planned, so this may break in the future") public abstract class JobControllerStub implements BackgroundResource { + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public OperationsStub getOperationsStub() { + throw new UnsupportedOperationException("Not implemented: getOperationsStub()"); + } + public UnaryCallable submitJobCallable() { throw new UnsupportedOperationException("Not implemented: submitJobCallable()"); } @@ -69,6 +78,17 @@ public UnaryCallable deleteJobCallable() { throw new UnsupportedOperationException("Not implemented: deleteJobCallable()"); } + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public OperationCallable + submitJobAsOperationOperationCallable() { + throw new UnsupportedOperationException( + "Not implemented: submitJobAsOperationOperationCallable()"); + } + + public UnaryCallable submitJobAsOperationCallable() { + throw new UnsupportedOperationException("Not implemented: submitJobAsOperationCallable()"); + } + @Override public abstract void close(); } diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStubSettings.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStubSettings.java index b17ca3bf..85342eaa 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStubSettings.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1/stub/JobControllerStubSettings.java @@ -26,10 +26,14 @@ import com.google.api.gax.grpc.GaxGrpcProperties; import com.google.api.gax.grpc.GrpcTransportChannel; import com.google.api.gax.grpc.InstantiatingGrpcChannelProvider; +import com.google.api.gax.grpc.ProtoOperationTransformers; +import com.google.api.gax.longrunning.OperationSnapshot; +import com.google.api.gax.longrunning.OperationTimedPollAlgorithm; import com.google.api.gax.retrying.RetrySettings; import com.google.api.gax.rpc.ApiCallContext; import com.google.api.gax.rpc.ApiClientHeaderProvider; import com.google.api.gax.rpc.ClientContext; +import com.google.api.gax.rpc.OperationCallSettings; import com.google.api.gax.rpc.PageContext; import com.google.api.gax.rpc.PagedCallSettings; import com.google.api.gax.rpc.PagedListDescriptor; @@ -43,6 +47,7 @@ import com.google.cloud.dataproc.v1.DeleteJobRequest; import com.google.cloud.dataproc.v1.GetJobRequest; import com.google.cloud.dataproc.v1.Job; +import com.google.cloud.dataproc.v1.JobMetadata; import com.google.cloud.dataproc.v1.ListJobsRequest; import com.google.cloud.dataproc.v1.ListJobsResponse; import com.google.cloud.dataproc.v1.SubmitJobRequest; @@ -51,6 +56,7 @@ import com.google.common.collect.ImmutableMap; import com.google.common.collect.ImmutableSet; import com.google.common.collect.Lists; +import com.google.longrunning.Operation; import com.google.protobuf.Empty; import java.io.IOException; import java.util.List; @@ -102,6 +108,9 @@ public class JobControllerStubSettings extends StubSettings updateJobSettings; private final UnaryCallSettings cancelJobSettings; private final UnaryCallSettings deleteJobSettings; + private final UnaryCallSettings submitJobAsOperationSettings; + private final OperationCallSettings + submitJobAsOperationOperationSettings; /** Returns the object with the settings used for calls to submitJob. */ public UnaryCallSettings submitJobSettings() { @@ -134,6 +143,18 @@ public UnaryCallSettings deleteJobSettings() { return deleteJobSettings; } + /** Returns the object with the settings used for calls to submitJobAsOperation. */ + public UnaryCallSettings submitJobAsOperationSettings() { + return submitJobAsOperationSettings; + } + + /** Returns the object with the settings used for calls to submitJobAsOperation. */ + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public OperationCallSettings + submitJobAsOperationOperationSettings() { + return submitJobAsOperationOperationSettings; + } + @BetaApi("A restructuring of stub classes is planned, so this may break in the future") public JobControllerStub createStub() throws IOException { if (getTransportChannelProvider() @@ -209,6 +230,9 @@ protected JobControllerStubSettings(Builder settingsBuilder) throws IOException updateJobSettings = settingsBuilder.updateJobSettings().build(); cancelJobSettings = settingsBuilder.cancelJobSettings().build(); deleteJobSettings = settingsBuilder.deleteJobSettings().build(); + submitJobAsOperationSettings = settingsBuilder.submitJobAsOperationSettings().build(); + submitJobAsOperationOperationSettings = + settingsBuilder.submitJobAsOperationOperationSettings().build(); } private static final PagedListDescriptor @@ -275,6 +299,10 @@ public static class Builder extends StubSettings.Builder updateJobSettings; private final UnaryCallSettings.Builder cancelJobSettings; private final UnaryCallSettings.Builder deleteJobSettings; + private final UnaryCallSettings.Builder + submitJobAsOperationSettings; + private final OperationCallSettings.Builder + submitJobAsOperationOperationSettings; private static final ImmutableMap> RETRYABLE_CODE_DEFINITIONS; @@ -289,6 +317,8 @@ public static class Builder extends StubSettings.BuildernewArrayList())); definitions.put( "non_idempotent", ImmutableSet.copyOf(Lists.newArrayList(StatusCode.Code.UNAVAILABLE))); @@ -333,6 +363,10 @@ protected Builder(ClientContext clientContext) { deleteJobSettings = UnaryCallSettings.newUnaryCallSettingsBuilder(); + submitJobAsOperationSettings = UnaryCallSettings.newUnaryCallSettingsBuilder(); + + submitJobAsOperationOperationSettings = OperationCallSettings.newBuilder(); + unaryMethodSettingsBuilders = ImmutableList.>of( submitJobSettings, @@ -340,7 +374,8 @@ protected Builder(ClientContext clientContext) { listJobsSettings, updateJobSettings, cancelJobSettings, - deleteJobSettings); + deleteJobSettings, + submitJobAsOperationSettings); initDefaults(this); } @@ -386,6 +421,32 @@ private static Builder initDefaults(Builder builder) { .setRetryableCodes(RETRYABLE_CODE_DEFINITIONS.get("non_idempotent")) .setRetrySettings(RETRY_PARAM_DEFINITIONS.get("default")); + builder + .submitJobAsOperationSettings() + .setRetryableCodes(RETRYABLE_CODE_DEFINITIONS.get("non_idempotent2")) + .setRetrySettings(RETRY_PARAM_DEFINITIONS.get("default")); + builder + .submitJobAsOperationOperationSettings() + .setInitialCallSettings( + UnaryCallSettings.newUnaryCallSettingsBuilder() + .setRetryableCodes(RETRYABLE_CODE_DEFINITIONS.get("non_idempotent2")) + .setRetrySettings(RETRY_PARAM_DEFINITIONS.get("default")) + .build()) + .setResponseTransformer(ProtoOperationTransformers.ResponseTransformer.create(Job.class)) + .setMetadataTransformer( + ProtoOperationTransformers.MetadataTransformer.create(JobMetadata.class)) + .setPollingAlgorithm( + OperationTimedPollAlgorithm.create( + RetrySettings.newBuilder() + .setInitialRetryDelay(Duration.ofMillis(500L)) + .setRetryDelayMultiplier(1.5) + .setMaxRetryDelay(Duration.ofMillis(5000L)) + .setInitialRpcTimeout(Duration.ZERO) // ignored + .setRpcTimeoutMultiplier(1.0) // ignored + .setMaxRpcTimeout(Duration.ZERO) // ignored + .setTotalTimeout(Duration.ofMillis(300000L)) + .build())); + return builder; } @@ -398,6 +459,9 @@ protected Builder(JobControllerStubSettings settings) { updateJobSettings = settings.updateJobSettings.toBuilder(); cancelJobSettings = settings.cancelJobSettings.toBuilder(); deleteJobSettings = settings.deleteJobSettings.toBuilder(); + submitJobAsOperationSettings = settings.submitJobAsOperationSettings.toBuilder(); + submitJobAsOperationOperationSettings = + settings.submitJobAsOperationOperationSettings.toBuilder(); unaryMethodSettingsBuilders = ImmutableList.>of( @@ -406,7 +470,8 @@ protected Builder(JobControllerStubSettings settings) { listJobsSettings, updateJobSettings, cancelJobSettings, - deleteJobSettings); + deleteJobSettings, + submitJobAsOperationSettings); } // NEXT_MAJOR_VER: remove 'throws Exception' @@ -456,6 +521,19 @@ public UnaryCallSettings.Builder deleteJobSettings() { return deleteJobSettings; } + /** Returns the builder for the settings used for calls to submitJobAsOperation. */ + public UnaryCallSettings.Builder submitJobAsOperationSettings() { + return submitJobAsOperationSettings; + } + + /** Returns the builder for the settings used for calls to submitJobAsOperation. */ + @BetaApi( + "The surface for use by generated code is not stable yet and may change in the future.") + public OperationCallSettings.Builder + submitJobAsOperationOperationSettings() { + return submitJobAsOperationOperationSettings; + } + @Override public JobControllerStubSettings build() throws IOException { return new JobControllerStubSettings(this); diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerClient.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerClient.java index 4cd016aa..574d4ad7 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerClient.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerClient.java @@ -682,7 +682,7 @@ public final UnaryCallable getClusterCallable() { // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * @@ -709,7 +709,7 @@ public final ListClustersPagedResponse listClusters(String projectId, String reg // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * @@ -756,7 +756,7 @@ public final ListClustersPagedResponse listClusters( // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * @@ -783,7 +783,7 @@ public final ListClustersPagedResponse listClusters(ListClustersRequest request) // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * @@ -810,7 +810,7 @@ public final ListClustersPagedResponse listClusters(ListClustersRequest request) // AUTO-GENERATED DOCUMENTATION AND METHOD /** - * Lists all regions/{region}/clusters in a project. + * Lists all regions/{region}/clusters in a project alphabetically. * *

Sample code: * diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerClient.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerClient.java index 50cbb9b7..dd63ae31 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerClient.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerClient.java @@ -20,14 +20,18 @@ import com.google.api.core.ApiFutures; import com.google.api.core.BetaApi; import com.google.api.gax.core.BackgroundResource; +import com.google.api.gax.longrunning.OperationFuture; import com.google.api.gax.paging.AbstractFixedSizeCollection; import com.google.api.gax.paging.AbstractPage; import com.google.api.gax.paging.AbstractPagedListResponse; +import com.google.api.gax.rpc.OperationCallable; import com.google.api.gax.rpc.PageContext; import com.google.api.gax.rpc.UnaryCallable; import com.google.cloud.dataproc.v1beta2.stub.JobControllerStub; import com.google.cloud.dataproc.v1beta2.stub.JobControllerStubSettings; import com.google.common.util.concurrent.MoreExecutors; +import com.google.longrunning.Operation; +import com.google.longrunning.OperationsClient; import com.google.protobuf.Empty; import java.io.IOException; import java.util.List; @@ -107,6 +111,7 @@ public class JobControllerClient implements BackgroundResource { private final JobControllerSettings settings; private final JobControllerStub stub; + private final OperationsClient operationsClient; /** Constructs an instance of JobControllerClient with default settings. */ public static final JobControllerClient create() throws IOException { @@ -139,12 +144,14 @@ public static final JobControllerClient create(JobControllerStub stub) { protected JobControllerClient(JobControllerSettings settings) throws IOException { this.settings = settings; this.stub = ((JobControllerStubSettings) settings.getStubSettings()).createStub(); + this.operationsClient = OperationsClient.create(this.stub.getOperationsStub()); } @BetaApi("A restructuring of stub classes is planned, so this may break in the future") protected JobControllerClient(JobControllerStub stub) { this.settings = null; this.stub = stub; + this.operationsClient = OperationsClient.create(this.stub.getOperationsStub()); } public final JobControllerSettings getSettings() { @@ -156,6 +163,16 @@ public JobControllerStub getStub() { return stub; } + /** + * Returns the OperationsClient that can be used to query the status of a long-running operation + * returned by another API method call. + */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public final OperationsClient getOperationsClient() { + return operationsClient; + } + // AUTO-GENERATED DOCUMENTATION AND METHOD /** * Submits a job to a cluster. @@ -709,6 +726,119 @@ public final UnaryCallable deleteJobCallable() { return stub.deleteJobCallable(); } + // AUTO-GENERATED DOCUMENTATION AND METHOD + /** + * Submits job to a cluster. + * + *

Sample code: + * + *


+   * try (JobControllerClient jobControllerClient = JobControllerClient.create()) {
+   *   String projectId = "";
+   *   String region = "";
+   *   Job job = Job.newBuilder().build();
+   *   Job response = jobControllerClient.submitJobAsOperationAsync(projectId, region, job).get();
+   * }
+   * 
+ * + * @param projectId Required. The ID of the Google Cloud Platform project that the job belongs to. + * @param region Required. The Dataproc region in which to handle the request. + * @param job Required. The job resource. + * @throws com.google.api.gax.rpc.ApiException if the remote call fails + */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public final OperationFuture submitJobAsOperationAsync( + String projectId, String region, Job job) { + SubmitJobRequest request = + SubmitJobRequest.newBuilder().setProjectId(projectId).setRegion(region).setJob(job).build(); + return submitJobAsOperationAsync(request); + } + + // AUTO-GENERATED DOCUMENTATION AND METHOD + /** + * Submits job to a cluster. + * + *

Sample code: + * + *


+   * try (JobControllerClient jobControllerClient = JobControllerClient.create()) {
+   *   String projectId = "";
+   *   String region = "";
+   *   Job job = Job.newBuilder().build();
+   *   SubmitJobRequest request = SubmitJobRequest.newBuilder()
+   *     .setProjectId(projectId)
+   *     .setRegion(region)
+   *     .setJob(job)
+   *     .build();
+   *   Job response = jobControllerClient.submitJobAsOperationAsync(request).get();
+   * }
+   * 
+ * + * @param request The request object containing all of the parameters for the API call. + * @throws com.google.api.gax.rpc.ApiException if the remote call fails + */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public final OperationFuture submitJobAsOperationAsync( + SubmitJobRequest request) { + return submitJobAsOperationOperationCallable().futureCall(request); + } + + // AUTO-GENERATED DOCUMENTATION AND METHOD + /** + * Submits job to a cluster. + * + *

Sample code: + * + *


+   * try (JobControllerClient jobControllerClient = JobControllerClient.create()) {
+   *   String projectId = "";
+   *   String region = "";
+   *   Job job = Job.newBuilder().build();
+   *   SubmitJobRequest request = SubmitJobRequest.newBuilder()
+   *     .setProjectId(projectId)
+   *     .setRegion(region)
+   *     .setJob(job)
+   *     .build();
+   *   OperationFuture<Job, JobMetadata> future = jobControllerClient.submitJobAsOperationOperationCallable().futureCall(request);
+   *   // Do something
+   *   Job response = future.get();
+   * }
+   * 
+ */ + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public final OperationCallable + submitJobAsOperationOperationCallable() { + return stub.submitJobAsOperationOperationCallable(); + } + + // AUTO-GENERATED DOCUMENTATION AND METHOD + /** + * Submits job to a cluster. + * + *

Sample code: + * + *


+   * try (JobControllerClient jobControllerClient = JobControllerClient.create()) {
+   *   String projectId = "";
+   *   String region = "";
+   *   Job job = Job.newBuilder().build();
+   *   SubmitJobRequest request = SubmitJobRequest.newBuilder()
+   *     .setProjectId(projectId)
+   *     .setRegion(region)
+   *     .setJob(job)
+   *     .build();
+   *   ApiFuture<Operation> future = jobControllerClient.submitJobAsOperationCallable().futureCall(request);
+   *   // Do something
+   *   Operation response = future.get();
+   * }
+   * 
+ */ + public final UnaryCallable submitJobAsOperationCallable() { + return stub.submitJobAsOperationCallable(); + } + @Override public final void close() { stub.close(); diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerSettings.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerSettings.java index 8d7efd56..d2532905 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerSettings.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerSettings.java @@ -25,10 +25,12 @@ import com.google.api.gax.rpc.ApiClientHeaderProvider; import com.google.api.gax.rpc.ClientContext; import com.google.api.gax.rpc.ClientSettings; +import com.google.api.gax.rpc.OperationCallSettings; import com.google.api.gax.rpc.PagedCallSettings; import com.google.api.gax.rpc.TransportChannelProvider; import com.google.api.gax.rpc.UnaryCallSettings; import com.google.cloud.dataproc.v1beta2.stub.JobControllerStubSettings; +import com.google.longrunning.Operation; import com.google.protobuf.Empty; import java.io.IOException; import java.util.List; @@ -99,6 +101,19 @@ public UnaryCallSettings deleteJobSettings() { return ((JobControllerStubSettings) getStubSettings()).deleteJobSettings(); } + /** Returns the object with the settings used for calls to submitJobAsOperation. */ + public UnaryCallSettings submitJobAsOperationSettings() { + return ((JobControllerStubSettings) getStubSettings()).submitJobAsOperationSettings(); + } + + /** Returns the object with the settings used for calls to submitJobAsOperation. */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public OperationCallSettings + submitJobAsOperationOperationSettings() { + return ((JobControllerStubSettings) getStubSettings()).submitJobAsOperationOperationSettings(); + } + public static final JobControllerSettings create(JobControllerStubSettings stub) throws IOException { return new JobControllerSettings.Builder(stub.toBuilder()).build(); @@ -227,6 +242,19 @@ public UnaryCallSettings.Builder deleteJobSettings() { return getStubSettingsBuilder().deleteJobSettings(); } + /** Returns the builder for the settings used for calls to submitJobAsOperation. */ + public UnaryCallSettings.Builder submitJobAsOperationSettings() { + return getStubSettingsBuilder().submitJobAsOperationSettings(); + } + + /** Returns the builder for the settings used for calls to submitJobAsOperation. */ + @BetaApi( + "The surface for long-running operations is not stable yet and may change in the future.") + public OperationCallSettings.Builder + submitJobAsOperationOperationSettings() { + return getStubSettingsBuilder().submitJobAsOperationOperationSettings(); + } + @Override public JobControllerSettings build() throws IOException { return new JobControllerSettings(this); diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcJobControllerStub.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcJobControllerStub.java index df041f62..3a92366c 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcJobControllerStub.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/GrpcJobControllerStub.java @@ -23,15 +23,19 @@ import com.google.api.gax.grpc.GrpcCallSettings; import com.google.api.gax.grpc.GrpcStubCallableFactory; import com.google.api.gax.rpc.ClientContext; +import com.google.api.gax.rpc.OperationCallable; import com.google.api.gax.rpc.UnaryCallable; import com.google.cloud.dataproc.v1beta2.CancelJobRequest; import com.google.cloud.dataproc.v1beta2.DeleteJobRequest; import com.google.cloud.dataproc.v1beta2.GetJobRequest; import com.google.cloud.dataproc.v1beta2.Job; +import com.google.cloud.dataproc.v1beta2.JobMetadata; import com.google.cloud.dataproc.v1beta2.ListJobsRequest; import com.google.cloud.dataproc.v1beta2.ListJobsResponse; import com.google.cloud.dataproc.v1beta2.SubmitJobRequest; import com.google.cloud.dataproc.v1beta2.UpdateJobRequest; +import com.google.longrunning.Operation; +import com.google.longrunning.stub.GrpcOperationsStub; import com.google.protobuf.Empty; import io.grpc.MethodDescriptor; import io.grpc.protobuf.ProtoUtils; @@ -92,8 +96,17 @@ public class GrpcJobControllerStub extends JobControllerStub { .setRequestMarshaller(ProtoUtils.marshaller(DeleteJobRequest.getDefaultInstance())) .setResponseMarshaller(ProtoUtils.marshaller(Empty.getDefaultInstance())) .build(); + private static final MethodDescriptor + submitJobAsOperationMethodDescriptor = + MethodDescriptor.newBuilder() + .setType(MethodDescriptor.MethodType.UNARY) + .setFullMethodName("google.cloud.dataproc.v1beta2.JobController/SubmitJobAsOperation") + .setRequestMarshaller(ProtoUtils.marshaller(SubmitJobRequest.getDefaultInstance())) + .setResponseMarshaller(ProtoUtils.marshaller(Operation.getDefaultInstance())) + .build(); private final BackgroundResource backgroundResources; + private final GrpcOperationsStub operationsStub; private final UnaryCallable submitJobCallable; private final UnaryCallable getJobCallable; @@ -102,6 +115,9 @@ public class GrpcJobControllerStub extends JobControllerStub { private final UnaryCallable updateJobCallable; private final UnaryCallable cancelJobCallable; private final UnaryCallable deleteJobCallable; + private final UnaryCallable submitJobAsOperationCallable; + private final OperationCallable + submitJobAsOperationOperationCallable; private final GrpcStubCallableFactory callableFactory; @@ -141,6 +157,7 @@ protected GrpcJobControllerStub( GrpcStubCallableFactory callableFactory) throws IOException { this.callableFactory = callableFactory; + this.operationsStub = GrpcOperationsStub.create(clientContext, callableFactory); GrpcCallSettings submitJobTransportSettings = GrpcCallSettings.newBuilder() @@ -166,6 +183,10 @@ protected GrpcJobControllerStub( GrpcCallSettings.newBuilder() .setMethodDescriptor(deleteJobMethodDescriptor) .build(); + GrpcCallSettings submitJobAsOperationTransportSettings = + GrpcCallSettings.newBuilder() + .setMethodDescriptor(submitJobAsOperationMethodDescriptor) + .build(); this.submitJobCallable = callableFactory.createUnaryCallable( @@ -188,10 +209,26 @@ protected GrpcJobControllerStub( this.deleteJobCallable = callableFactory.createUnaryCallable( deleteJobTransportSettings, settings.deleteJobSettings(), clientContext); + this.submitJobAsOperationCallable = + callableFactory.createUnaryCallable( + submitJobAsOperationTransportSettings, + settings.submitJobAsOperationSettings(), + clientContext); + this.submitJobAsOperationOperationCallable = + callableFactory.createOperationCallable( + submitJobAsOperationTransportSettings, + settings.submitJobAsOperationOperationSettings(), + clientContext, + this.operationsStub); backgroundResources = new BackgroundResourceAggregation(clientContext.getBackgroundResources()); } + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public GrpcOperationsStub getOperationsStub() { + return operationsStub; + } + public UnaryCallable submitJobCallable() { return submitJobCallable; } @@ -220,6 +257,16 @@ public UnaryCallable deleteJobCallable() { return deleteJobCallable; } + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public OperationCallable + submitJobAsOperationOperationCallable() { + return submitJobAsOperationOperationCallable; + } + + public UnaryCallable submitJobAsOperationCallable() { + return submitJobAsOperationCallable; + } + @Override public final void close() { shutdown(); diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStub.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStub.java index 9df843bd..9054e313 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStub.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStub.java @@ -19,15 +19,19 @@ import com.google.api.core.BetaApi; import com.google.api.gax.core.BackgroundResource; +import com.google.api.gax.rpc.OperationCallable; import com.google.api.gax.rpc.UnaryCallable; import com.google.cloud.dataproc.v1beta2.CancelJobRequest; import com.google.cloud.dataproc.v1beta2.DeleteJobRequest; import com.google.cloud.dataproc.v1beta2.GetJobRequest; import com.google.cloud.dataproc.v1beta2.Job; +import com.google.cloud.dataproc.v1beta2.JobMetadata; import com.google.cloud.dataproc.v1beta2.ListJobsRequest; import com.google.cloud.dataproc.v1beta2.ListJobsResponse; import com.google.cloud.dataproc.v1beta2.SubmitJobRequest; import com.google.cloud.dataproc.v1beta2.UpdateJobRequest; +import com.google.longrunning.Operation; +import com.google.longrunning.stub.OperationsStub; import com.google.protobuf.Empty; import javax.annotation.Generated; @@ -41,6 +45,11 @@ @BetaApi("A restructuring of stub classes is planned, so this may break in the future") public abstract class JobControllerStub implements BackgroundResource { + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public OperationsStub getOperationsStub() { + throw new UnsupportedOperationException("Not implemented: getOperationsStub()"); + } + public UnaryCallable submitJobCallable() { throw new UnsupportedOperationException("Not implemented: submitJobCallable()"); } @@ -69,6 +78,17 @@ public UnaryCallable deleteJobCallable() { throw new UnsupportedOperationException("Not implemented: deleteJobCallable()"); } + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public OperationCallable + submitJobAsOperationOperationCallable() { + throw new UnsupportedOperationException( + "Not implemented: submitJobAsOperationOperationCallable()"); + } + + public UnaryCallable submitJobAsOperationCallable() { + throw new UnsupportedOperationException("Not implemented: submitJobAsOperationCallable()"); + } + @Override public abstract void close(); } diff --git a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStubSettings.java b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStubSettings.java index 96bd987d..fe138db5 100644 --- a/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStubSettings.java +++ b/google-cloud-dataproc/src/main/java/com/google/cloud/dataproc/v1beta2/stub/JobControllerStubSettings.java @@ -26,10 +26,14 @@ import com.google.api.gax.grpc.GaxGrpcProperties; import com.google.api.gax.grpc.GrpcTransportChannel; import com.google.api.gax.grpc.InstantiatingGrpcChannelProvider; +import com.google.api.gax.grpc.ProtoOperationTransformers; +import com.google.api.gax.longrunning.OperationSnapshot; +import com.google.api.gax.longrunning.OperationTimedPollAlgorithm; import com.google.api.gax.retrying.RetrySettings; import com.google.api.gax.rpc.ApiCallContext; import com.google.api.gax.rpc.ApiClientHeaderProvider; import com.google.api.gax.rpc.ClientContext; +import com.google.api.gax.rpc.OperationCallSettings; import com.google.api.gax.rpc.PageContext; import com.google.api.gax.rpc.PagedCallSettings; import com.google.api.gax.rpc.PagedListDescriptor; @@ -43,6 +47,7 @@ import com.google.cloud.dataproc.v1beta2.DeleteJobRequest; import com.google.cloud.dataproc.v1beta2.GetJobRequest; import com.google.cloud.dataproc.v1beta2.Job; +import com.google.cloud.dataproc.v1beta2.JobMetadata; import com.google.cloud.dataproc.v1beta2.ListJobsRequest; import com.google.cloud.dataproc.v1beta2.ListJobsResponse; import com.google.cloud.dataproc.v1beta2.SubmitJobRequest; @@ -51,6 +56,7 @@ import com.google.common.collect.ImmutableMap; import com.google.common.collect.ImmutableSet; import com.google.common.collect.Lists; +import com.google.longrunning.Operation; import com.google.protobuf.Empty; import java.io.IOException; import java.util.List; @@ -102,6 +108,9 @@ public class JobControllerStubSettings extends StubSettings updateJobSettings; private final UnaryCallSettings cancelJobSettings; private final UnaryCallSettings deleteJobSettings; + private final UnaryCallSettings submitJobAsOperationSettings; + private final OperationCallSettings + submitJobAsOperationOperationSettings; /** Returns the object with the settings used for calls to submitJob. */ public UnaryCallSettings submitJobSettings() { @@ -134,6 +143,18 @@ public UnaryCallSettings deleteJobSettings() { return deleteJobSettings; } + /** Returns the object with the settings used for calls to submitJobAsOperation. */ + public UnaryCallSettings submitJobAsOperationSettings() { + return submitJobAsOperationSettings; + } + + /** Returns the object with the settings used for calls to submitJobAsOperation. */ + @BetaApi("The surface for use by generated code is not stable yet and may change in the future.") + public OperationCallSettings + submitJobAsOperationOperationSettings() { + return submitJobAsOperationOperationSettings; + } + @BetaApi("A restructuring of stub classes is planned, so this may break in the future") public JobControllerStub createStub() throws IOException { if (getTransportChannelProvider() @@ -209,6 +230,9 @@ protected JobControllerStubSettings(Builder settingsBuilder) throws IOException updateJobSettings = settingsBuilder.updateJobSettings().build(); cancelJobSettings = settingsBuilder.cancelJobSettings().build(); deleteJobSettings = settingsBuilder.deleteJobSettings().build(); + submitJobAsOperationSettings = settingsBuilder.submitJobAsOperationSettings().build(); + submitJobAsOperationOperationSettings = + settingsBuilder.submitJobAsOperationOperationSettings().build(); } private static final PagedListDescriptor @@ -275,6 +299,10 @@ public static class Builder extends StubSettings.Builder updateJobSettings; private final UnaryCallSettings.Builder cancelJobSettings; private final UnaryCallSettings.Builder deleteJobSettings; + private final UnaryCallSettings.Builder + submitJobAsOperationSettings; + private final OperationCallSettings.Builder + submitJobAsOperationOperationSettings; private static final ImmutableMap> RETRYABLE_CODE_DEFINITIONS; @@ -289,6 +317,8 @@ public static class Builder extends StubSettings.BuildernewArrayList())); definitions.put( "non_idempotent", ImmutableSet.copyOf(Lists.newArrayList(StatusCode.Code.UNAVAILABLE))); @@ -333,6 +363,10 @@ protected Builder(ClientContext clientContext) { deleteJobSettings = UnaryCallSettings.newUnaryCallSettingsBuilder(); + submitJobAsOperationSettings = UnaryCallSettings.newUnaryCallSettingsBuilder(); + + submitJobAsOperationOperationSettings = OperationCallSettings.newBuilder(); + unaryMethodSettingsBuilders = ImmutableList.>of( submitJobSettings, @@ -340,7 +374,8 @@ protected Builder(ClientContext clientContext) { listJobsSettings, updateJobSettings, cancelJobSettings, - deleteJobSettings); + deleteJobSettings, + submitJobAsOperationSettings); initDefaults(this); } @@ -386,6 +421,32 @@ private static Builder initDefaults(Builder builder) { .setRetryableCodes(RETRYABLE_CODE_DEFINITIONS.get("non_idempotent")) .setRetrySettings(RETRY_PARAM_DEFINITIONS.get("default")); + builder + .submitJobAsOperationSettings() + .setRetryableCodes(RETRYABLE_CODE_DEFINITIONS.get("non_idempotent2")) + .setRetrySettings(RETRY_PARAM_DEFINITIONS.get("default")); + builder + .submitJobAsOperationOperationSettings() + .setInitialCallSettings( + UnaryCallSettings.newUnaryCallSettingsBuilder() + .setRetryableCodes(RETRYABLE_CODE_DEFINITIONS.get("non_idempotent2")) + .setRetrySettings(RETRY_PARAM_DEFINITIONS.get("default")) + .build()) + .setResponseTransformer(ProtoOperationTransformers.ResponseTransformer.create(Job.class)) + .setMetadataTransformer( + ProtoOperationTransformers.MetadataTransformer.create(JobMetadata.class)) + .setPollingAlgorithm( + OperationTimedPollAlgorithm.create( + RetrySettings.newBuilder() + .setInitialRetryDelay(Duration.ofMillis(500L)) + .setRetryDelayMultiplier(1.5) + .setMaxRetryDelay(Duration.ofMillis(5000L)) + .setInitialRpcTimeout(Duration.ZERO) // ignored + .setRpcTimeoutMultiplier(1.0) // ignored + .setMaxRpcTimeout(Duration.ZERO) // ignored + .setTotalTimeout(Duration.ofMillis(300000L)) + .build())); + return builder; } @@ -398,6 +459,9 @@ protected Builder(JobControllerStubSettings settings) { updateJobSettings = settings.updateJobSettings.toBuilder(); cancelJobSettings = settings.cancelJobSettings.toBuilder(); deleteJobSettings = settings.deleteJobSettings.toBuilder(); + submitJobAsOperationSettings = settings.submitJobAsOperationSettings.toBuilder(); + submitJobAsOperationOperationSettings = + settings.submitJobAsOperationOperationSettings.toBuilder(); unaryMethodSettingsBuilders = ImmutableList.>of( @@ -406,7 +470,8 @@ protected Builder(JobControllerStubSettings settings) { listJobsSettings, updateJobSettings, cancelJobSettings, - deleteJobSettings); + deleteJobSettings, + submitJobAsOperationSettings); } // NEXT_MAJOR_VER: remove 'throws Exception' @@ -456,6 +521,19 @@ public UnaryCallSettings.Builder deleteJobSettings() { return deleteJobSettings; } + /** Returns the builder for the settings used for calls to submitJobAsOperation. */ + public UnaryCallSettings.Builder submitJobAsOperationSettings() { + return submitJobAsOperationSettings; + } + + /** Returns the builder for the settings used for calls to submitJobAsOperation. */ + @BetaApi( + "The surface for use by generated code is not stable yet and may change in the future.") + public OperationCallSettings.Builder + submitJobAsOperationOperationSettings() { + return submitJobAsOperationOperationSettings; + } + @Override public JobControllerStubSettings build() throws IOException { return new JobControllerStubSettings(this); diff --git a/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/JobControllerClientTest.java b/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/JobControllerClientTest.java index f8dec138..bbd73369 100644 --- a/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/JobControllerClientTest.java +++ b/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/JobControllerClientTest.java @@ -24,8 +24,11 @@ import com.google.api.gax.grpc.testing.MockServiceHelper; import com.google.api.gax.rpc.ApiClientHeaderProvider; import com.google.api.gax.rpc.InvalidArgumentException; +import com.google.api.gax.rpc.StatusCode; import com.google.common.collect.Lists; +import com.google.longrunning.Operation; import com.google.protobuf.AbstractMessage; +import com.google.protobuf.Any; import com.google.protobuf.Empty; import com.google.protobuf.FieldMask; import io.grpc.Status; @@ -34,6 +37,7 @@ import java.util.Arrays; import java.util.List; import java.util.UUID; +import java.util.concurrent.ExecutionException; import org.junit.After; import org.junit.AfterClass; import org.junit.Assert; @@ -96,11 +100,13 @@ public void submitJobTest() { String driverOutputResourceUri = "driverOutputResourceUri-542229086"; String driverControlFilesUri = "driverControlFilesUri207057643"; String jobUuid = "jobUuid-1615012099"; + boolean done = true; Job expectedResponse = Job.newBuilder() .setDriverOutputResourceUri(driverOutputResourceUri) .setDriverControlFilesUri(driverControlFilesUri) .setJobUuid(jobUuid) + .setDone(done) .build(); mockJobController.addResponse(expectedResponse); @@ -148,11 +154,13 @@ public void getJobTest() { String driverOutputResourceUri = "driverOutputResourceUri-542229086"; String driverControlFilesUri = "driverControlFilesUri207057643"; String jobUuid = "jobUuid-1615012099"; + boolean done = true; Job expectedResponse = Job.newBuilder() .setDriverOutputResourceUri(driverOutputResourceUri) .setDriverControlFilesUri(driverControlFilesUri) .setJobUuid(jobUuid) + .setDone(done) .build(); mockJobController.addResponse(expectedResponse); @@ -296,11 +304,13 @@ public void updateJobTest() { String driverOutputResourceUri = "driverOutputResourceUri-542229086"; String driverControlFilesUri = "driverControlFilesUri207057643"; String jobUuid = "jobUuid-1615012099"; + boolean done = true; Job expectedResponse = Job.newBuilder() .setDriverOutputResourceUri(driverOutputResourceUri) .setDriverControlFilesUri(driverControlFilesUri) .setJobUuid(jobUuid) + .setDone(done) .build(); mockJobController.addResponse(expectedResponse); @@ -370,11 +380,13 @@ public void cancelJobTest() { String driverOutputResourceUri = "driverOutputResourceUri-542229086"; String driverControlFilesUri = "driverControlFilesUri207057643"; String jobUuid = "jobUuid-1615012099"; + boolean done = true; Job expectedResponse = Job.newBuilder() .setDriverOutputResourceUri(driverOutputResourceUri) .setDriverControlFilesUri(driverControlFilesUri) .setJobUuid(jobUuid) + .setDone(done) .build(); mockJobController.addResponse(expectedResponse); @@ -458,4 +470,66 @@ public void deleteJobExceptionTest() throws Exception { // Expected exception } } + + @Test + @SuppressWarnings("all") + public void submitJobAsOperationTest() throws Exception { + String driverOutputResourceUri = "driverOutputResourceUri-542229086"; + String driverControlFilesUri = "driverControlFilesUri207057643"; + String jobUuid = "jobUuid-1615012099"; + boolean done = true; + Job expectedResponse = + Job.newBuilder() + .setDriverOutputResourceUri(driverOutputResourceUri) + .setDriverControlFilesUri(driverControlFilesUri) + .setJobUuid(jobUuid) + .setDone(done) + .build(); + Operation resultOperation = + Operation.newBuilder() + .setName("submitJobAsOperationTest") + .setDone(true) + .setResponse(Any.pack(expectedResponse)) + .build(); + mockJobController.addResponse(resultOperation); + + String projectId = "projectId-1969970175"; + String region = "region-934795532"; + Job job = Job.newBuilder().build(); + + Job actualResponse = client.submitJobAsOperationAsync(projectId, region, job).get(); + Assert.assertEquals(expectedResponse, actualResponse); + + List actualRequests = mockJobController.getRequests(); + Assert.assertEquals(1, actualRequests.size()); + SubmitJobRequest actualRequest = (SubmitJobRequest) actualRequests.get(0); + + Assert.assertEquals(projectId, actualRequest.getProjectId()); + Assert.assertEquals(region, actualRequest.getRegion()); + Assert.assertEquals(job, actualRequest.getJob()); + Assert.assertTrue( + channelProvider.isHeaderSent( + ApiClientHeaderProvider.getDefaultApiClientHeaderKey(), + GaxGrpcProperties.getDefaultApiClientHeaderPattern())); + } + + @Test + @SuppressWarnings("all") + public void submitJobAsOperationExceptionTest() throws Exception { + StatusRuntimeException exception = new StatusRuntimeException(Status.INVALID_ARGUMENT); + mockJobController.addException(exception); + + try { + String projectId = "projectId-1969970175"; + String region = "region-934795532"; + Job job = Job.newBuilder().build(); + + client.submitJobAsOperationAsync(projectId, region, job).get(); + Assert.fail("No exception raised"); + } catch (ExecutionException e) { + Assert.assertEquals(InvalidArgumentException.class, e.getCause().getClass()); + InvalidArgumentException apiException = (InvalidArgumentException) e.getCause(); + Assert.assertEquals(StatusCode.Code.INVALID_ARGUMENT, apiException.getStatusCode().getCode()); + } + } } diff --git a/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockJobControllerImpl.java b/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockJobControllerImpl.java index 4298b284..93f5ed54 100644 --- a/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockJobControllerImpl.java +++ b/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1/MockJobControllerImpl.java @@ -17,6 +17,7 @@ import com.google.api.core.BetaApi; import com.google.cloud.dataproc.v1.JobControllerGrpc.JobControllerImplBase; +import com.google.longrunning.Operation; import com.google.protobuf.AbstractMessage; import com.google.protobuf.Empty; import io.grpc.stub.StreamObserver; @@ -71,6 +72,21 @@ public void submitJob(SubmitJobRequest request, StreamObserver responseObse } } + @Override + public void submitJobAsOperation( + SubmitJobRequest request, StreamObserver responseObserver) { + Object response = responses.remove(); + if (response instanceof Operation) { + requests.add(request); + responseObserver.onNext((Operation) response); + responseObserver.onCompleted(); + } else if (response instanceof Exception) { + responseObserver.onError((Exception) response); + } else { + responseObserver.onError(new IllegalArgumentException("Unrecognized response type")); + } + } + @Override public void getJob(GetJobRequest request, StreamObserver responseObserver) { Object response = responses.remove(); diff --git a/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/JobControllerClientTest.java b/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/JobControllerClientTest.java index 64e4cbd8..7dce00fc 100644 --- a/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/JobControllerClientTest.java +++ b/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/JobControllerClientTest.java @@ -24,8 +24,11 @@ import com.google.api.gax.grpc.testing.MockServiceHelper; import com.google.api.gax.rpc.ApiClientHeaderProvider; import com.google.api.gax.rpc.InvalidArgumentException; +import com.google.api.gax.rpc.StatusCode; import com.google.common.collect.Lists; +import com.google.longrunning.Operation; import com.google.protobuf.AbstractMessage; +import com.google.protobuf.Any; import com.google.protobuf.Empty; import com.google.protobuf.FieldMask; import io.grpc.Status; @@ -34,6 +37,7 @@ import java.util.Arrays; import java.util.List; import java.util.UUID; +import java.util.concurrent.ExecutionException; import org.junit.After; import org.junit.AfterClass; import org.junit.Assert; @@ -97,12 +101,14 @@ public void submitJobTest() { String driverOutputResourceUri = "driverOutputResourceUri-542229086"; String driverControlFilesUri = "driverControlFilesUri207057643"; String jobUuid = "jobUuid-1615012099"; + boolean done = true; Job expectedResponse = Job.newBuilder() .setSubmittedBy(submittedBy) .setDriverOutputResourceUri(driverOutputResourceUri) .setDriverControlFilesUri(driverControlFilesUri) .setJobUuid(jobUuid) + .setDone(done) .build(); mockJobController.addResponse(expectedResponse); @@ -151,12 +157,14 @@ public void getJobTest() { String driverOutputResourceUri = "driverOutputResourceUri-542229086"; String driverControlFilesUri = "driverControlFilesUri207057643"; String jobUuid = "jobUuid-1615012099"; + boolean done = true; Job expectedResponse = Job.newBuilder() .setSubmittedBy(submittedBy) .setDriverOutputResourceUri(driverOutputResourceUri) .setDriverControlFilesUri(driverControlFilesUri) .setJobUuid(jobUuid) + .setDone(done) .build(); mockJobController.addResponse(expectedResponse); @@ -301,12 +309,14 @@ public void updateJobTest() { String driverOutputResourceUri = "driverOutputResourceUri-542229086"; String driverControlFilesUri = "driverControlFilesUri207057643"; String jobUuid = "jobUuid-1615012099"; + boolean done = true; Job expectedResponse = Job.newBuilder() .setSubmittedBy(submittedBy) .setDriverOutputResourceUri(driverOutputResourceUri) .setDriverControlFilesUri(driverControlFilesUri) .setJobUuid(jobUuid) + .setDone(done) .build(); mockJobController.addResponse(expectedResponse); @@ -377,12 +387,14 @@ public void cancelJobTest() { String driverOutputResourceUri = "driverOutputResourceUri-542229086"; String driverControlFilesUri = "driverControlFilesUri207057643"; String jobUuid = "jobUuid-1615012099"; + boolean done = true; Job expectedResponse = Job.newBuilder() .setSubmittedBy(submittedBy) .setDriverOutputResourceUri(driverOutputResourceUri) .setDriverControlFilesUri(driverControlFilesUri) .setJobUuid(jobUuid) + .setDone(done) .build(); mockJobController.addResponse(expectedResponse); @@ -466,4 +478,68 @@ public void deleteJobExceptionTest() throws Exception { // Expected exception } } + + @Test + @SuppressWarnings("all") + public void submitJobAsOperationTest() throws Exception { + String submittedBy = "submittedBy-2047729125"; + String driverOutputResourceUri = "driverOutputResourceUri-542229086"; + String driverControlFilesUri = "driverControlFilesUri207057643"; + String jobUuid = "jobUuid-1615012099"; + boolean done = true; + Job expectedResponse = + Job.newBuilder() + .setSubmittedBy(submittedBy) + .setDriverOutputResourceUri(driverOutputResourceUri) + .setDriverControlFilesUri(driverControlFilesUri) + .setJobUuid(jobUuid) + .setDone(done) + .build(); + Operation resultOperation = + Operation.newBuilder() + .setName("submitJobAsOperationTest") + .setDone(true) + .setResponse(Any.pack(expectedResponse)) + .build(); + mockJobController.addResponse(resultOperation); + + String projectId = "projectId-1969970175"; + String region = "region-934795532"; + Job job = Job.newBuilder().build(); + + Job actualResponse = client.submitJobAsOperationAsync(projectId, region, job).get(); + Assert.assertEquals(expectedResponse, actualResponse); + + List actualRequests = mockJobController.getRequests(); + Assert.assertEquals(1, actualRequests.size()); + SubmitJobRequest actualRequest = (SubmitJobRequest) actualRequests.get(0); + + Assert.assertEquals(projectId, actualRequest.getProjectId()); + Assert.assertEquals(region, actualRequest.getRegion()); + Assert.assertEquals(job, actualRequest.getJob()); + Assert.assertTrue( + channelProvider.isHeaderSent( + ApiClientHeaderProvider.getDefaultApiClientHeaderKey(), + GaxGrpcProperties.getDefaultApiClientHeaderPattern())); + } + + @Test + @SuppressWarnings("all") + public void submitJobAsOperationExceptionTest() throws Exception { + StatusRuntimeException exception = new StatusRuntimeException(Status.INVALID_ARGUMENT); + mockJobController.addException(exception); + + try { + String projectId = "projectId-1969970175"; + String region = "region-934795532"; + Job job = Job.newBuilder().build(); + + client.submitJobAsOperationAsync(projectId, region, job).get(); + Assert.fail("No exception raised"); + } catch (ExecutionException e) { + Assert.assertEquals(InvalidArgumentException.class, e.getCause().getClass()); + InvalidArgumentException apiException = (InvalidArgumentException) e.getCause(); + Assert.assertEquals(StatusCode.Code.INVALID_ARGUMENT, apiException.getStatusCode().getCode()); + } + } } diff --git a/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockJobControllerImpl.java b/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockJobControllerImpl.java index 01a1d26c..c8c5c0a8 100644 --- a/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockJobControllerImpl.java +++ b/google-cloud-dataproc/src/test/java/com/google/cloud/dataproc/v1beta2/MockJobControllerImpl.java @@ -17,6 +17,7 @@ import com.google.api.core.BetaApi; import com.google.cloud.dataproc.v1beta2.JobControllerGrpc.JobControllerImplBase; +import com.google.longrunning.Operation; import com.google.protobuf.AbstractMessage; import com.google.protobuf.Empty; import io.grpc.stub.StreamObserver; @@ -71,6 +72,21 @@ public void submitJob(SubmitJobRequest request, StreamObserver responseObse } } + @Override + public void submitJobAsOperation( + SubmitJobRequest request, StreamObserver responseObserver) { + Object response = responses.remove(); + if (response instanceof Operation) { + requests.add(request); + responseObserver.onNext((Operation) response); + responseObserver.onCompleted(); + } else if (response instanceof Exception) { + responseObserver.onError((Exception) response); + } else { + responseObserver.onError(new IllegalArgumentException("Unrecognized response type")); + } + } + @Override public void getJob(GetJobRequest request, StreamObserver responseObserver) { Object response = responses.remove(); diff --git a/grpc-google-cloud-dataproc-v1/clirr-ignored-differences.xml b/grpc-google-cloud-dataproc-v1/clirr-ignored-differences.xml deleted file mode 100644 index 192dad2e..00000000 --- a/grpc-google-cloud-dataproc-v1/clirr-ignored-differences.xml +++ /dev/null @@ -1,10 +0,0 @@ - - - - - - 6001 - com/google/cloud/dataproc/v1/*Grpc - METHOD_* - - \ No newline at end of file diff --git a/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerGrpc.java b/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerGrpc.java index f211cfc9..7a2eb402 100644 --- a/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerGrpc.java +++ b/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterControllerGrpc.java @@ -414,7 +414,7 @@ public void getCluster( * * *
-     * Lists all regions/{region}/clusters in a project.
+     * Lists all regions/{region}/clusters in a project alphabetically.
      * 
*/ public void listClusters( @@ -579,7 +579,7 @@ public void getCluster( * * *
-     * Lists all regions/{region}/clusters in a project.
+     * Lists all regions/{region}/clusters in a project alphabetically.
      * 
*/ public void listClusters( @@ -694,7 +694,7 @@ public com.google.cloud.dataproc.v1.Cluster getCluster( * * *
-     * Lists all regions/{region}/clusters in a project.
+     * Lists all regions/{region}/clusters in a project alphabetically.
      * 
*/ public com.google.cloud.dataproc.v1.ListClustersResponse listClusters( @@ -803,7 +803,7 @@ protected ClusterControllerFutureStub build( * * *
-     * Lists all regions/{region}/clusters in a project.
+     * Lists all regions/{region}/clusters in a project alphabetically.
      * 
*/ public com.google.common.util.concurrent.ListenableFuture< diff --git a/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobControllerGrpc.java b/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobControllerGrpc.java index 7e79627c..749d4504 100644 --- a/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobControllerGrpc.java +++ b/grpc-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobControllerGrpc.java @@ -80,6 +80,50 @@ private JobControllerGrpc() {} return getSubmitJobMethod; } + private static volatile io.grpc.MethodDescriptor< + com.google.cloud.dataproc.v1.SubmitJobRequest, com.google.longrunning.Operation> + getSubmitJobAsOperationMethod; + + @io.grpc.stub.annotations.RpcMethod( + fullMethodName = SERVICE_NAME + '/' + "SubmitJobAsOperation", + requestType = com.google.cloud.dataproc.v1.SubmitJobRequest.class, + responseType = com.google.longrunning.Operation.class, + methodType = io.grpc.MethodDescriptor.MethodType.UNARY) + public static io.grpc.MethodDescriptor< + com.google.cloud.dataproc.v1.SubmitJobRequest, com.google.longrunning.Operation> + getSubmitJobAsOperationMethod() { + io.grpc.MethodDescriptor< + com.google.cloud.dataproc.v1.SubmitJobRequest, com.google.longrunning.Operation> + getSubmitJobAsOperationMethod; + if ((getSubmitJobAsOperationMethod = JobControllerGrpc.getSubmitJobAsOperationMethod) == null) { + synchronized (JobControllerGrpc.class) { + if ((getSubmitJobAsOperationMethod = JobControllerGrpc.getSubmitJobAsOperationMethod) + == null) { + JobControllerGrpc.getSubmitJobAsOperationMethod = + getSubmitJobAsOperationMethod = + io.grpc.MethodDescriptor + . + newBuilder() + .setType(io.grpc.MethodDescriptor.MethodType.UNARY) + .setFullMethodName( + generateFullMethodName(SERVICE_NAME, "SubmitJobAsOperation")) + .setSampledToLocalTracing(true) + .setRequestMarshaller( + io.grpc.protobuf.ProtoUtils.marshaller( + com.google.cloud.dataproc.v1.SubmitJobRequest.getDefaultInstance())) + .setResponseMarshaller( + io.grpc.protobuf.ProtoUtils.marshaller( + com.google.longrunning.Operation.getDefaultInstance())) + .setSchemaDescriptor( + new JobControllerMethodDescriptorSupplier("SubmitJobAsOperation")) + .build(); + } + } + } + return getSubmitJobAsOperationMethod; + } + private static volatile io.grpc.MethodDescriptor< com.google.cloud.dataproc.v1.GetJobRequest, com.google.cloud.dataproc.v1.Job> getGetJobMethod; @@ -350,6 +394,19 @@ public void submitJob( asyncUnimplementedUnaryCall(getSubmitJobMethod(), responseObserver); } + /** + * + * + *
+     * Submits job to a cluster.
+     * 
+ */ + public void submitJobAsOperation( + com.google.cloud.dataproc.v1.SubmitJobRequest request, + io.grpc.stub.StreamObserver responseObserver) { + asyncUnimplementedUnaryCall(getSubmitJobAsOperationMethod(), responseObserver); + } + /** * * @@ -430,6 +487,12 @@ public final io.grpc.ServerServiceDefinition bindService() { new MethodHandlers< com.google.cloud.dataproc.v1.SubmitJobRequest, com.google.cloud.dataproc.v1.Job>(this, METHODID_SUBMIT_JOB))) + .addMethod( + getSubmitJobAsOperationMethod(), + asyncUnaryCall( + new MethodHandlers< + com.google.cloud.dataproc.v1.SubmitJobRequest, + com.google.longrunning.Operation>(this, METHODID_SUBMIT_JOB_AS_OPERATION))) .addMethod( getGetJobMethod(), asyncUnaryCall( @@ -496,6 +559,22 @@ public void submitJob( getChannel().newCall(getSubmitJobMethod(), getCallOptions()), request, responseObserver); } + /** + * + * + *
+     * Submits job to a cluster.
+     * 
+ */ + public void submitJobAsOperation( + com.google.cloud.dataproc.v1.SubmitJobRequest request, + io.grpc.stub.StreamObserver responseObserver) { + asyncUnaryCall( + getChannel().newCall(getSubmitJobAsOperationMethod(), getCallOptions()), + request, + responseObserver); + } + /** * * @@ -604,6 +683,19 @@ public com.google.cloud.dataproc.v1.Job submitJob( return blockingUnaryCall(getChannel(), getSubmitJobMethod(), getCallOptions(), request); } + /** + * + * + *
+     * Submits job to a cluster.
+     * 
+ */ + public com.google.longrunning.Operation submitJobAsOperation( + com.google.cloud.dataproc.v1.SubmitJobRequest request) { + return blockingUnaryCall( + getChannel(), getSubmitJobAsOperationMethod(), getCallOptions(), request); + } + /** * * @@ -701,6 +793,19 @@ protected JobControllerFutureStub build( return futureUnaryCall(getChannel().newCall(getSubmitJobMethod(), getCallOptions()), request); } + /** + * + * + *
+     * Submits job to a cluster.
+     * 
+ */ + public com.google.common.util.concurrent.ListenableFuture + submitJobAsOperation(com.google.cloud.dataproc.v1.SubmitJobRequest request) { + return futureUnaryCall( + getChannel().newCall(getSubmitJobAsOperationMethod(), getCallOptions()), request); + } + /** * * @@ -769,11 +874,12 @@ public com.google.common.util.concurrent.ListenableFuture implements io.grpc.stub.ServerCalls.UnaryMethod, @@ -797,6 +903,11 @@ public void invoke(Req request, io.grpc.stub.StreamObserver responseObserv (com.google.cloud.dataproc.v1.SubmitJobRequest) request, (io.grpc.stub.StreamObserver) responseObserver); break; + case METHODID_SUBMIT_JOB_AS_OPERATION: + serviceImpl.submitJobAsOperation( + (com.google.cloud.dataproc.v1.SubmitJobRequest) request, + (io.grpc.stub.StreamObserver) responseObserver); + break; case METHODID_GET_JOB: serviceImpl.getJob( (com.google.cloud.dataproc.v1.GetJobRequest) request, @@ -888,6 +999,7 @@ public static io.grpc.ServiceDescriptor getServiceDescriptor() { io.grpc.ServiceDescriptor.newBuilder(SERVICE_NAME) .setSchemaDescriptor(new JobControllerFileDescriptorSupplier()) .addMethod(getSubmitJobMethod()) + .addMethod(getSubmitJobAsOperationMethod()) .addMethod(getGetJobMethod()) .addMethod(getListJobsMethod()) .addMethod(getUpdateJobMethod()) diff --git a/grpc-google-cloud-dataproc-v1beta2/clirr-ignored-differences.xml b/grpc-google-cloud-dataproc-v1beta2/clirr-ignored-differences.xml deleted file mode 100644 index 323afd96..00000000 --- a/grpc-google-cloud-dataproc-v1beta2/clirr-ignored-differences.xml +++ /dev/null @@ -1,10 +0,0 @@ - - - - - - 6001 - com/google/cloud/dataproc/v1beta2/*Grpc - METHOD_* - - \ No newline at end of file diff --git a/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerGrpc.java b/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerGrpc.java index ee0dbde8..7acc1f4a 100644 --- a/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerGrpc.java +++ b/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterControllerGrpc.java @@ -424,7 +424,7 @@ public void getCluster( * * *
-     * Lists all regions/{region}/clusters in a project.
+     * Lists all regions/{region}/clusters in a project alphabetically.
      * 
*/ public void listClusters( @@ -589,7 +589,7 @@ public void getCluster( * * *
-     * Lists all regions/{region}/clusters in a project.
+     * Lists all regions/{region}/clusters in a project alphabetically.
      * 
*/ public void listClusters( @@ -704,7 +704,7 @@ public com.google.cloud.dataproc.v1beta2.Cluster getCluster( * * *
-     * Lists all regions/{region}/clusters in a project.
+     * Lists all regions/{region}/clusters in a project alphabetically.
      * 
*/ public com.google.cloud.dataproc.v1beta2.ListClustersResponse listClusters( @@ -814,7 +814,7 @@ protected ClusterControllerFutureStub build( * * *
-     * Lists all regions/{region}/clusters in a project.
+     * Lists all regions/{region}/clusters in a project alphabetically.
      * 
*/ public com.google.common.util.concurrent.ListenableFuture< diff --git a/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerGrpc.java b/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerGrpc.java index b1cf41b7..b0fdbc1f 100644 --- a/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerGrpc.java +++ b/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobControllerGrpc.java @@ -82,6 +82,51 @@ private JobControllerGrpc() {} return getSubmitJobMethod; } + private static volatile io.grpc.MethodDescriptor< + com.google.cloud.dataproc.v1beta2.SubmitJobRequest, com.google.longrunning.Operation> + getSubmitJobAsOperationMethod; + + @io.grpc.stub.annotations.RpcMethod( + fullMethodName = SERVICE_NAME + '/' + "SubmitJobAsOperation", + requestType = com.google.cloud.dataproc.v1beta2.SubmitJobRequest.class, + responseType = com.google.longrunning.Operation.class, + methodType = io.grpc.MethodDescriptor.MethodType.UNARY) + public static io.grpc.MethodDescriptor< + com.google.cloud.dataproc.v1beta2.SubmitJobRequest, com.google.longrunning.Operation> + getSubmitJobAsOperationMethod() { + io.grpc.MethodDescriptor< + com.google.cloud.dataproc.v1beta2.SubmitJobRequest, com.google.longrunning.Operation> + getSubmitJobAsOperationMethod; + if ((getSubmitJobAsOperationMethod = JobControllerGrpc.getSubmitJobAsOperationMethod) == null) { + synchronized (JobControllerGrpc.class) { + if ((getSubmitJobAsOperationMethod = JobControllerGrpc.getSubmitJobAsOperationMethod) + == null) { + JobControllerGrpc.getSubmitJobAsOperationMethod = + getSubmitJobAsOperationMethod = + io.grpc.MethodDescriptor + . + newBuilder() + .setType(io.grpc.MethodDescriptor.MethodType.UNARY) + .setFullMethodName( + generateFullMethodName(SERVICE_NAME, "SubmitJobAsOperation")) + .setSampledToLocalTracing(true) + .setRequestMarshaller( + io.grpc.protobuf.ProtoUtils.marshaller( + com.google.cloud.dataproc.v1beta2.SubmitJobRequest + .getDefaultInstance())) + .setResponseMarshaller( + io.grpc.protobuf.ProtoUtils.marshaller( + com.google.longrunning.Operation.getDefaultInstance())) + .setSchemaDescriptor( + new JobControllerMethodDescriptorSupplier("SubmitJobAsOperation")) + .build(); + } + } + } + return getSubmitJobAsOperationMethod; + } + private static volatile io.grpc.MethodDescriptor< com.google.cloud.dataproc.v1beta2.GetJobRequest, com.google.cloud.dataproc.v1beta2.Job> getGetJobMethod; @@ -360,6 +405,19 @@ public void submitJob( asyncUnimplementedUnaryCall(getSubmitJobMethod(), responseObserver); } + /** + * + * + *
+     * Submits job to a cluster.
+     * 
+ */ + public void submitJobAsOperation( + com.google.cloud.dataproc.v1beta2.SubmitJobRequest request, + io.grpc.stub.StreamObserver responseObserver) { + asyncUnimplementedUnaryCall(getSubmitJobAsOperationMethod(), responseObserver); + } + /** * * @@ -440,6 +498,12 @@ public final io.grpc.ServerServiceDefinition bindService() { new MethodHandlers< com.google.cloud.dataproc.v1beta2.SubmitJobRequest, com.google.cloud.dataproc.v1beta2.Job>(this, METHODID_SUBMIT_JOB))) + .addMethod( + getSubmitJobAsOperationMethod(), + asyncUnaryCall( + new MethodHandlers< + com.google.cloud.dataproc.v1beta2.SubmitJobRequest, + com.google.longrunning.Operation>(this, METHODID_SUBMIT_JOB_AS_OPERATION))) .addMethod( getGetJobMethod(), asyncUnaryCall( @@ -507,6 +571,22 @@ public void submitJob( getChannel().newCall(getSubmitJobMethod(), getCallOptions()), request, responseObserver); } + /** + * + * + *
+     * Submits job to a cluster.
+     * 
+ */ + public void submitJobAsOperation( + com.google.cloud.dataproc.v1beta2.SubmitJobRequest request, + io.grpc.stub.StreamObserver responseObserver) { + asyncUnaryCall( + getChannel().newCall(getSubmitJobAsOperationMethod(), getCallOptions()), + request, + responseObserver); + } + /** * * @@ -615,6 +695,19 @@ public com.google.cloud.dataproc.v1beta2.Job submitJob( return blockingUnaryCall(getChannel(), getSubmitJobMethod(), getCallOptions(), request); } + /** + * + * + *
+     * Submits job to a cluster.
+     * 
+ */ + public com.google.longrunning.Operation submitJobAsOperation( + com.google.cloud.dataproc.v1beta2.SubmitJobRequest request) { + return blockingUnaryCall( + getChannel(), getSubmitJobAsOperationMethod(), getCallOptions(), request); + } + /** * * @@ -712,6 +805,19 @@ protected JobControllerFutureStub build( return futureUnaryCall(getChannel().newCall(getSubmitJobMethod(), getCallOptions()), request); } + /** + * + * + *
+     * Submits job to a cluster.
+     * 
+ */ + public com.google.common.util.concurrent.ListenableFuture + submitJobAsOperation(com.google.cloud.dataproc.v1beta2.SubmitJobRequest request) { + return futureUnaryCall( + getChannel().newCall(getSubmitJobAsOperationMethod(), getCallOptions()), request); + } + /** * * @@ -780,11 +886,12 @@ public com.google.common.util.concurrent.ListenableFuture implements io.grpc.stub.ServerCalls.UnaryMethod, @@ -809,6 +916,11 @@ public void invoke(Req request, io.grpc.stub.StreamObserver responseObserv (io.grpc.stub.StreamObserver) responseObserver); break; + case METHODID_SUBMIT_JOB_AS_OPERATION: + serviceImpl.submitJobAsOperation( + (com.google.cloud.dataproc.v1beta2.SubmitJobRequest) request, + (io.grpc.stub.StreamObserver) responseObserver); + break; case METHODID_GET_JOB: serviceImpl.getJob( (com.google.cloud.dataproc.v1beta2.GetJobRequest) request, @@ -903,6 +1015,7 @@ public static io.grpc.ServiceDescriptor getServiceDescriptor() { io.grpc.ServiceDescriptor.newBuilder(SERVICE_NAME) .setSchemaDescriptor(new JobControllerFileDescriptorSupplier()) .addMethod(getSubmitJobMethod()) + .addMethod(getSubmitJobAsOperationMethod()) .addMethod(getGetJobMethod()) .addMethod(getListJobsMethod()) .addMethod(getUpdateJobMethod()) diff --git a/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateServiceGrpc.java b/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateServiceGrpc.java index 0909a3e9..0aeb05ba 100644 --- a/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateServiceGrpc.java +++ b/grpc-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplateServiceGrpc.java @@ -523,8 +523,7 @@ public void instantiateWorkflowTemplate( *
      * Instantiates a template and begins execution.
      * This method is equivalent to executing the sequence
-     * [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate],
-     * [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate],
+     * [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate],
      * [DeleteWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.DeleteWorkflowTemplate].
      * The returned Operation can be used to track execution of
      * workflow by polling
@@ -739,8 +738,7 @@ public void instantiateWorkflowTemplate(
      * 
      * Instantiates a template and begins execution.
      * This method is equivalent to executing the sequence
-     * [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate],
-     * [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate],
+     * [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate],
      * [DeleteWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.DeleteWorkflowTemplate].
      * The returned Operation can be used to track execution of
      * workflow by polling
@@ -903,8 +901,7 @@ public com.google.longrunning.Operation instantiateWorkflowTemplate(
      * 
      * Instantiates a template and begins execution.
      * This method is equivalent to executing the sequence
-     * [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate],
-     * [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate],
+     * [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate],
      * [DeleteWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.DeleteWorkflowTemplate].
      * The returned Operation can be used to track execution of
      * workflow by polling
@@ -1057,8 +1054,7 @@ protected WorkflowTemplateServiceFutureStub build(
      * 
      * Instantiates a template and begins execution.
      * This method is equivalent to executing the sequence
-     * [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate],
-     * [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate],
+     * [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate],
      * [DeleteWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.DeleteWorkflowTemplate].
      * The returned Operation can be used to track execution of
      * workflow by polling
diff --git a/proto-google-cloud-dataproc-v1/clirr-ignored-differences.xml b/proto-google-cloud-dataproc-v1/clirr-ignored-differences.xml
index b045425c..fa53d8b6 100644
--- a/proto-google-cloud-dataproc-v1/clirr-ignored-differences.xml
+++ b/proto-google-cloud-dataproc-v1/clirr-ignored-differences.xml
@@ -16,9 +16,4 @@
     com/google/cloud/dataproc/v1/*OrBuilder
     boolean has*(*)
   
-  
-    7002
-    com/google/cloud/dataproc/v1/AutoscalingPolicyName$Builder
-    * *etRegion(*)
-  
 
diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java
index 2d641c5f..001a4284 100644
--- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java
+++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfig.java
@@ -23,7 +23,8 @@
  *
  * 
  * Specifies the type and number of accelerator cards attached to the instances
- * of an instance. See [GPUs on Compute Engine](/compute/docs/gpus/).
+ * of an instance. See [GPUs on Compute
+ * Engine](https://cloud.google.com/compute/docs/gpus/).
  * 
* * Protobuf type {@code google.cloud.dataproc.v1.AcceleratorConfig} @@ -126,14 +127,14 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. *
@@ -160,14 +161,14 @@ public java.lang.String getAcceleratorTypeUri() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. *
@@ -378,7 +379,8 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * *
    * Specifies the type and number of accelerator cards attached to the instances
-   * of an instance. See [GPUs on Compute Engine](/compute/docs/gpus/).
+   * of an instance. See [GPUs on Compute
+   * Engine](https://cloud.google.com/compute/docs/gpus/).
    * 
* * Protobuf type {@code google.cloud.dataproc.v1.AcceleratorConfig} @@ -545,14 +547,14 @@ public Builder mergeFrom( * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. *
@@ -579,14 +581,14 @@ public java.lang.String getAcceleratorTypeUri() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. *
@@ -613,14 +615,14 @@ public com.google.protobuf.ByteString getAcceleratorTypeUriBytes() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. *
@@ -646,14 +648,14 @@ public Builder setAcceleratorTypeUri(java.lang.String value) { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. *
@@ -675,14 +677,14 @@ public Builder clearAcceleratorTypeUri() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java index 2661c639..5e557214 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AcceleratorConfigOrBuilder.java @@ -30,14 +30,14 @@ public interface AcceleratorConfigOrBuilder * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * @@ -54,14 +54,14 @@ public interface AcceleratorConfigOrBuilder * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). * Examples: * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPoliciesProto.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPoliciesProto.java index a1126b55..7dda762c 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPoliciesProto.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/AutoscalingPoliciesProto.java @@ -109,67 +109,67 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "in_worker_fraction\030\004 \001(\001B\003\340A\001\"s\n$Instanc" + "eGroupAutoscalingPolicyConfig\022\032\n\rmin_ins" + "tances\030\001 \001(\005B\003\340A\001\022\032\n\rmax_instances\030\002 \001(\005" - + "B\003\340A\002\022\023\n\006weight\030\003 \001(\005B\003\340A\001\"\240\001\n\036CreateAut" + + "B\003\340A\002\022\023\n\006weight\030\003 \001(\005B\003\340A\001\"\245\001\n\036CreateAut" + "oscalingPolicyRequest\022A\n\006parent\030\001 \001(\tB1\340" + "A\002\372A+\022)dataproc.googleapis.com/Autoscali" - + "ngPolicy\022;\n\006policy\030\002 \001(\0132+.google.cloud." - + "dataproc.v1.AutoscalingPolicy\"^\n\033GetAuto" - + "scalingPolicyRequest\022?\n\004name\030\001 \001(\tB1\340A\002\372" - + "A+\n)dataproc.googleapis.com/AutoscalingP" - + "olicy\"b\n\036UpdateAutoscalingPolicyRequest\022" - + "@\n\006policy\030\001 \001(\0132+.google.cloud.dataproc." - + "v1.AutoscalingPolicyB\003\340A\002\"a\n\036DeleteAutos" - + "calingPolicyRequest\022?\n\004name\030\001 \001(\tB1\340A\002\372A" - + "+\n)dataproc.googleapis.com/AutoscalingPo" - + "licy\"\224\001\n\036ListAutoscalingPoliciesRequest\022" - + "A\n\006parent\030\001 \001(\tB1\340A\002\372A+\022)dataproc.google" - + "apis.com/AutoscalingPolicy\022\026\n\tpage_size\030" - + "\002 \001(\005B\003\340A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001\"\203\001\n\037" - + "ListAutoscalingPoliciesResponse\022B\n\010polic" - + "ies\030\001 \003(\0132+.google.cloud.dataproc.v1.Aut" - + "oscalingPolicyB\003\340A\003\022\034\n\017next_page_token\030\002" - + " \001(\tB\003\340A\0032\256\013\n\030AutoscalingPolicyService\022\234" - + "\002\n\027CreateAutoscalingPolicy\0228.google.clou" - + "d.dataproc.v1.CreateAutoscalingPolicyReq" - + "uest\032+.google.cloud.dataproc.v1.Autoscal" - + "ingPolicy\"\231\001\202\323\344\223\002\202\001\"7/v1/{parent=project" - + "s/*/locations/*}/autoscalingPolicies:\006po" - + "licyZ?\"5/v1/{parent=projects/*/regions/*" - + "}/autoscalingPolicies:\006policy\332A\rparent,p" - + "olicy\022\243\002\n\027UpdateAutoscalingPolicy\0228.goog" - + "le.cloud.dataproc.v1.UpdateAutoscalingPo" - + "licyRequest\032+.google.cloud.dataproc.v1.A" - + "utoscalingPolicy\"\240\001\202\323\344\223\002\220\001\032>/v1/{policy." + + "ngPolicy\022@\n\006policy\030\002 \001(\0132+.google.cloud." + + "dataproc.v1.AutoscalingPolicyB\003\340A\002\"^\n\033Ge" + + "tAutoscalingPolicyRequest\022?\n\004name\030\001 \001(\tB" + + "1\340A\002\372A+\n)dataproc.googleapis.com/Autosca" + + "lingPolicy\"b\n\036UpdateAutoscalingPolicyReq" + + "uest\022@\n\006policy\030\001 \001(\0132+.google.cloud.data" + + "proc.v1.AutoscalingPolicyB\003\340A\002\"a\n\036Delete" + + "AutoscalingPolicyRequest\022?\n\004name\030\001 \001(\tB1" + + "\340A\002\372A+\n)dataproc.googleapis.com/Autoscal" + + "ingPolicy\"\224\001\n\036ListAutoscalingPoliciesReq" + + "uest\022A\n\006parent\030\001 \001(\tB1\340A\002\372A+\022)dataproc.g" + + "oogleapis.com/AutoscalingPolicy\022\026\n\tpage_" + + "size\030\002 \001(\005B\003\340A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001" + + "\"\203\001\n\037ListAutoscalingPoliciesResponse\022B\n\010" + + "policies\030\001 \003(\0132+.google.cloud.dataproc.v" + + "1.AutoscalingPolicyB\003\340A\003\022\034\n\017next_page_to" + + "ken\030\002 \001(\tB\003\340A\0032\256\013\n\030AutoscalingPolicyServ" + + "ice\022\234\002\n\027CreateAutoscalingPolicy\0228.google" + + ".cloud.dataproc.v1.CreateAutoscalingPoli" + + "cyRequest\032+.google.cloud.dataproc.v1.Aut" + + "oscalingPolicy\"\231\001\202\323\344\223\002\202\001\"7/v1/{parent=pr" + + "ojects/*/locations/*}/autoscalingPolicie" + + "s:\006policyZ?\"5/v1/{parent=projects/*/regi" + + "ons/*}/autoscalingPolicies:\006policy\332A\rpar" + + "ent,policy\022\243\002\n\027UpdateAutoscalingPolicy\0228" + + ".google.cloud.dataproc.v1.UpdateAutoscal" + + "ingPolicyRequest\032+.google.cloud.dataproc" + + ".v1.AutoscalingPolicy\"\240\001\202\323\344\223\002\220\001\032>/v1/{po" + + "licy.name=projects/*/locations/*/autosca" + + "lingPolicies/*}:\006policyZF\032 - * Contains cluster daemon metrics such as HDFS and YARN stats. + * Output only. Contains cluster daemon metrics such as HDFS and YARN stats. * **Beta Feature**: This report is available for testing purposes only. It * may be changed before final release. * * - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * * * @return Whether the metrics field is set. */ @@ -680,12 +682,14 @@ public boolean hasMetrics() { * * *
-   * Contains cluster daemon metrics such as HDFS and YARN stats.
+   * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
    * **Beta Feature**: This report is available for testing purposes only. It
    * may be changed before final release.
    * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * * * @return The metrics. */ @@ -698,12 +702,14 @@ public com.google.cloud.dataproc.v1.ClusterMetrics getMetrics() { * * *
-   * Contains cluster daemon metrics such as HDFS and YARN stats.
+   * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
    * **Beta Feature**: This report is available for testing purposes only. It
    * may be changed before final release.
    * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ public com.google.cloud.dataproc.v1.ClusterMetricsOrBuilder getMetricsOrBuilder() { return getMetrics(); @@ -2576,12 +2582,14 @@ public Builder setClusterUuidBytes(com.google.protobuf.ByteString value) { * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * * * @return Whether the metrics field is set. */ @@ -2592,12 +2600,14 @@ public boolean hasMetrics() { * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * * * @return The metrics. */ @@ -2614,12 +2624,14 @@ public com.google.cloud.dataproc.v1.ClusterMetrics getMetrics() { * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ public Builder setMetrics(com.google.cloud.dataproc.v1.ClusterMetrics value) { if (metricsBuilder_ == null) { @@ -2638,12 +2650,14 @@ public Builder setMetrics(com.google.cloud.dataproc.v1.ClusterMetrics value) { * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ public Builder setMetrics(com.google.cloud.dataproc.v1.ClusterMetrics.Builder builderForValue) { if (metricsBuilder_ == null) { @@ -2659,12 +2673,14 @@ public Builder setMetrics(com.google.cloud.dataproc.v1.ClusterMetrics.Builder bu * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ public Builder mergeMetrics(com.google.cloud.dataproc.v1.ClusterMetrics value) { if (metricsBuilder_ == null) { @@ -2687,12 +2703,14 @@ public Builder mergeMetrics(com.google.cloud.dataproc.v1.ClusterMetrics value) { * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ public Builder clearMetrics() { if (metricsBuilder_ == null) { @@ -2709,12 +2727,14 @@ public Builder clearMetrics() { * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ public com.google.cloud.dataproc.v1.ClusterMetrics.Builder getMetricsBuilder() { @@ -2725,12 +2745,14 @@ public com.google.cloud.dataproc.v1.ClusterMetrics.Builder getMetricsBuilder() { * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ public com.google.cloud.dataproc.v1.ClusterMetricsOrBuilder getMetricsOrBuilder() { if (metricsBuilder_ != null) { @@ -2745,12 +2767,14 @@ public com.google.cloud.dataproc.v1.ClusterMetricsOrBuilder getMetricsOrBuilder( * * *
-     * Contains cluster daemon metrics such as HDFS and YARN stats.
+     * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
      * **Beta Feature**: This report is available for testing purposes only. It
      * may be changed before final release.
      * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1.ClusterMetrics, diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java index 07936bde..3d5e8524 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfig.java @@ -288,7 +288,7 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -318,7 +318,7 @@ public java.lang.String getConfigBucket() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1586,7 +1586,7 @@ public Builder mergeFrom( * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1616,7 +1616,7 @@ public java.lang.String getConfigBucket() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1646,7 +1646,7 @@ public com.google.protobuf.ByteString getConfigBucketBytes() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1675,7 +1675,7 @@ public Builder setConfigBucket(java.lang.String value) { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1700,7 +1700,7 @@ public Builder clearConfigBucket() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java index 0cc4b2d4..76f2b5e8 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterConfigOrBuilder.java @@ -35,7 +35,7 @@ public interface ClusterConfigOrBuilder * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -55,7 +55,7 @@ public interface ClusterConfigOrBuilder * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java index 9eeb72f0..daaa60c4 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClusterOrBuilder.java @@ -337,12 +337,14 @@ public interface ClusterOrBuilder * * *
-   * Contains cluster daemon metrics such as HDFS and YARN stats.
+   * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
    * **Beta Feature**: This report is available for testing purposes only. It
    * may be changed before final release.
    * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * * * @return Whether the metrics field is set. */ @@ -351,12 +353,14 @@ public interface ClusterOrBuilder * * *
-   * Contains cluster daemon metrics such as HDFS and YARN stats.
+   * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
    * **Beta Feature**: This report is available for testing purposes only. It
    * may be changed before final release.
    * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * * * @return The metrics. */ @@ -365,12 +369,14 @@ public interface ClusterOrBuilder * * *
-   * Contains cluster daemon metrics such as HDFS and YARN stats.
+   * Output only. Contains cluster daemon metrics such as HDFS and YARN stats.
    * **Beta Feature**: This report is available for testing purposes only. It
    * may be changed before final release.
    * 
* - * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9; + * + * .google.cloud.dataproc.v1.ClusterMetrics metrics = 9 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * */ com.google.cloud.dataproc.v1.ClusterMetricsOrBuilder getMetricsOrBuilder(); } diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java index 7dd95f73..2ea622c9 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/ClustersProto.java @@ -164,7 +164,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "le/longrunning/operations.proto\032\036google/" + "protobuf/duration.proto\032 google/protobuf" + "/field_mask.proto\032\037google/protobuf/times" - + "tamp.proto\"\310\003\n\007Cluster\022\027\n\nproject_id\030\001 \001" + + "tamp.proto\"\315\003\n\007Cluster\022\027\n\nproject_id\030\001 \001" + "(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\022<\n\006co" + "nfig\030\003 \001(\0132\'.google.cloud.dataproc.v1.Cl" + "usterConfigB\003\340A\002\022B\n\006labels\030\010 \003(\0132-.googl" @@ -173,186 +173,186 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "proc.v1.ClusterStatusB\003\340A\003\022D\n\016status_his" + "tory\030\007 \003(\0132\'.google.cloud.dataproc.v1.Cl" + "usterStatusB\003\340A\003\022\031\n\014cluster_uuid\030\006 \001(\tB\003" - + "\340A\003\0229\n\007metrics\030\t \001(\0132(.google.cloud.data" - + "proc.v1.ClusterMetrics\032-\n\013LabelsEntry\022\013\n" - + "\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\260\006\n\rCluste" - + "rConfig\022\032\n\rconfig_bucket\030\001 \001(\tB\003\340A\001\022K\n\022g" - + "ce_cluster_config\030\010 \001(\0132*.google.cloud.d" - + "ataproc.v1.GceClusterConfigB\003\340A\001\022I\n\rmast" - + "er_config\030\t \001(\0132-.google.cloud.dataproc." - + "v1.InstanceGroupConfigB\003\340A\001\022I\n\rworker_co" - + "nfig\030\n \001(\0132-.google.cloud.dataproc.v1.In" - + "stanceGroupConfigB\003\340A\001\022S\n\027secondary_work" - + "er_config\030\014 \001(\0132-.google.cloud.dataproc." - + "v1.InstanceGroupConfigB\003\340A\001\022F\n\017software_" - + "config\030\r \001(\0132(.google.cloud.dataproc.v1." - + "SoftwareConfigB\003\340A\001\022W\n\026initialization_ac" - + "tions\030\013 \003(\01322.google.cloud.dataproc.v1.N" - + "odeInitializationActionB\003\340A\001\022J\n\021encrypti" - + "on_config\030\017 \001(\0132*.google.cloud.dataproc." - + "v1.EncryptionConfigB\003\340A\001\022L\n\022autoscaling_" - + "config\030\022 \001(\0132+.google.cloud.dataproc.v1." - + "AutoscalingConfigB\003\340A\001\022F\n\017security_confi" - + "g\030\020 \001(\0132(.google.cloud.dataproc.v1.Secur" - + "ityConfigB\003\340A\001\022H\n\020lifecycle_config\030\021 \001(\013" - + "2).google.cloud.dataproc.v1.LifecycleCon" - + "figB\003\340A\001\",\n\021AutoscalingConfig\022\027\n\npolicy_" - + "uri\030\001 \001(\tB\003\340A\001\"4\n\020EncryptionConfig\022 \n\023gc" - + "e_pd_kms_key_name\030\001 \001(\tB\003\340A\001\"\237\003\n\020GceClus" - + "terConfig\022\025\n\010zone_uri\030\001 \001(\tB\003\340A\001\022\030\n\013netw" - + "ork_uri\030\002 \001(\tB\003\340A\001\022\033\n\016subnetwork_uri\030\006 \001" - + "(\tB\003\340A\001\022\035\n\020internal_ip_only\030\007 \001(\010B\003\340A\001\022\034" - + "\n\017service_account\030\010 \001(\tB\003\340A\001\022#\n\026service_" - + "account_scopes\030\003 \003(\tB\003\340A\001\022\014\n\004tags\030\004 \003(\t\022" - + "J\n\010metadata\030\005 \003(\01328.google.cloud.datapro" - + "c.v1.GceClusterConfig.MetadataEntry\022P\n\024r" - + "eservation_affinity\030\013 \001(\0132-.google.cloud" - + ".dataproc.v1.ReservationAffinityB\003\340A\001\032/\n" - + "\rMetadataEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001" - + "(\t:\0028\001\"\232\003\n\023InstanceGroupConfig\022\032\n\rnum_in" - + "stances\030\001 \001(\005B\003\340A\001\022\033\n\016instance_names\030\002 \003" - + "(\tB\003\340A\003\022\026\n\timage_uri\030\003 \001(\tB\003\340A\001\022\035\n\020machi" - + "ne_type_uri\030\004 \001(\tB\003\340A\001\022>\n\013disk_config\030\005 " - + "\001(\0132$.google.cloud.dataproc.v1.DiskConfi" - + "gB\003\340A\001\022\033\n\016is_preemptible\030\006 \001(\010B\003\340A\001\022O\n\024m" - + "anaged_group_config\030\007 \001(\0132,.google.cloud" - + ".dataproc.v1.ManagedGroupConfigB\003\340A\003\022F\n\014" - + "accelerators\030\010 \003(\0132+.google.cloud.datapr" - + "oc.v1.AcceleratorConfigB\003\340A\001\022\035\n\020min_cpu_" - + "platform\030\t \001(\tB\003\340A\001\"c\n\022ManagedGroupConfi" - + "g\022#\n\026instance_template_name\030\001 \001(\tB\003\340A\003\022(" - + "\n\033instance_group_manager_name\030\002 \001(\tB\003\340A\003" - + "\"L\n\021AcceleratorConfig\022\034\n\024accelerator_typ" - + "e_uri\030\001 \001(\t\022\031\n\021accelerator_count\030\002 \001(\005\"f" - + "\n\nDiskConfig\022\033\n\016boot_disk_type\030\003 \001(\tB\003\340A" - + "\001\022\036\n\021boot_disk_size_gb\030\001 \001(\005B\003\340A\001\022\033\n\016num" - + "_local_ssds\030\002 \001(\005B\003\340A\001\"s\n\030NodeInitializa" - + "tionAction\022\034\n\017executable_file\030\001 \001(\tB\003\340A\002" - + "\0229\n\021execution_timeout\030\002 \001(\0132\031.google.pro" - + "tobuf.DurationB\003\340A\001\"\204\003\n\rClusterStatus\022A\n" - + "\005state\030\001 \001(\0162-.google.cloud.dataproc.v1." - + "ClusterStatus.StateB\003\340A\003\022\026\n\006detail\030\002 \001(\t" - + "B\006\340A\003\340A\001\0229\n\020state_start_time\030\003 \001(\0132\032.goo" - + "gle.protobuf.TimestampB\003\340A\003\022G\n\010substate\030" - + "\004 \001(\01620.google.cloud.dataproc.v1.Cluster" - + "Status.SubstateB\003\340A\003\"V\n\005State\022\013\n\007UNKNOWN" - + "\020\000\022\014\n\010CREATING\020\001\022\013\n\007RUNNING\020\002\022\t\n\005ERROR\020\003" - + "\022\014\n\010DELETING\020\004\022\014\n\010UPDATING\020\005\"<\n\010Substate" - + "\022\017\n\013UNSPECIFIED\020\000\022\r\n\tUNHEALTHY\020\001\022\020\n\014STAL" - + "E_STATUS\020\002\"S\n\016SecurityConfig\022A\n\017kerberos" - + "_config\030\001 \001(\0132(.google.cloud.dataproc.v1" - + ".KerberosConfig\"\220\004\n\016KerberosConfig\022\034\n\017en" - + "able_kerberos\030\001 \001(\010B\003\340A\001\022(\n\033root_princip" - + "al_password_uri\030\002 \001(\tB\003\340A\002\022\030\n\013kms_key_ur" - + "i\030\003 \001(\tB\003\340A\002\022\031\n\014keystore_uri\030\004 \001(\tB\003\340A\001\022" - + "\033\n\016truststore_uri\030\005 \001(\tB\003\340A\001\022\"\n\025keystore" - + "_password_uri\030\006 \001(\tB\003\340A\001\022\035\n\020key_password" - + "_uri\030\007 \001(\tB\003\340A\001\022$\n\027truststore_password_u" - + "ri\030\010 \001(\tB\003\340A\001\022$\n\027cross_realm_trust_realm" - + "\030\t \001(\tB\003\340A\001\022\"\n\025cross_realm_trust_kdc\030\n \001" - + "(\tB\003\340A\001\022+\n\036cross_realm_trust_admin_serve" - + "r\030\013 \001(\tB\003\340A\001\0222\n%cross_realm_trust_shared" - + "_password_uri\030\014 \001(\tB\003\340A\001\022\033\n\016kdc_db_key_u" - + "ri\030\r \001(\tB\003\340A\001\022\037\n\022tgt_lifetime_hours\030\016 \001(" - + "\005B\003\340A\001\022\022\n\005realm\030\017 \001(\tB\003\340A\001\"\371\001\n\016SoftwareC" - + "onfig\022\032\n\rimage_version\030\001 \001(\tB\003\340A\001\022Q\n\npro" - + "perties\030\002 \003(\01328.google.cloud.dataproc.v1" - + ".SoftwareConfig.PropertiesEntryB\003\340A\001\022E\n\023" - + "optional_components\030\003 \003(\0162#.google.cloud" - + ".dataproc.v1.ComponentB\003\340A\001\0321\n\017Propertie" - + "sEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\203" - + "\002\n\017LifecycleConfig\0227\n\017idle_delete_ttl\030\001 " - + "\001(\0132\031.google.protobuf.DurationB\003\340A\001\022;\n\020a" - + "uto_delete_time\030\002 \001(\0132\032.google.protobuf." - + "TimestampB\003\340A\001H\000\0229\n\017auto_delete_ttl\030\003 \001(" - + "\0132\031.google.protobuf.DurationB\003\340A\001H\000\0228\n\017i" - + "dle_start_time\030\004 \001(\0132\032.google.protobuf.T" - + "imestampB\003\340A\003B\005\n\003ttl\"\232\002\n\016ClusterMetrics\022" - + "O\n\014hdfs_metrics\030\001 \003(\01329.google.cloud.dat" - + "aproc.v1.ClusterMetrics.HdfsMetricsEntry" - + "\022O\n\014yarn_metrics\030\002 \003(\01329.google.cloud.da" - + "taproc.v1.ClusterMetrics.YarnMetricsEntr" - + "y\0322\n\020HdfsMetricsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005va" - + "lue\030\002 \001(\003:\0028\001\0322\n\020YarnMetricsEntry\022\013\n\003key" - + "\030\001 \001(\t\022\r\n\005value\030\002 \001(\003:\0028\001\"\226\001\n\024CreateClus" + + "\340A\003\022>\n\007metrics\030\t \001(\0132(.google.cloud.data" + + "proc.v1.ClusterMetricsB\003\340A\003\032-\n\013LabelsEnt" + + "ry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\260\006\n\rC" + + "lusterConfig\022\032\n\rconfig_bucket\030\001 \001(\tB\003\340A\001" + + "\022K\n\022gce_cluster_config\030\010 \001(\0132*.google.cl" + + "oud.dataproc.v1.GceClusterConfigB\003\340A\001\022I\n" + + "\rmaster_config\030\t \001(\0132-.google.cloud.data" + + "proc.v1.InstanceGroupConfigB\003\340A\001\022I\n\rwork" + + "er_config\030\n \001(\0132-.google.cloud.dataproc." + + "v1.InstanceGroupConfigB\003\340A\001\022S\n\027secondary" + + "_worker_config\030\014 \001(\0132-.google.cloud.data" + + "proc.v1.InstanceGroupConfigB\003\340A\001\022F\n\017soft" + + "ware_config\030\r \001(\0132(.google.cloud.datapro" + + "c.v1.SoftwareConfigB\003\340A\001\022W\n\026initializati" + + "on_actions\030\013 \003(\01322.google.cloud.dataproc" + + ".v1.NodeInitializationActionB\003\340A\001\022J\n\021enc" + + "ryption_config\030\017 \001(\0132*.google.cloud.data" + + "proc.v1.EncryptionConfigB\003\340A\001\022L\n\022autosca" + + "ling_config\030\022 \001(\0132+.google.cloud.datapro" + + "c.v1.AutoscalingConfigB\003\340A\001\022F\n\017security_" + + "config\030\020 \001(\0132(.google.cloud.dataproc.v1." + + "SecurityConfigB\003\340A\001\022H\n\020lifecycle_config\030" + + "\021 \001(\0132).google.cloud.dataproc.v1.Lifecyc" + + "leConfigB\003\340A\001\",\n\021AutoscalingConfig\022\027\n\npo" + + "licy_uri\030\001 \001(\tB\003\340A\001\"4\n\020EncryptionConfig\022" + + " \n\023gce_pd_kms_key_name\030\001 \001(\tB\003\340A\001\"\237\003\n\020Gc" + + "eClusterConfig\022\025\n\010zone_uri\030\001 \001(\tB\003\340A\001\022\030\n" + + "\013network_uri\030\002 \001(\tB\003\340A\001\022\033\n\016subnetwork_ur" + + "i\030\006 \001(\tB\003\340A\001\022\035\n\020internal_ip_only\030\007 \001(\010B\003" + + "\340A\001\022\034\n\017service_account\030\010 \001(\tB\003\340A\001\022#\n\026ser" + + "vice_account_scopes\030\003 \003(\tB\003\340A\001\022\014\n\004tags\030\004" + + " \003(\t\022J\n\010metadata\030\005 \003(\01328.google.cloud.da" + + "taproc.v1.GceClusterConfig.MetadataEntry" + + "\022P\n\024reservation_affinity\030\013 \001(\0132-.google." + + "cloud.dataproc.v1.ReservationAffinityB\003\340" + + "A\001\032/\n\rMetadataEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005valu" + + "e\030\002 \001(\t:\0028\001\"\232\003\n\023InstanceGroupConfig\022\032\n\rn" + + "um_instances\030\001 \001(\005B\003\340A\001\022\033\n\016instance_name" + + "s\030\002 \003(\tB\003\340A\003\022\026\n\timage_uri\030\003 \001(\tB\003\340A\001\022\035\n\020" + + "machine_type_uri\030\004 \001(\tB\003\340A\001\022>\n\013disk_conf" + + "ig\030\005 \001(\0132$.google.cloud.dataproc.v1.Disk" + + "ConfigB\003\340A\001\022\033\n\016is_preemptible\030\006 \001(\010B\003\340A\003" + + "\022O\n\024managed_group_config\030\007 \001(\0132,.google." + + "cloud.dataproc.v1.ManagedGroupConfigB\003\340A" + + "\003\022F\n\014accelerators\030\010 \003(\0132+.google.cloud.d" + + "ataproc.v1.AcceleratorConfigB\003\340A\001\022\035\n\020min" + + "_cpu_platform\030\t \001(\tB\003\340A\001\"c\n\022ManagedGroup" + + "Config\022#\n\026instance_template_name\030\001 \001(\tB\003" + + "\340A\003\022(\n\033instance_group_manager_name\030\002 \001(\t" + + "B\003\340A\003\"L\n\021AcceleratorConfig\022\034\n\024accelerato" + + "r_type_uri\030\001 \001(\t\022\031\n\021accelerator_count\030\002 " + + "\001(\005\"f\n\nDiskConfig\022\033\n\016boot_disk_type\030\003 \001(" + + "\tB\003\340A\001\022\036\n\021boot_disk_size_gb\030\001 \001(\005B\003\340A\001\022\033" + + "\n\016num_local_ssds\030\002 \001(\005B\003\340A\001\"s\n\030NodeIniti" + + "alizationAction\022\034\n\017executable_file\030\001 \001(\t" + + "B\003\340A\002\0229\n\021execution_timeout\030\002 \001(\0132\031.googl" + + "e.protobuf.DurationB\003\340A\001\"\204\003\n\rClusterStat" + + "us\022A\n\005state\030\001 \001(\0162-.google.cloud.datapro" + + "c.v1.ClusterStatus.StateB\003\340A\003\022\026\n\006detail\030" + + "\002 \001(\tB\006\340A\003\340A\001\0229\n\020state_start_time\030\003 \001(\0132" + + "\032.google.protobuf.TimestampB\003\340A\003\022G\n\010subs" + + "tate\030\004 \001(\01620.google.cloud.dataproc.v1.Cl" + + "usterStatus.SubstateB\003\340A\003\"V\n\005State\022\013\n\007UN" + + "KNOWN\020\000\022\014\n\010CREATING\020\001\022\013\n\007RUNNING\020\002\022\t\n\005ER" + + "ROR\020\003\022\014\n\010DELETING\020\004\022\014\n\010UPDATING\020\005\"<\n\010Sub" + + "state\022\017\n\013UNSPECIFIED\020\000\022\r\n\tUNHEALTHY\020\001\022\020\n" + + "\014STALE_STATUS\020\002\"S\n\016SecurityConfig\022A\n\017ker" + + "beros_config\030\001 \001(\0132(.google.cloud.datapr" + + "oc.v1.KerberosConfig\"\220\004\n\016KerberosConfig\022" + + "\034\n\017enable_kerberos\030\001 \001(\010B\003\340A\001\022(\n\033root_pr" + + "incipal_password_uri\030\002 \001(\tB\003\340A\002\022\030\n\013kms_k" + + "ey_uri\030\003 \001(\tB\003\340A\002\022\031\n\014keystore_uri\030\004 \001(\tB" + + "\003\340A\001\022\033\n\016truststore_uri\030\005 \001(\tB\003\340A\001\022\"\n\025key" + + "store_password_uri\030\006 \001(\tB\003\340A\001\022\035\n\020key_pas" + + "sword_uri\030\007 \001(\tB\003\340A\001\022$\n\027truststore_passw" + + "ord_uri\030\010 \001(\tB\003\340A\001\022$\n\027cross_realm_trust_" + + "realm\030\t \001(\tB\003\340A\001\022\"\n\025cross_realm_trust_kd" + + "c\030\n \001(\tB\003\340A\001\022+\n\036cross_realm_trust_admin_" + + "server\030\013 \001(\tB\003\340A\001\0222\n%cross_realm_trust_s" + + "hared_password_uri\030\014 \001(\tB\003\340A\001\022\033\n\016kdc_db_" + + "key_uri\030\r \001(\tB\003\340A\001\022\037\n\022tgt_lifetime_hours" + + "\030\016 \001(\005B\003\340A\001\022\022\n\005realm\030\017 \001(\tB\003\340A\001\"\371\001\n\016Soft" + + "wareConfig\022\032\n\rimage_version\030\001 \001(\tB\003\340A\001\022Q" + + "\n\nproperties\030\002 \003(\01328.google.cloud.datapr" + + "oc.v1.SoftwareConfig.PropertiesEntryB\003\340A" + + "\001\022E\n\023optional_components\030\003 \003(\0162#.google." + + "cloud.dataproc.v1.ComponentB\003\340A\001\0321\n\017Prop" + + "ertiesEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:" + + "\0028\001\"\203\002\n\017LifecycleConfig\0227\n\017idle_delete_t" + + "tl\030\001 \001(\0132\031.google.protobuf.DurationB\003\340A\001" + + "\022;\n\020auto_delete_time\030\002 \001(\0132\032.google.prot" + + "obuf.TimestampB\003\340A\001H\000\0229\n\017auto_delete_ttl" + + "\030\003 \001(\0132\031.google.protobuf.DurationB\003\340A\001H\000" + + "\0228\n\017idle_start_time\030\004 \001(\0132\032.google.proto" + + "buf.TimestampB\003\340A\003B\005\n\003ttl\"\232\002\n\016ClusterMet" + + "rics\022O\n\014hdfs_metrics\030\001 \003(\01329.google.clou" + + "d.dataproc.v1.ClusterMetrics.HdfsMetrics" + + "Entry\022O\n\014yarn_metrics\030\002 \003(\01329.google.clo" + + "ud.dataproc.v1.ClusterMetrics.YarnMetric" + + "sEntry\0322\n\020HdfsMetricsEntry\022\013\n\003key\030\001 \001(\t\022" + + "\r\n\005value\030\002 \001(\003:\0028\001\0322\n\020YarnMetricsEntry\022\013" + + "\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\003:\0028\001\"\226\001\n\024Creat" + + "eClusterRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002" + + "\022\023\n\006region\030\003 \001(\tB\003\340A\002\0227\n\007cluster\030\002 \001(\0132!" + + ".google.cloud.dataproc.v1.ClusterB\003\340A\002\022\027" + + "\n\nrequest_id\030\004 \001(\tB\003\340A\001\"\256\002\n\024UpdateCluste" + + "rRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006reg" + + "ion\030\005 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A" + + "\002\0227\n\007cluster\030\003 \001(\0132!.google.cloud.datapr" + + "oc.v1.ClusterB\003\340A\002\022E\n\035graceful_decommiss" + + "ion_timeout\030\006 \001(\0132\031.google.protobuf.Dura" + + "tionB\003\340A\001\0224\n\013update_mask\030\004 \001(\0132\032.google." + + "protobuf.FieldMaskB\003\340A\002\022\027\n\nrequest_id\030\007 " + + "\001(\tB\003\340A\001\"\223\001\n\024DeleteClusterRequest\022\027\n\npro" + + "ject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022" + + "\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\022\031\n\014cluster_uu" + + "id\030\004 \001(\tB\003\340A\001\022\027\n\nrequest_id\030\005 \001(\tB\003\340A\001\"\\" + + "\n\021GetClusterRequest\022\027\n\nproject_id\030\001 \001(\tB" + + "\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_nam" + + "e\030\002 \001(\tB\003\340A\002\"\211\001\n\023ListClustersRequest\022\027\n\n" + + "project_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\004 \001(\tB\003\340" + + "A\002\022\023\n\006filter\030\005 \001(\tB\003\340A\001\022\026\n\tpage_size\030\002 \001" + + "(\005B\003\340A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001\"n\n\024List" + + "ClustersResponse\0228\n\010clusters\030\001 \003(\0132!.goo" + + "gle.cloud.dataproc.v1.ClusterB\003\340A\003\022\034\n\017ne" + + "xt_page_token\030\002 \001(\tB\003\340A\003\"a\n\026DiagnoseClus" + "terRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006r" - + "egion\030\003 \001(\tB\003\340A\002\0227\n\007cluster\030\002 \001(\0132!.goog" - + "le.cloud.dataproc.v1.ClusterB\003\340A\002\022\027\n\nreq" - + "uest_id\030\004 \001(\tB\003\340A\001\"\256\002\n\024UpdateClusterRequ" - + "est\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\005" - + " \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\0227\n\007" - + "cluster\030\003 \001(\0132!.google.cloud.dataproc.v1" - + ".ClusterB\003\340A\002\022E\n\035graceful_decommission_t" - + "imeout\030\006 \001(\0132\031.google.protobuf.DurationB" - + "\003\340A\001\0224\n\013update_mask\030\004 \001(\0132\032.google.proto" - + "buf.FieldMaskB\003\340A\002\022\027\n\nrequest_id\030\007 \001(\tB\003" - + "\340A\001\"\223\001\n\024DeleteClusterRequest\022\027\n\nproject_" - + "id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cl" - + "uster_name\030\002 \001(\tB\003\340A\002\022\031\n\014cluster_uuid\030\004 " - + "\001(\tB\003\340A\001\022\027\n\nrequest_id\030\005 \001(\tB\003\340A\001\"\\\n\021Get" - + "ClusterRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022" - + "\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001" - + "(\tB\003\340A\002\"\211\001\n\023ListClustersRequest\022\027\n\nproje" - + "ct_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\004 \001(\tB\003\340A\002\022\023\n" - + "\006filter\030\005 \001(\tB\003\340A\001\022\026\n\tpage_size\030\002 \001(\005B\003\340" - + "A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001\"n\n\024ListClust" - + "ersResponse\0228\n\010clusters\030\001 \003(\0132!.google.c" - + "loud.dataproc.v1.ClusterB\003\340A\003\022\034\n\017next_pa" - + "ge_token\030\002 \001(\tB\003\340A\003\"a\n\026DiagnoseClusterRe" - + "quest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region" - + "\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\"1" - + "\n\026DiagnoseClusterResults\022\027\n\noutput_uri\030\001" - + " \001(\tB\003\340A\003\"\370\001\n\023ReservationAffinity\022Y\n\030con" - + "sume_reservation_type\030\001 \001(\01622.google.clo" - + "ud.dataproc.v1.ReservationAffinity.TypeB" - + "\003\340A\001\022\020\n\003key\030\002 \001(\tB\003\340A\001\022\023\n\006values\030\003 \003(\tB\003" - + "\340A\001\"_\n\004Type\022\024\n\020TYPE_UNSPECIFIED\020\000\022\022\n\016NO_" - + "RESERVATION\020\001\022\023\n\017ANY_RESERVATION\020\002\022\030\n\024SP" - + "ECIFIC_RESERVATION\020\0032\343\014\n\021ClusterControll" - + "er\022\200\002\n\rCreateCluster\022..google.cloud.data" - + "proc.v1.CreateClusterRequest\032\035.google.lo" - + "ngrunning.Operation\"\237\001\202\323\344\223\002>\"3/v1/projec" - + "ts/{project_id}/regions/{region}/cluster" - + "s:\007cluster\332A\031project_id,region,cluster\312A" - + "<\n\007Cluster\0221google.cloud.dataproc.v1.Clu" - + "sterOperationMetadata\022\250\002\n\rUpdateCluster\022" - + "..google.cloud.dataproc.v1.UpdateCluster" - + "Request\032\035.google.longrunning.Operation\"\307" - + "\001\202\323\344\223\002M2B/v1/projects/{project_id}/regio" - + "ns/{region}/clusters/{cluster_name}:\007clu" - + "ster\332A2project_id,region,cluster_name,cl" - + "uster,update_mask\312A<\n\007Cluster\0221google.cl" - + "oud.dataproc.v1.ClusterOperationMetadata" - + "\022\231\002\n\rDeleteCluster\022..google.cloud.datapr" - + "oc.v1.DeleteClusterRequest\032\035.google.long" - + "running.Operation\"\270\001\202\323\344\223\002D*B/v1/projects" - + "/{project_id}/regions/{region}/clusters/" - + "{cluster_name}\332A\036project_id,region,clust" - + "er_name\312AJ\n\025google.protobuf.Empty\0221googl" - + "e.cloud.dataproc.v1.ClusterOperationMeta" - + "data\022\311\001\n\nGetCluster\022+.google.cloud.datap" - + "roc.v1.GetClusterRequest\032!.google.cloud." - + "dataproc.v1.Cluster\"k\202\323\344\223\002D\022B/v1/project" - + "s/{project_id}/regions/{region}/clusters" - + "/{cluster_name}\332A\036project_id,region,clus" - + "ter_name\022\331\001\n\014ListClusters\022-.google.cloud" - + ".dataproc.v1.ListClustersRequest\032..googl" - + "e.cloud.dataproc.v1.ListClustersResponse" - + "\"j\202\323\344\223\0025\0223/v1/projects/{project_id}/regi" - + "ons/{region}/clusters\332A\021project_id,regio" - + "n\332A\030project_id,region,filter\022\216\002\n\017Diagnos" - + "eCluster\0220.google.cloud.dataproc.v1.Diag" - + "noseClusterRequest\032\035.google.longrunning." - + "Operation\"\251\001\202\323\344\223\002P\"K/v1/projects/{projec" - + "t_id}/regions/{region}/clusters/{cluster" - + "_name}:diagnose:\001*\332A\036project_id,region,c" - + "luster_name\312A/\n\025google.protobuf.Empty\022\026D" - + "iagnoseClusterResults\032K\312A\027dataproc.googl" - + "eapis.com\322A.https://www.googleapis.com/a" - + "uth/cloud-platformBq\n\034com.google.cloud.d" - + "ataproc.v1B\rClustersProtoP\001Z@google.gola" - + "ng.org/genproto/googleapis/cloud/datapro" - + "c/v1;dataprocb\006proto3" + + "egion\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003" + + "\340A\002\"1\n\026DiagnoseClusterResults\022\027\n\noutput_" + + "uri\030\001 \001(\tB\003\340A\003\"\370\001\n\023ReservationAffinity\022Y" + + "\n\030consume_reservation_type\030\001 \001(\01622.googl" + + "e.cloud.dataproc.v1.ReservationAffinity." + + "TypeB\003\340A\001\022\020\n\003key\030\002 \001(\tB\003\340A\001\022\023\n\006values\030\003 " + + "\003(\tB\003\340A\001\"_\n\004Type\022\024\n\020TYPE_UNSPECIFIED\020\000\022\022" + + "\n\016NO_RESERVATION\020\001\022\023\n\017ANY_RESERVATION\020\002\022" + + "\030\n\024SPECIFIC_RESERVATION\020\0032\343\014\n\021ClusterCon" + + "troller\022\200\002\n\rCreateCluster\022..google.cloud" + + ".dataproc.v1.CreateClusterRequest\032\035.goog" + + "le.longrunning.Operation\"\237\001\202\323\344\223\002>\"3/v1/p" + + "rojects/{project_id}/regions/{region}/cl" + + "usters:\007cluster\332A\031project_id,region,clus" + + "ter\312A<\n\007Cluster\0221google.cloud.dataproc.v" + + "1.ClusterOperationMetadata\022\250\002\n\rUpdateClu" + + "ster\022..google.cloud.dataproc.v1.UpdateCl" + + "usterRequest\032\035.google.longrunning.Operat" + + "ion\"\307\001\202\323\344\223\002M2B/v1/projects/{project_id}/" + + "regions/{region}/clusters/{cluster_name}" + + ":\007cluster\332A2project_id,region,cluster_na" + + "me,cluster,update_mask\312A<\n\007Cluster\0221goog" + + "le.cloud.dataproc.v1.ClusterOperationMet" + + "adata\022\231\002\n\rDeleteCluster\022..google.cloud.d" + + "ataproc.v1.DeleteClusterRequest\032\035.google" + + ".longrunning.Operation\"\270\001\202\323\344\223\002D*B/v1/pro" + + "jects/{project_id}/regions/{region}/clus" + + "ters/{cluster_name}\332A\036project_id,region," + + "cluster_name\312AJ\n\025google.protobuf.Empty\0221" + + "google.cloud.dataproc.v1.ClusterOperatio" + + "nMetadata\022\311\001\n\nGetCluster\022+.google.cloud." + + "dataproc.v1.GetClusterRequest\032!.google.c" + + "loud.dataproc.v1.Cluster\"k\202\323\344\223\002D\022B/v1/pr" + + "ojects/{project_id}/regions/{region}/clu" + + "sters/{cluster_name}\332A\036project_id,region" + + ",cluster_name\022\331\001\n\014ListClusters\022-.google." + + "cloud.dataproc.v1.ListClustersRequest\032.." + + "google.cloud.dataproc.v1.ListClustersRes" + + "ponse\"j\202\323\344\223\0025\0223/v1/projects/{project_id}" + + "/regions/{region}/clusters\332A\021project_id," + + "region\332A\030project_id,region,filter\022\216\002\n\017Di" + + "agnoseCluster\0220.google.cloud.dataproc.v1" + + ".DiagnoseClusterRequest\032\035.google.longrun" + + "ning.Operation\"\251\001\202\323\344\223\002P\"K/v1/projects/{p" + + "roject_id}/regions/{region}/clusters/{cl" + + "uster_name}:diagnose:\001*\332A\036project_id,reg" + + "ion,cluster_name\312A/\n\025google.protobuf.Emp" + + "ty\022\026DiagnoseClusterResults\032K\312A\027dataproc." + + "googleapis.com\322A.https://www.googleapis." + + "com/auth/cloud-platformBq\n\034com.google.cl" + + "oud.dataproc.v1B\rClustersProtoP\001Z@google" + + ".golang.org/genproto/googleapis/cloud/da" + + "taproc/v1;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Component.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Component.java index dcf1e190..09e877dc 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Component.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Component.java @@ -68,6 +68,16 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * JUPYTER = 1; */ JUPYTER(1), + /** + * + * + *
+   * The Presto query engine.
+   * 
+ * + * PRESTO = 6; + */ + PRESTO(6), /** * * @@ -78,6 +88,16 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * ZEPPELIN = 4; */ ZEPPELIN(4), + /** + * + * + *
+   * The Zookeeper service.
+   * 
+ * + * ZOOKEEPER = 8; + */ + ZOOKEEPER(8), UNRECOGNIZED(-1), ; @@ -121,6 +141,16 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * JUPYTER = 1; */ public static final int JUPYTER_VALUE = 1; + /** + * + * + *
+   * The Presto query engine.
+   * 
+ * + * PRESTO = 6; + */ + public static final int PRESTO_VALUE = 6; /** * * @@ -131,6 +161,16 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * ZEPPELIN = 4; */ public static final int ZEPPELIN_VALUE = 4; + /** + * + * + *
+   * The Zookeeper service.
+   * 
+ * + * ZOOKEEPER = 8; + */ + public static final int ZOOKEEPER_VALUE = 8; public final int getNumber() { if (this == UNRECOGNIZED) { @@ -164,8 +204,12 @@ public static Component forNumber(int value) { return HIVE_WEBHCAT; case 1: return JUPYTER; + case 6: + return PRESTO; case 4: return ZEPPELIN; + case 8: + return ZOOKEEPER; default: return null; } diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequest.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequest.java index aa7fed2b..3d7fcd7c 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequest.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequest.java @@ -199,10 +199,12 @@ public com.google.protobuf.ByteString getParentBytes() { * * *
-   * The autoscaling policy to create.
+   * Required. The autoscaling policy to create.
    * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * * * @return Whether the policy field is set. */ @@ -213,10 +215,12 @@ public boolean hasPolicy() { * * *
-   * The autoscaling policy to create.
+   * Required. The autoscaling policy to create.
    * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * * * @return The policy. */ @@ -229,10 +233,12 @@ public com.google.cloud.dataproc.v1.AutoscalingPolicy getPolicy() { * * *
-   * The autoscaling policy to create.
+   * Required. The autoscaling policy to create.
    * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ public com.google.cloud.dataproc.v1.AutoscalingPolicyOrBuilder getPolicyOrBuilder() { return getPolicy(); @@ -746,10 +752,12 @@ public Builder setParentBytes(com.google.protobuf.ByteString value) { * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * * * @return Whether the policy field is set. */ @@ -760,10 +768,12 @@ public boolean hasPolicy() { * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * * * @return The policy. */ @@ -780,10 +790,12 @@ public com.google.cloud.dataproc.v1.AutoscalingPolicy getPolicy() { * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ public Builder setPolicy(com.google.cloud.dataproc.v1.AutoscalingPolicy value) { if (policyBuilder_ == null) { @@ -802,10 +814,12 @@ public Builder setPolicy(com.google.cloud.dataproc.v1.AutoscalingPolicy value) { * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ public Builder setPolicy( com.google.cloud.dataproc.v1.AutoscalingPolicy.Builder builderForValue) { @@ -822,10 +836,12 @@ public Builder setPolicy( * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ public Builder mergePolicy(com.google.cloud.dataproc.v1.AutoscalingPolicy value) { if (policyBuilder_ == null) { @@ -848,10 +864,12 @@ public Builder mergePolicy(com.google.cloud.dataproc.v1.AutoscalingPolicy value) * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ public Builder clearPolicy() { if (policyBuilder_ == null) { @@ -868,10 +886,12 @@ public Builder clearPolicy() { * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ public com.google.cloud.dataproc.v1.AutoscalingPolicy.Builder getPolicyBuilder() { @@ -882,10 +902,12 @@ public com.google.cloud.dataproc.v1.AutoscalingPolicy.Builder getPolicyBuilder() * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ public com.google.cloud.dataproc.v1.AutoscalingPolicyOrBuilder getPolicyOrBuilder() { if (policyBuilder_ != null) { @@ -900,10 +922,12 @@ public com.google.cloud.dataproc.v1.AutoscalingPolicyOrBuilder getPolicyOrBuilde * * *
-     * The autoscaling policy to create.
+     * Required. The autoscaling policy to create.
      * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1.AutoscalingPolicy, diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequestOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequestOrBuilder.java index 74b43346..a7571aef 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequestOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/CreateAutoscalingPolicyRequestOrBuilder.java @@ -70,10 +70,12 @@ public interface CreateAutoscalingPolicyRequestOrBuilder * * *
-   * The autoscaling policy to create.
+   * Required. The autoscaling policy to create.
    * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * * * @return Whether the policy field is set. */ @@ -82,10 +84,12 @@ public interface CreateAutoscalingPolicyRequestOrBuilder * * *
-   * The autoscaling policy to create.
+   * Required. The autoscaling policy to create.
    * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * * * @return The policy. */ @@ -94,10 +98,12 @@ public interface CreateAutoscalingPolicyRequestOrBuilder * * *
-   * The autoscaling policy to create.
+   * Required. The autoscaling policy to create.
    * 
* - * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2; + * + * .google.cloud.dataproc.v1.AutoscalingPolicy policy = 2 [(.google.api.field_behavior) = REQUIRED]; + * */ com.google.cloud.dataproc.v1.AutoscalingPolicyOrBuilder getPolicyOrBuilder(); } diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java index c8ece660..96dcb263 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfig.java @@ -285,7 +285,8 @@ public com.google.protobuf.ByteString getZoneUriBytes() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -315,7 +316,8 @@ public java.lang.String getNetworkUri() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -424,14 +426,14 @@ public boolean getInternalIpOnly() { * *
    * Optional. The [Dataproc service
-   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
    * (also see [VM Data Plane
-   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
    * used by Dataproc cluster VM instances to access Google Cloud Platform
    * services.
    * If not specified, the
    * [Compute Engine default service
-   * account](/compute/docs/access/service-accounts#default_service_account)
+   * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
    * is used.
    * 
* @@ -455,14 +457,14 @@ public java.lang.String getServiceAccount() { * *
    * Optional. The [Dataproc service
-   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
    * (also see [VM Data Plane
-   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
    * used by Dataproc cluster VM instances to access Google Cloud Platform
    * services.
    * If not specified, the
    * [Compute Engine default service
-   * account](/compute/docs/access/service-accounts#default_service_account)
+   * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
    * is used.
    * 
* @@ -593,8 +595,8 @@ public com.google.protobuf.ByteString getServiceAccountScopesBytes(int index) { * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -608,8 +610,8 @@ public com.google.protobuf.ProtocolStringList getTagsList() { * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -623,8 +625,8 @@ public int getTagsCount() { * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -639,8 +641,8 @@ public java.lang.String getTags(int index) { * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -1481,7 +1483,8 @@ public Builder setZoneUriBytes(com.google.protobuf.ByteString value) { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1511,7 +1514,8 @@ public java.lang.String getNetworkUri() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1541,7 +1545,8 @@ public com.google.protobuf.ByteString getNetworkUriBytes() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1570,7 +1575,8 @@ public Builder setNetworkUri(java.lang.String value) { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1595,7 +1601,8 @@ public Builder clearNetworkUri() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1821,14 +1828,14 @@ public Builder clearInternalIpOnly() { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -1852,14 +1859,14 @@ public java.lang.String getServiceAccount() { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -1883,14 +1890,14 @@ public com.google.protobuf.ByteString getServiceAccountBytes() { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -1913,14 +1920,14 @@ public Builder setServiceAccount(java.lang.String value) { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -1939,14 +1946,14 @@ public Builder clearServiceAccount() { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -2246,8 +2253,8 @@ private void ensureTagsIsMutable() { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2261,8 +2268,8 @@ public com.google.protobuf.ProtocolStringList getTagsList() { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2276,8 +2283,8 @@ public int getTagsCount() { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2292,8 +2299,8 @@ public java.lang.String getTags(int index) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2308,8 +2315,8 @@ public com.google.protobuf.ByteString getTagsBytes(int index) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2331,8 +2338,8 @@ public Builder setTags(int index, java.lang.String value) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2353,8 +2360,8 @@ public Builder addTags(java.lang.String value) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2372,8 +2379,8 @@ public Builder addAllTags(java.lang.Iterable values) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2390,8 +2397,8 @@ public Builder clearTags() { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java index 2dfec7d0..0580305d 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/GceClusterConfigOrBuilder.java @@ -72,7 +72,8 @@ public interface GceClusterConfigOrBuilder * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -92,7 +93,8 @@ public interface GceClusterConfigOrBuilder * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -163,14 +165,14 @@ public interface GceClusterConfigOrBuilder * *
    * Optional. The [Dataproc service
-   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
    * (also see [VM Data Plane
-   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
    * used by Dataproc cluster VM instances to access Google Cloud Platform
    * services.
    * If not specified, the
    * [Compute Engine default service
-   * account](/compute/docs/access/service-accounts#default_service_account)
+   * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
    * is used.
    * 
* @@ -184,14 +186,14 @@ public interface GceClusterConfigOrBuilder * *
    * Optional. The [Dataproc service
-   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
    * (also see [VM Data Plane
-   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
    * used by Dataproc cluster VM instances to access Google Cloud Platform
    * services.
    * If not specified, the
    * [Compute Engine default service
-   * account](/compute/docs/access/service-accounts#default_service_account)
+   * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
    * is used.
    * 
* @@ -300,8 +302,8 @@ public interface GceClusterConfigOrBuilder * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -313,8 +315,8 @@ public interface GceClusterConfigOrBuilder * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -326,8 +328,8 @@ public interface GceClusterConfigOrBuilder * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -340,8 +342,8 @@ public interface GceClusterConfigOrBuilder * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java index 1d3ab0c1..9e64c69e 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfig.java @@ -290,9 +290,18 @@ public com.google.protobuf.ByteString getInstanceNamesBytes(int index) { * * *
-   * Optional. The Compute Engine image resource used for cluster
-   * instances. It can be specified or may be inferred from
-   * `SoftwareConfig.image_version`.
+   * Optional. The Compute Engine image resource used for cluster instances.
+   * The URI can represent an image or image family.
+   * Image examples:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+   * * `projects/[project_id]/global/images/[image-id]`
+   * * `image-id`
+   * Image family examples. Dataproc will use the most recent
+   * image from the family:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * If the URI is unspecified, it will be inferred from
+   * `SoftwareConfig.image_version` or the system default.
    * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -314,9 +323,18 @@ public java.lang.String getImageUri() { * * *
-   * Optional. The Compute Engine image resource used for cluster
-   * instances. It can be specified or may be inferred from
-   * `SoftwareConfig.image_version`.
+   * Optional. The Compute Engine image resource used for cluster instances.
+   * The URI can represent an image or image family.
+   * Image examples:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+   * * `projects/[project_id]/global/images/[image-id]`
+   * * `image-id`
+   * Image family examples. Dataproc will use the most recent
+   * image from the family:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * If the URI is unspecified, it will be inferred from
+   * `SoftwareConfig.image_version` or the system default.
    * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -348,7 +366,7 @@ public com.google.protobuf.ByteString getImageUriBytes() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -379,7 +397,7 @@ public java.lang.String getMachineTypeUri() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -457,11 +475,11 @@ public com.google.cloud.dataproc.v1.DiskConfigOrBuilder getDiskConfigOrBuilder() * * *
-   * Optional. Specifies that this instance group contains preemptible
+   * Output only. Specifies that this instance group contains preemptible
    * instances.
    * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @return The isPreemptible. */ @@ -613,8 +631,8 @@ public com.google.cloud.dataproc.v1.AcceleratorConfigOrBuilder getAcceleratorsOr * *
    * Optional. Specifies the minimum cpu platform for the Instance Group.
-   * See [Dataproc&rarr;Minimum CPU Platform]
-   * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+   * See [Dataproc -> Minimum CPU
+   * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; @@ -637,8 +655,8 @@ public java.lang.String getMinCpuPlatform() { * *
    * Optional. Specifies the minimum cpu platform for the Instance Group.
-   * See [Dataproc&rarr;Minimum CPU Platform]
-   * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+   * See [Dataproc -> Minimum CPU
+   * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; @@ -1420,9 +1438,18 @@ public Builder addInstanceNamesBytes(com.google.protobuf.ByteString value) { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1444,9 +1471,18 @@ public java.lang.String getImageUri() { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1468,9 +1504,18 @@ public com.google.protobuf.ByteString getImageUriBytes() { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1491,9 +1536,18 @@ public Builder setImageUri(java.lang.String value) { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1510,9 +1564,18 @@ public Builder clearImageUri() { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1543,7 +1606,7 @@ public Builder setImageUriBytes(com.google.protobuf.ByteString value) { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1574,7 +1637,7 @@ public java.lang.String getMachineTypeUri() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1605,7 +1668,7 @@ public com.google.protobuf.ByteString getMachineTypeUriBytes() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1635,7 +1698,7 @@ public Builder setMachineTypeUri(java.lang.String value) { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1661,7 +1724,7 @@ public Builder clearMachineTypeUri() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1890,11 +1953,11 @@ public com.google.cloud.dataproc.v1.DiskConfigOrBuilder getDiskConfigOrBuilder() * * *
-     * Optional. Specifies that this instance group contains preemptible
+     * Output only. Specifies that this instance group contains preemptible
      * instances.
      * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @return The isPreemptible. */ @@ -1905,11 +1968,11 @@ public boolean getIsPreemptible() { * * *
-     * Optional. Specifies that this instance group contains preemptible
+     * Output only. Specifies that this instance group contains preemptible
      * instances.
      * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @param value The isPreemptible to set. * @return This builder for chaining. @@ -1924,11 +1987,11 @@ public Builder setIsPreemptible(boolean value) { * * *
-     * Optional. Specifies that this instance group contains preemptible
+     * Output only. Specifies that this instance group contains preemptible
      * instances.
      * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @return This builder for chaining. */ @@ -2578,8 +2641,8 @@ public com.google.cloud.dataproc.v1.AcceleratorConfig.Builder addAcceleratorsBui * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; @@ -2602,8 +2665,8 @@ public java.lang.String getMinCpuPlatform() { * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; @@ -2626,8 +2689,8 @@ public com.google.protobuf.ByteString getMinCpuPlatformBytes() { * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; @@ -2649,8 +2712,8 @@ public Builder setMinCpuPlatform(java.lang.String value) { * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; @@ -2668,8 +2731,8 @@ public Builder clearMinCpuPlatform() { * *
      * Optional. Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java index 905c3a07..dda45e3f 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/InstanceGroupConfigOrBuilder.java @@ -96,9 +96,18 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Optional. The Compute Engine image resource used for cluster
-   * instances. It can be specified or may be inferred from
-   * `SoftwareConfig.image_version`.
+   * Optional. The Compute Engine image resource used for cluster instances.
+   * The URI can represent an image or image family.
+   * Image examples:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+   * * `projects/[project_id]/global/images/[image-id]`
+   * * `image-id`
+   * Image family examples. Dataproc will use the most recent
+   * image from the family:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * If the URI is unspecified, it will be inferred from
+   * `SoftwareConfig.image_version` or the system default.
    * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -110,9 +119,18 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Optional. The Compute Engine image resource used for cluster
-   * instances. It can be specified or may be inferred from
-   * `SoftwareConfig.image_version`.
+   * Optional. The Compute Engine image resource used for cluster instances.
+   * The URI can represent an image or image family.
+   * Image examples:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+   * * `projects/[project_id]/global/images/[image-id]`
+   * * `image-id`
+   * Image family examples. Dataproc will use the most recent
+   * image from the family:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * If the URI is unspecified, it will be inferred from
+   * `SoftwareConfig.image_version` or the system default.
    * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -132,7 +150,7 @@ public interface InstanceGroupConfigOrBuilder * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -153,7 +171,7 @@ public interface InstanceGroupConfigOrBuilder * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -209,11 +227,11 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Optional. Specifies that this instance group contains preemptible
+   * Output only. Specifies that this instance group contains preemptible
    * instances.
    * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @return The isPreemptible. */ @@ -338,8 +356,8 @@ public interface InstanceGroupConfigOrBuilder * *
    * Optional. Specifies the minimum cpu platform for the Instance Group.
-   * See [Dataproc&rarr;Minimum CPU Platform]
-   * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+   * See [Dataproc -> Minimum CPU
+   * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; @@ -352,8 +370,8 @@ public interface InstanceGroupConfigOrBuilder * *
    * Optional. Specifies the minimum cpu platform for the Instance Group.
-   * See [Dataproc&rarr;Minimum CPU Platform]
-   * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+   * See [Dataproc -> Minimum CPU
+   * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* * string min_cpu_platform = 9 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java index 39fcea9d..f2a011f4 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/Job.java @@ -324,6 +324,11 @@ private Job( typeJobCase_ = 23; break; } + case 192: + { + done_ = input.readBool(); + break; + } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { @@ -1514,6 +1519,26 @@ public com.google.protobuf.ByteString getJobUuidBytes() { } } + public static final int DONE_FIELD_NUMBER = 24; + private boolean done_; + /** + * + * + *
+   * Output only. Indicates whether the job is completed. If the value is `false`,
+   * the job is still in progress. If `true`, the job is completed, and
+   * `status.state` field will indicate if it was successful, failed,
+   * or cancelled.
+   * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The done. + */ + public boolean getDone() { + return done_; + } + private byte memoizedIsInitialized = -1; @java.lang.Override @@ -1581,6 +1606,9 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io if (typeJobCase_ == 23) { output.writeMessage(23, (com.google.cloud.dataproc.v1.PrestoJob) typeJob_); } + if (done_ != false) { + output.writeBool(24, done_); + } unknownFields.writeTo(output); } @@ -1668,6 +1696,9 @@ public int getSerializedSize() { com.google.protobuf.CodedOutputStream.computeMessageSize( 23, (com.google.cloud.dataproc.v1.PrestoJob) typeJob_); } + if (done_ != false) { + size += com.google.protobuf.CodedOutputStream.computeBoolSize(24, done_); + } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; @@ -1705,6 +1736,7 @@ public boolean equals(final java.lang.Object obj) { if (!getScheduling().equals(other.getScheduling())) return false; } if (!getJobUuid().equals(other.getJobUuid())) return false; + if (getDone() != other.getDone()) return false; if (!getTypeJobCase().equals(other.getTypeJobCase())) return false; switch (typeJobCase_) { case 3: @@ -1779,6 +1811,8 @@ public int hashCode() { } hash = (37 * hash) + JOB_UUID_FIELD_NUMBER; hash = (53 * hash) + getJobUuid().hashCode(); + hash = (37 * hash) + DONE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(getDone()); switch (typeJobCase_) { case 3: hash = (37 * hash) + HADOOP_JOB_FIELD_NUMBER; @@ -2025,6 +2059,8 @@ public Builder clear() { } jobUuid_ = ""; + done_ = false; + typeJobCase_ = 0; typeJob_ = null; return this; @@ -2153,6 +2189,7 @@ public com.google.cloud.dataproc.v1.Job buildPartial() { result.scheduling_ = schedulingBuilder_.build(); } result.jobUuid_ = jobUuid_; + result.done_ = done_; result.typeJobCase_ = typeJobCase_; onBuilt(); return result; @@ -2282,6 +2319,9 @@ public Builder mergeFrom(com.google.cloud.dataproc.v1.Job other) { jobUuid_ = other.jobUuid_; onChanged(); } + if (other.getDone() != false) { + setDone(other.getDone()); + } switch (other.getTypeJobCase()) { case HADOOP_JOB: { @@ -6381,6 +6421,66 @@ public Builder setJobUuidBytes(com.google.protobuf.ByteString value) { return this; } + private boolean done_; + /** + * + * + *
+     * Output only. Indicates whether the job is completed. If the value is `false`,
+     * the job is still in progress. If `true`, the job is completed, and
+     * `status.state` field will indicate if it was successful, failed,
+     * or cancelled.
+     * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The done. + */ + public boolean getDone() { + return done_; + } + /** + * + * + *
+     * Output only. Indicates whether the job is completed. If the value is `false`,
+     * the job is still in progress. If `true`, the job is completed, and
+     * `status.state` field will indicate if it was successful, failed,
+     * or cancelled.
+     * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The done to set. + * @return This builder for chaining. + */ + public Builder setDone(boolean value) { + + done_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. Indicates whether the job is completed. If the value is `false`,
+     * the job is still in progress. If `true`, the job is completed, and
+     * `status.state` field will indicate if it was successful, failed,
+     * or cancelled.
+     * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return This builder for chaining. + */ + public Builder clearDone() { + + done_ = false; + onChanged(); + return this; + } + @java.lang.Override public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobMetadata.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobMetadata.java new file mode 100644 index 00000000..65e2f132 --- /dev/null +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobMetadata.java @@ -0,0 +1,1381 @@ +/* + * Copyright 2020 Google LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dataproc/v1/jobs.proto + +package com.google.cloud.dataproc.v1; + +/** + * + * + *
+ * Job Operation metadata.
+ * 
+ * + * Protobuf type {@code google.cloud.dataproc.v1.JobMetadata} + */ +public final class JobMetadata extends com.google.protobuf.GeneratedMessageV3 + implements + // @@protoc_insertion_point(message_implements:google.cloud.dataproc.v1.JobMetadata) + JobMetadataOrBuilder { + private static final long serialVersionUID = 0L; + // Use JobMetadata.newBuilder() to construct. + private JobMetadata(com.google.protobuf.GeneratedMessageV3.Builder builder) { + super(builder); + } + + private JobMetadata() { + jobId_ = ""; + operationType_ = ""; + } + + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance(UnusedPrivateParameter unused) { + return new JobMetadata(); + } + + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet getUnknownFields() { + return this.unknownFields; + } + + private JobMetadata( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + case 10: + { + java.lang.String s = input.readStringRequireUtf8(); + + jobId_ = s; + break; + } + case 18: + { + com.google.cloud.dataproc.v1.JobStatus.Builder subBuilder = null; + if (status_ != null) { + subBuilder = status_.toBuilder(); + } + status_ = + input.readMessage( + com.google.cloud.dataproc.v1.JobStatus.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom(status_); + status_ = subBuilder.buildPartial(); + } + + break; + } + case 26: + { + java.lang.String s = input.readStringRequireUtf8(); + + operationType_ = s; + break; + } + case 34: + { + com.google.protobuf.Timestamp.Builder subBuilder = null; + if (startTime_ != null) { + subBuilder = startTime_.toBuilder(); + } + startTime_ = + input.readMessage(com.google.protobuf.Timestamp.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom(startTime_); + startTime_ = subBuilder.buildPartial(); + } + + break; + } + default: + { + if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException(e).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1.JobsProto + .internal_static_google_cloud_dataproc_v1_JobMetadata_descriptor; + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1.JobsProto + .internal_static_google_cloud_dataproc_v1_JobMetadata_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1.JobMetadata.class, + com.google.cloud.dataproc.v1.JobMetadata.Builder.class); + } + + public static final int JOB_ID_FIELD_NUMBER = 1; + private volatile java.lang.Object jobId_; + /** + * + * + *
+   * Output only. The job id.
+   * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The jobId. + */ + public java.lang.String getJobId() { + java.lang.Object ref = jobId_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + jobId_ = s; + return s; + } + } + /** + * + * + *
+   * Output only. The job id.
+   * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for jobId. + */ + public com.google.protobuf.ByteString getJobIdBytes() { + java.lang.Object ref = jobId_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + jobId_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + public static final int STATUS_FIELD_NUMBER = 2; + private com.google.cloud.dataproc.v1.JobStatus status_; + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the status field is set. + */ + public boolean hasStatus() { + return status_ != null; + } + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The status. + */ + public com.google.cloud.dataproc.v1.JobStatus getStatus() { + return status_ == null ? com.google.cloud.dataproc.v1.JobStatus.getDefaultInstance() : status_; + } + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.cloud.dataproc.v1.JobStatusOrBuilder getStatusOrBuilder() { + return getStatus(); + } + + public static final int OPERATION_TYPE_FIELD_NUMBER = 3; + private volatile java.lang.Object operationType_; + /** + * + * + *
+   * Output only. Operation type.
+   * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The operationType. + */ + public java.lang.String getOperationType() { + java.lang.Object ref = operationType_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + operationType_ = s; + return s; + } + } + /** + * + * + *
+   * Output only. Operation type.
+   * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for operationType. + */ + public com.google.protobuf.ByteString getOperationTypeBytes() { + java.lang.Object ref = operationType_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + operationType_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + public static final int START_TIME_FIELD_NUMBER = 4; + private com.google.protobuf.Timestamp startTime_; + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the startTime field is set. + */ + public boolean hasStartTime() { + return startTime_ != null; + } + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The startTime. + */ + public com.google.protobuf.Timestamp getStartTime() { + return startTime_ == null ? com.google.protobuf.Timestamp.getDefaultInstance() : startTime_; + } + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.protobuf.TimestampOrBuilder getStartTimeOrBuilder() { + return getStartTime(); + } + + private byte memoizedIsInitialized = -1; + + @java.lang.Override + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; + + memoizedIsInitialized = 1; + return true; + } + + @java.lang.Override + public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { + if (!getJobIdBytes().isEmpty()) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, jobId_); + } + if (status_ != null) { + output.writeMessage(2, getStatus()); + } + if (!getOperationTypeBytes().isEmpty()) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 3, operationType_); + } + if (startTime_ != null) { + output.writeMessage(4, getStartTime()); + } + unknownFields.writeTo(output); + } + + @java.lang.Override + public int getSerializedSize() { + int size = memoizedSize; + if (size != -1) return size; + + size = 0; + if (!getJobIdBytes().isEmpty()) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, jobId_); + } + if (status_ != null) { + size += com.google.protobuf.CodedOutputStream.computeMessageSize(2, getStatus()); + } + if (!getOperationTypeBytes().isEmpty()) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, operationType_); + } + if (startTime_ != null) { + size += com.google.protobuf.CodedOutputStream.computeMessageSize(4, getStartTime()); + } + size += unknownFields.getSerializedSize(); + memoizedSize = size; + return size; + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof com.google.cloud.dataproc.v1.JobMetadata)) { + return super.equals(obj); + } + com.google.cloud.dataproc.v1.JobMetadata other = (com.google.cloud.dataproc.v1.JobMetadata) obj; + + if (!getJobId().equals(other.getJobId())) return false; + if (hasStatus() != other.hasStatus()) return false; + if (hasStatus()) { + if (!getStatus().equals(other.getStatus())) return false; + } + if (!getOperationType().equals(other.getOperationType())) return false; + if (hasStartTime() != other.hasStartTime()) return false; + if (hasStartTime()) { + if (!getStartTime().equals(other.getStartTime())) return false; + } + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (37 * hash) + JOB_ID_FIELD_NUMBER; + hash = (53 * hash) + getJobId().hashCode(); + if (hasStatus()) { + hash = (37 * hash) + STATUS_FIELD_NUMBER; + hash = (53 * hash) + getStatus().hashCode(); + } + hash = (37 * hash) + OPERATION_TYPE_FIELD_NUMBER; + hash = (53 * hash) + getOperationType().hashCode(); + if (hasStartTime()) { + hash = (37 * hash) + START_TIME_FIELD_NUMBER; + hash = (53 * hash) + getStartTime().hashCode(); + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom(java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom( + java.nio.ByteBuffer data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom( + byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom(java.io.InputStream input) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseDelimitedFrom( + java.io.InputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseDelimitedFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom( + com.google.protobuf.CodedInputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1.JobMetadata parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + @java.lang.Override + public Builder newBuilderForType() { + return newBuilder(); + } + + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } + + public static Builder newBuilder(com.google.cloud.dataproc.v1.JobMetadata prototype) { + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); + } + + @java.lang.Override + protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * + * + *
+   * Job Operation metadata.
+   * 
+ * + * Protobuf type {@code google.cloud.dataproc.v1.JobMetadata} + */ + public static final class Builder extends com.google.protobuf.GeneratedMessageV3.Builder + implements + // @@protoc_insertion_point(builder_implements:google.cloud.dataproc.v1.JobMetadata) + com.google.cloud.dataproc.v1.JobMetadataOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1.JobsProto + .internal_static_google_cloud_dataproc_v1_JobMetadata_descriptor; + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1.JobsProto + .internal_static_google_cloud_dataproc_v1_JobMetadata_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1.JobMetadata.class, + com.google.cloud.dataproc.v1.JobMetadata.Builder.class); + } + + // Construct using com.google.cloud.dataproc.v1.JobMetadata.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders) {} + } + + @java.lang.Override + public Builder clear() { + super.clear(); + jobId_ = ""; + + if (statusBuilder_ == null) { + status_ = null; + } else { + status_ = null; + statusBuilder_ = null; + } + operationType_ = ""; + + if (startTimeBuilder_ == null) { + startTime_ = null; + } else { + startTime_ = null; + startTimeBuilder_ = null; + } + return this; + } + + @java.lang.Override + public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { + return com.google.cloud.dataproc.v1.JobsProto + .internal_static_google_cloud_dataproc_v1_JobMetadata_descriptor; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1.JobMetadata getDefaultInstanceForType() { + return com.google.cloud.dataproc.v1.JobMetadata.getDefaultInstance(); + } + + @java.lang.Override + public com.google.cloud.dataproc.v1.JobMetadata build() { + com.google.cloud.dataproc.v1.JobMetadata result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1.JobMetadata buildPartial() { + com.google.cloud.dataproc.v1.JobMetadata result = + new com.google.cloud.dataproc.v1.JobMetadata(this); + result.jobId_ = jobId_; + if (statusBuilder_ == null) { + result.status_ = status_; + } else { + result.status_ = statusBuilder_.build(); + } + result.operationType_ = operationType_; + if (startTimeBuilder_ == null) { + result.startTime_ = startTime_; + } else { + result.startTime_ = startTimeBuilder_.build(); + } + onBuilt(); + return result; + } + + @java.lang.Override + public Builder clone() { + return super.clone(); + } + + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.setField(field, value); + } + + @java.lang.Override + public Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + + @java.lang.Override + public Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.addRepeatedField(field, value); + } + + @java.lang.Override + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof com.google.cloud.dataproc.v1.JobMetadata) { + return mergeFrom((com.google.cloud.dataproc.v1.JobMetadata) other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(com.google.cloud.dataproc.v1.JobMetadata other) { + if (other == com.google.cloud.dataproc.v1.JobMetadata.getDefaultInstance()) return this; + if (!other.getJobId().isEmpty()) { + jobId_ = other.jobId_; + onChanged(); + } + if (other.hasStatus()) { + mergeStatus(other.getStatus()); + } + if (!other.getOperationType().isEmpty()) { + operationType_ = other.operationType_; + onChanged(); + } + if (other.hasStartTime()) { + mergeStartTime(other.getStartTime()); + } + this.mergeUnknownFields(other.unknownFields); + onChanged(); + return this; + } + + @java.lang.Override + public final boolean isInitialized() { + return true; + } + + @java.lang.Override + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.cloud.dataproc.v1.JobMetadata parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (com.google.cloud.dataproc.v1.JobMetadata) e.getUnfinishedMessage(); + throw e.unwrapIOException(); + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + + private java.lang.Object jobId_ = ""; + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The jobId. + */ + public java.lang.String getJobId() { + java.lang.Object ref = jobId_; + if (!(ref instanceof java.lang.String)) { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + jobId_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for jobId. + */ + public com.google.protobuf.ByteString getJobIdBytes() { + java.lang.Object ref = jobId_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + jobId_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The jobId to set. + * @return This builder for chaining. + */ + public Builder setJobId(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + + jobId_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return This builder for chaining. + */ + public Builder clearJobId() { + + jobId_ = getDefaultInstance().getJobId(); + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The bytes for jobId to set. + * @return This builder for chaining. + */ + public Builder setJobIdBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + + jobId_ = value; + onChanged(); + return this; + } + + private com.google.cloud.dataproc.v1.JobStatus status_; + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.JobStatus, + com.google.cloud.dataproc.v1.JobStatus.Builder, + com.google.cloud.dataproc.v1.JobStatusOrBuilder> + statusBuilder_; + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the status field is set. + */ + public boolean hasStatus() { + return statusBuilder_ != null || status_ != null; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The status. + */ + public com.google.cloud.dataproc.v1.JobStatus getStatus() { + if (statusBuilder_ == null) { + return status_ == null + ? com.google.cloud.dataproc.v1.JobStatus.getDefaultInstance() + : status_; + } else { + return statusBuilder_.getMessage(); + } + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder setStatus(com.google.cloud.dataproc.v1.JobStatus value) { + if (statusBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + status_ = value; + onChanged(); + } else { + statusBuilder_.setMessage(value); + } + + return this; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder setStatus(com.google.cloud.dataproc.v1.JobStatus.Builder builderForValue) { + if (statusBuilder_ == null) { + status_ = builderForValue.build(); + onChanged(); + } else { + statusBuilder_.setMessage(builderForValue.build()); + } + + return this; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder mergeStatus(com.google.cloud.dataproc.v1.JobStatus value) { + if (statusBuilder_ == null) { + if (status_ != null) { + status_ = + com.google.cloud.dataproc.v1.JobStatus.newBuilder(status_) + .mergeFrom(value) + .buildPartial(); + } else { + status_ = value; + } + onChanged(); + } else { + statusBuilder_.mergeFrom(value); + } + + return this; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder clearStatus() { + if (statusBuilder_ == null) { + status_ = null; + onChanged(); + } else { + status_ = null; + statusBuilder_ = null; + } + + return this; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.cloud.dataproc.v1.JobStatus.Builder getStatusBuilder() { + + onChanged(); + return getStatusFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.cloud.dataproc.v1.JobStatusOrBuilder getStatusOrBuilder() { + if (statusBuilder_ != null) { + return statusBuilder_.getMessageOrBuilder(); + } else { + return status_ == null + ? com.google.cloud.dataproc.v1.JobStatus.getDefaultInstance() + : status_; + } + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.JobStatus, + com.google.cloud.dataproc.v1.JobStatus.Builder, + com.google.cloud.dataproc.v1.JobStatusOrBuilder> + getStatusFieldBuilder() { + if (statusBuilder_ == null) { + statusBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.JobStatus, + com.google.cloud.dataproc.v1.JobStatus.Builder, + com.google.cloud.dataproc.v1.JobStatusOrBuilder>( + getStatus(), getParentForChildren(), isClean()); + status_ = null; + } + return statusBuilder_; + } + + private java.lang.Object operationType_ = ""; + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The operationType. + */ + public java.lang.String getOperationType() { + java.lang.Object ref = operationType_; + if (!(ref instanceof java.lang.String)) { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + operationType_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for operationType. + */ + public com.google.protobuf.ByteString getOperationTypeBytes() { + java.lang.Object ref = operationType_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + operationType_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The operationType to set. + * @return This builder for chaining. + */ + public Builder setOperationType(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + + operationType_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return This builder for chaining. + */ + public Builder clearOperationType() { + + operationType_ = getDefaultInstance().getOperationType(); + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The bytes for operationType to set. + * @return This builder for chaining. + */ + public Builder setOperationTypeBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + + operationType_ = value; + onChanged(); + return this; + } + + private com.google.protobuf.Timestamp startTime_; + private com.google.protobuf.SingleFieldBuilderV3< + com.google.protobuf.Timestamp, + com.google.protobuf.Timestamp.Builder, + com.google.protobuf.TimestampOrBuilder> + startTimeBuilder_; + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the startTime field is set. + */ + public boolean hasStartTime() { + return startTimeBuilder_ != null || startTime_ != null; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The startTime. + */ + public com.google.protobuf.Timestamp getStartTime() { + if (startTimeBuilder_ == null) { + return startTime_ == null ? com.google.protobuf.Timestamp.getDefaultInstance() : startTime_; + } else { + return startTimeBuilder_.getMessage(); + } + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder setStartTime(com.google.protobuf.Timestamp value) { + if (startTimeBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + startTime_ = value; + onChanged(); + } else { + startTimeBuilder_.setMessage(value); + } + + return this; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder setStartTime(com.google.protobuf.Timestamp.Builder builderForValue) { + if (startTimeBuilder_ == null) { + startTime_ = builderForValue.build(); + onChanged(); + } else { + startTimeBuilder_.setMessage(builderForValue.build()); + } + + return this; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder mergeStartTime(com.google.protobuf.Timestamp value) { + if (startTimeBuilder_ == null) { + if (startTime_ != null) { + startTime_ = + com.google.protobuf.Timestamp.newBuilder(startTime_).mergeFrom(value).buildPartial(); + } else { + startTime_ = value; + } + onChanged(); + } else { + startTimeBuilder_.mergeFrom(value); + } + + return this; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder clearStartTime() { + if (startTimeBuilder_ == null) { + startTime_ = null; + onChanged(); + } else { + startTime_ = null; + startTimeBuilder_ = null; + } + + return this; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.protobuf.Timestamp.Builder getStartTimeBuilder() { + + onChanged(); + return getStartTimeFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.protobuf.TimestampOrBuilder getStartTimeOrBuilder() { + if (startTimeBuilder_ != null) { + return startTimeBuilder_.getMessageOrBuilder(); + } else { + return startTime_ == null ? com.google.protobuf.Timestamp.getDefaultInstance() : startTime_; + } + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.protobuf.Timestamp, + com.google.protobuf.Timestamp.Builder, + com.google.protobuf.TimestampOrBuilder> + getStartTimeFieldBuilder() { + if (startTimeBuilder_ == null) { + startTimeBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.protobuf.Timestamp, + com.google.protobuf.Timestamp.Builder, + com.google.protobuf.TimestampOrBuilder>( + getStartTime(), getParentForChildren(), isClean()); + startTime_ = null; + } + return startTimeBuilder_; + } + + @java.lang.Override + public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + + // @@protoc_insertion_point(builder_scope:google.cloud.dataproc.v1.JobMetadata) + } + + // @@protoc_insertion_point(class_scope:google.cloud.dataproc.v1.JobMetadata) + private static final com.google.cloud.dataproc.v1.JobMetadata DEFAULT_INSTANCE; + + static { + DEFAULT_INSTANCE = new com.google.cloud.dataproc.v1.JobMetadata(); + } + + public static com.google.cloud.dataproc.v1.JobMetadata getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + private static final com.google.protobuf.Parser PARSER = + new com.google.protobuf.AbstractParser() { + @java.lang.Override + public JobMetadata parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new JobMetadata(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1.JobMetadata getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } +} diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobMetadataOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobMetadataOrBuilder.java new file mode 100644 index 00000000..f1343e20 --- /dev/null +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobMetadataOrBuilder.java @@ -0,0 +1,154 @@ +/* + * Copyright 2020 Google LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dataproc/v1/jobs.proto + +package com.google.cloud.dataproc.v1; + +public interface JobMetadataOrBuilder + extends + // @@protoc_insertion_point(interface_extends:google.cloud.dataproc.v1.JobMetadata) + com.google.protobuf.MessageOrBuilder { + + /** + * + * + *
+   * Output only. The job id.
+   * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The jobId. + */ + java.lang.String getJobId(); + /** + * + * + *
+   * Output only. The job id.
+   * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for jobId. + */ + com.google.protobuf.ByteString getJobIdBytes(); + + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the status field is set. + */ + boolean hasStatus(); + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The status. + */ + com.google.cloud.dataproc.v1.JobStatus getStatus(); + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + com.google.cloud.dataproc.v1.JobStatusOrBuilder getStatusOrBuilder(); + + /** + * + * + *
+   * Output only. Operation type.
+   * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The operationType. + */ + java.lang.String getOperationType(); + /** + * + * + *
+   * Output only. Operation type.
+   * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for operationType. + */ + com.google.protobuf.ByteString getOperationTypeBytes(); + + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the startTime field is set. + */ + boolean hasStartTime(); + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The startTime. + */ + com.google.protobuf.Timestamp getStartTime(); + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + com.google.protobuf.TimestampOrBuilder getStartTimeOrBuilder(); +} diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobOrBuilder.java index b1b8548b..f70e389f 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobOrBuilder.java @@ -832,5 +832,21 @@ public interface JobOrBuilder */ com.google.protobuf.ByteString getJobUuidBytes(); + /** + * + * + *
+   * Output only. Indicates whether the job is completed. If the value is `false`,
+   * the job is still in progress. If `true`, the job is completed, and
+   * `status.state` field will indicate if it was successful, failed,
+   * or cancelled.
+   * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The done. + */ + boolean getDone(); + public com.google.cloud.dataproc.v1.Job.TypeJobCase getTypeJobCase(); } diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobsProto.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobsProto.java index 9a960862..78929c12 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobsProto.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/JobsProto.java @@ -147,6 +147,10 @@ public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry r internal_static_google_cloud_dataproc_v1_SubmitJobRequest_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_google_cloud_dataproc_v1_SubmitJobRequest_fieldAccessorTable; + static final com.google.protobuf.Descriptors.Descriptor + internal_static_google_cloud_dataproc_v1_JobMetadata_descriptor; + static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_google_cloud_dataproc_v1_JobMetadata_fieldAccessorTable; static final com.google.protobuf.Descriptors.Descriptor internal_static_google_cloud_dataproc_v1_GetJobRequest_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable @@ -300,7 +304,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "State\022\025\n\021STATE_UNSPECIFIED\020\000\022\007\n\003NEW\020\001\022\016\n" + "\nNEW_SAVING\020\002\022\r\n\tSUBMITTED\020\003\022\014\n\010ACCEPTED" + "\020\004\022\013\n\007RUNNING\020\005\022\014\n\010FINISHED\020\006\022\n\n\006FAILED\020" - + "\007\022\n\n\006KILLED\020\010\"\354\010\n\003Job\022>\n\treference\030\001 \001(\013" + + "\007\022\n\n\006KILLED\020\010\"\377\010\n\003Job\022>\n\treference\030\001 \001(\013" + "2&.google.cloud.dataproc.v1.JobReference" + "B\003\340A\001\022>\n\tplacement\030\002 \001(\0132&.google.cloud." + "dataproc.v1.JobPlacementB\003\340A\002\022>\n\nhadoop_" @@ -327,67 +331,78 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "e.cloud.dataproc.v1.Job.LabelsEntryB\003\340A\001" + "\022@\n\nscheduling\030\024 \001(\0132\'.google.cloud.data" + "proc.v1.JobSchedulingB\003\340A\001\022\025\n\010job_uuid\030\026" - + " \001(\tB\003\340A\003\032-\n\013LabelsEntry\022\013\n\003key\030\001 \001(\t\022\r\n" - + "\005value\030\002 \001(\t:\0028\001B\n\n\010type_job\"3\n\rJobSched" - + "uling\022\"\n\025max_failures_per_hour\030\001 \001(\005B\003\340A" - + "\001\"\212\001\n\020SubmitJobRequest\022\027\n\nproject_id\030\001 \001" - + "(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022/\n\003job\030\002 \001(" - + "\0132\035.google.cloud.dataproc.v1.JobB\003\340A\002\022\027\n" - + "\nrequest_id\030\004 \001(\tB\003\340A\001\"R\n\rGetJobRequest\022" - + "\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\t" - + "B\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\002\"\263\002\n\017ListJobsR" - + "equest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006regio" - + "n\030\006 \001(\tB\003\340A\002\022\026\n\tpage_size\030\002 \001(\005B\003\340A\001\022\027\n\n" - + "page_token\030\003 \001(\tB\003\340A\001\022\031\n\014cluster_name\030\004 " - + "\001(\tB\003\340A\001\022Y\n\021job_state_matcher\030\005 \001(\01629.go" - + "ogle.cloud.dataproc.v1.ListJobsRequest.J" - + "obStateMatcherB\003\340A\001\022\023\n\006filter\030\007 \001(\tB\003\340A\001" - + "\"6\n\017JobStateMatcher\022\007\n\003ALL\020\000\022\n\n\006ACTIVE\020\001" - + "\022\016\n\nNON_ACTIVE\020\002\"\274\001\n\020UpdateJobRequest\022\027\n" - + "\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\002 \001(\tB\003" - + "\340A\002\022\023\n\006job_id\030\003 \001(\tB\003\340A\002\022/\n\003job\030\004 \001(\0132\035." - + "google.cloud.dataproc.v1.JobB\003\340A\002\0224\n\013upd" - + "ate_mask\030\005 \001(\0132\032.google.protobuf.FieldMa" - + "skB\003\340A\002\"b\n\020ListJobsResponse\0220\n\004jobs\030\001 \003(" - + "\0132\035.google.cloud.dataproc.v1.JobB\003\340A\003\022\034\n" - + "\017next_page_token\030\002 \001(\tB\003\340A\001\"U\n\020CancelJob" - + "Request\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006regi" - + "on\030\003 \001(\tB\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\002\"U\n\020De" - + "leteJobRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022" - + "\023\n\006region\030\003 \001(\tB\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A" - + "\0022\233\t\n\rJobController\022\261\001\n\tSubmitJob\022*.goog" - + "le.cloud.dataproc.v1.SubmitJobRequest\032\035." - + "google.cloud.dataproc.v1.Job\"Y\202\323\344\223\002;\"6/v" - + "1/projects/{project_id}/regions/{region}" - + "/jobs:submit:\001*\332A\025project_id,region,job\022" - + "\255\001\n\006GetJob\022\'.google.cloud.dataproc.v1.Ge" - + "tJobRequest\032\035.google.cloud.dataproc.v1.J" - + "ob\"[\202\323\344\223\002:\0228/v1/projects/{project_id}/re" - + "gions/{region}/jobs/{job_id}\332A\030project_i" - + "d,region,job_id\022\311\001\n\010ListJobs\022).google.cl" - + "oud.dataproc.v1.ListJobsRequest\032*.google" - + ".cloud.dataproc.v1.ListJobsResponse\"f\202\323\344" - + "\223\0021\022//v1/projects/{project_id}/regions/{" - + "region}/jobs\332A\021project_id,region\332A\030proje" - + "ct_id,region,filter\022\235\001\n\tUpdateJob\022*.goog" - + "le.cloud.dataproc.v1.UpdateJobRequest\032\035." - + "google.cloud.dataproc.v1.Job\"E\202\323\344\223\002?28/v" - + "1/projects/{project_id}/regions/{region}" - + "/jobs/{job_id}:\003job\022\275\001\n\tCancelJob\022*.goog" - + "le.cloud.dataproc.v1.CancelJobRequest\032\035." - + "google.cloud.dataproc.v1.Job\"e\202\323\344\223\002D\"?/v" - + "1/projects/{project_id}/regions/{region}" - + "/jobs/{job_id}:cancel:\001*\332A\030project_id,re" - + "gion,job_id\022\254\001\n\tDeleteJob\022*.google.cloud" - + ".dataproc.v1.DeleteJobRequest\032\026.google.p" - + "rotobuf.Empty\"[\202\323\344\223\002:*8/v1/projects/{pro" - + "ject_id}/regions/{region}/jobs/{job_id}\332" - + "A\030project_id,region,job_id\032K\312A\027dataproc." - + "googleapis.com\322A.https://www.googleapis." - + "com/auth/cloud-platformBm\n\034com.google.cl" - + "oud.dataproc.v1B\tJobsProtoP\001Z@google.gol" - + "ang.org/genproto/googleapis/cloud/datapr" - + "oc/v1;dataprocb\006proto3" + + " \001(\tB\003\340A\003\022\021\n\004done\030\030 \001(\010B\003\340A\003\032-\n\013LabelsEn" + + "try\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001B\n\n\010t" + + "ype_job\"3\n\rJobScheduling\022\"\n\025max_failures" + + "_per_hour\030\001 \001(\005B\003\340A\001\"\212\001\n\020SubmitJobReques" + + "t\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001" + + "(\tB\003\340A\002\022/\n\003job\030\002 \001(\0132\035.google.cloud.data" + + "proc.v1.JobB\003\340A\002\022\027\n\nrequest_id\030\004 \001(\tB\003\340A" + + "\001\"\256\001\n\013JobMetadata\022\023\n\006job_id\030\001 \001(\tB\003\340A\003\0228" + + "\n\006status\030\002 \001(\0132#.google.cloud.dataproc.v" + + "1.JobStatusB\003\340A\003\022\033\n\016operation_type\030\003 \001(\t" + + "B\003\340A\003\0223\n\nstart_time\030\004 \001(\0132\032.google.proto" + + "buf.TimestampB\003\340A\003\"R\n\rGetJobRequest\022\027\n\np" + + "roject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A" + + "\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\002\"\263\002\n\017ListJobsReque" + + "st\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\006 " + + "\001(\tB\003\340A\002\022\026\n\tpage_size\030\002 \001(\005B\003\340A\001\022\027\n\npage" + + "_token\030\003 \001(\tB\003\340A\001\022\031\n\014cluster_name\030\004 \001(\tB" + + "\003\340A\001\022Y\n\021job_state_matcher\030\005 \001(\01629.google" + + ".cloud.dataproc.v1.ListJobsRequest.JobSt" + + "ateMatcherB\003\340A\001\022\023\n\006filter\030\007 \001(\tB\003\340A\001\"6\n\017" + + "JobStateMatcher\022\007\n\003ALL\020\000\022\n\n\006ACTIVE\020\001\022\016\n\n" + + "NON_ACTIVE\020\002\"\274\001\n\020UpdateJobRequest\022\027\n\npro" + + "ject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\002 \001(\tB\003\340A\002\022" + + "\023\n\006job_id\030\003 \001(\tB\003\340A\002\022/\n\003job\030\004 \001(\0132\035.goog" + + "le.cloud.dataproc.v1.JobB\003\340A\002\0224\n\013update_" + + "mask\030\005 \001(\0132\032.google.protobuf.FieldMaskB\003" + + "\340A\002\"b\n\020ListJobsResponse\0220\n\004jobs\030\001 \003(\0132\035." + + "google.cloud.dataproc.v1.JobB\003\340A\003\022\034\n\017nex" + + "t_page_token\030\002 \001(\tB\003\340A\001\"U\n\020CancelJobRequ" + + "est\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003" + + " \001(\tB\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\002\"U\n\020Delete" + + "JobRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006r" + + "egion\030\003 \001(\tB\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\0022\374\n" + + "\n\rJobController\022\261\001\n\tSubmitJob\022*.google.c" + + "loud.dataproc.v1.SubmitJobRequest\032\035.goog" + + "le.cloud.dataproc.v1.Job\"Y\202\323\344\223\002;\"6/v1/pr" + + "ojects/{project_id}/regions/{region}/job" + + "s:submit:\001*\332A\025project_id,region,job\022\336\001\n\024" + + "SubmitJobAsOperation\022*.google.cloud.data" + + "proc.v1.SubmitJobRequest\032\035.google.longru" + + "nning.Operation\"{\202\323\344\223\002F\"A/v1/projects/{p" + + "roject_id}/regions/{region}/jobs:submitA" + + "sOperation:\001*\332A\027project_id, region, job\312" + + "A\022\n\003Job\022\013JobMetadata\022\255\001\n\006GetJob\022\'.google" + + ".cloud.dataproc.v1.GetJobRequest\032\035.googl" + + "e.cloud.dataproc.v1.Job\"[\202\323\344\223\002:\0228/v1/pro" + + "jects/{project_id}/regions/{region}/jobs" + + "/{job_id}\332A\030project_id,region,job_id\022\311\001\n" + + "\010ListJobs\022).google.cloud.dataproc.v1.Lis" + + "tJobsRequest\032*.google.cloud.dataproc.v1." + + "ListJobsResponse\"f\202\323\344\223\0021\022//v1/projects/{" + + "project_id}/regions/{region}/jobs\332A\021proj" + + "ect_id,region\332A\030project_id,region,filter" + + "\022\235\001\n\tUpdateJob\022*.google.cloud.dataproc.v" + + "1.UpdateJobRequest\032\035.google.cloud.datapr" + + "oc.v1.Job\"E\202\323\344\223\002?28/v1/projects/{project" + + "_id}/regions/{region}/jobs/{job_id}:\003job" + + "\022\275\001\n\tCancelJob\022*.google.cloud.dataproc.v" + + "1.CancelJobRequest\032\035.google.cloud.datapr" + + "oc.v1.Job\"e\202\323\344\223\002D\"?/v1/projects/{project" + + "_id}/regions/{region}/jobs/{job_id}:canc" + + "el:\001*\332A\030project_id,region,job_id\022\254\001\n\tDel" + + "eteJob\022*.google.cloud.dataproc.v1.Delete" + + "JobRequest\032\026.google.protobuf.Empty\"[\202\323\344\223" + + "\002:*8/v1/projects/{project_id}/regions/{r" + + "egion}/jobs/{job_id}\332A\030project_id,region" + + ",job_id\032K\312A\027dataproc.googleapis.com\322A.ht" + + "tps://www.googleapis.com/auth/cloud-plat" + + "formBm\n\034com.google.cloud.dataproc.v1B\tJo" + + "bsProtoP\001Z@google.golang.org/genproto/go" + + "ogleapis/cloud/dataproc/v1;dataprocb\006pro" + + "to3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( @@ -682,6 +697,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Labels", "Scheduling", "JobUuid", + "Done", "TypeJob", }); internal_static_google_cloud_dataproc_v1_Job_LabelsEntry_descriptor = @@ -708,8 +724,16 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { new java.lang.String[] { "ProjectId", "Region", "Job", "RequestId", }); - internal_static_google_cloud_dataproc_v1_GetJobRequest_descriptor = + internal_static_google_cloud_dataproc_v1_JobMetadata_descriptor = getDescriptor().getMessageTypes().get(17); + internal_static_google_cloud_dataproc_v1_JobMetadata_fieldAccessorTable = + new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_google_cloud_dataproc_v1_JobMetadata_descriptor, + new java.lang.String[] { + "JobId", "Status", "OperationType", "StartTime", + }); + internal_static_google_cloud_dataproc_v1_GetJobRequest_descriptor = + getDescriptor().getMessageTypes().get(18); internal_static_google_cloud_dataproc_v1_GetJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1_GetJobRequest_descriptor, @@ -717,7 +741,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "JobId", }); internal_static_google_cloud_dataproc_v1_ListJobsRequest_descriptor = - getDescriptor().getMessageTypes().get(18); + getDescriptor().getMessageTypes().get(19); internal_static_google_cloud_dataproc_v1_ListJobsRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1_ListJobsRequest_descriptor, @@ -731,7 +755,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Filter", }); internal_static_google_cloud_dataproc_v1_UpdateJobRequest_descriptor = - getDescriptor().getMessageTypes().get(19); + getDescriptor().getMessageTypes().get(20); internal_static_google_cloud_dataproc_v1_UpdateJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1_UpdateJobRequest_descriptor, @@ -739,7 +763,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "JobId", "Job", "UpdateMask", }); internal_static_google_cloud_dataproc_v1_ListJobsResponse_descriptor = - getDescriptor().getMessageTypes().get(20); + getDescriptor().getMessageTypes().get(21); internal_static_google_cloud_dataproc_v1_ListJobsResponse_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1_ListJobsResponse_descriptor, @@ -747,7 +771,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Jobs", "NextPageToken", }); internal_static_google_cloud_dataproc_v1_CancelJobRequest_descriptor = - getDescriptor().getMessageTypes().get(21); + getDescriptor().getMessageTypes().get(22); internal_static_google_cloud_dataproc_v1_CancelJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1_CancelJobRequest_descriptor, @@ -755,7 +779,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "JobId", }); internal_static_google_cloud_dataproc_v1_DeleteJobRequest_descriptor = - getDescriptor().getMessageTypes().get(22); + getDescriptor().getMessageTypes().get(23); internal_static_google_cloud_dataproc_v1_DeleteJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1_DeleteJobRequest_descriptor, @@ -769,6 +793,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { registry.add(com.google.api.AnnotationsProto.http); registry.add(com.google.api.ClientProto.methodSignature); registry.add(com.google.api.ClientProto.oauthScopes); + registry.add(com.google.longrunning.OperationsProto.operationInfo); com.google.protobuf.Descriptors.FileDescriptor.internalUpdateFileDescriptor( descriptor, registry); com.google.api.AnnotationsProto.getDescriptor(); diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfig.java index 4bb46159..590fdf3a 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfig.java @@ -223,7 +223,8 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * * *
-   * Optional. Flag to indicate whether to Kerberize the cluster.
+   * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+   * this field to true to enable Kerberos on a cluster.
    * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1458,7 +1459,8 @@ public Builder mergeFrom( * * *
-     * Optional. Flag to indicate whether to Kerberize the cluster.
+     * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+     * this field to true to enable Kerberos on a cluster.
      * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1472,7 +1474,8 @@ public boolean getEnableKerberos() { * * *
-     * Optional. Flag to indicate whether to Kerberize the cluster.
+     * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+     * this field to true to enable Kerberos on a cluster.
      * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1490,7 +1493,8 @@ public Builder setEnableKerberos(boolean value) { * * *
-     * Optional. Flag to indicate whether to Kerberize the cluster.
+     * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+     * this field to true to enable Kerberos on a cluster.
      * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfigOrBuilder.java index 9f4dc82f..426c5c3a 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/KerberosConfigOrBuilder.java @@ -27,7 +27,8 @@ public interface KerberosConfigOrBuilder * * *
-   * Optional. Flag to indicate whether to Kerberize the cluster.
+   * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+   * this field to true to enable Kerberos on a cluster.
    * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJob.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJob.java index db3dfd1f..3c6be289 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJob.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJob.java @@ -214,6 +214,38 @@ private OrderedJob( prerequisiteStepIds_.add(s); break; } + case 90: + { + com.google.cloud.dataproc.v1.SparkRJob.Builder subBuilder = null; + if (jobTypeCase_ == 11) { + subBuilder = ((com.google.cloud.dataproc.v1.SparkRJob) jobType_).toBuilder(); + } + jobType_ = + input.readMessage( + com.google.cloud.dataproc.v1.SparkRJob.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom((com.google.cloud.dataproc.v1.SparkRJob) jobType_); + jobType_ = subBuilder.buildPartial(); + } + jobTypeCase_ = 11; + break; + } + case 98: + { + com.google.cloud.dataproc.v1.PrestoJob.Builder subBuilder = null; + if (jobTypeCase_ == 12) { + subBuilder = ((com.google.cloud.dataproc.v1.PrestoJob) jobType_).toBuilder(); + } + jobType_ = + input.readMessage( + com.google.cloud.dataproc.v1.PrestoJob.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom((com.google.cloud.dataproc.v1.PrestoJob) jobType_); + jobType_ = subBuilder.buildPartial(); + } + jobTypeCase_ = 12; + break; + } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { @@ -274,7 +306,9 @@ public enum JobTypeCase PYSPARK_JOB(4), HIVE_JOB(5), PIG_JOB(6), + SPARK_R_JOB(11), SPARK_SQL_JOB(7), + PRESTO_JOB(12), JOBTYPE_NOT_SET(0); private final int value; @@ -303,8 +337,12 @@ public static JobTypeCase forNumber(int value) { return HIVE_JOB; case 6: return PIG_JOB; + case 11: + return SPARK_R_JOB; case 7: return SPARK_SQL_JOB; + case 12: + return PRESTO_JOB; case 0: return JOBTYPE_NOT_SET; default: @@ -524,6 +562,54 @@ public com.google.cloud.dataproc.v1.PigJobOrBuilder getPigJobOrBuilder() { return com.google.cloud.dataproc.v1.PigJob.getDefaultInstance(); } + public static final int SPARK_R_JOB_FIELD_NUMBER = 11; + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + * + * @return Whether the sparkRJob field is set. + */ + public boolean hasSparkRJob() { + return jobTypeCase_ == 11; + } + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + * + * @return The sparkRJob. + */ + public com.google.cloud.dataproc.v1.SparkRJob getSparkRJob() { + if (jobTypeCase_ == 11) { + return (com.google.cloud.dataproc.v1.SparkRJob) jobType_; + } + return com.google.cloud.dataproc.v1.SparkRJob.getDefaultInstance(); + } + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + public com.google.cloud.dataproc.v1.SparkRJobOrBuilder getSparkRJobOrBuilder() { + if (jobTypeCase_ == 11) { + return (com.google.cloud.dataproc.v1.SparkRJob) jobType_; + } + return com.google.cloud.dataproc.v1.SparkRJob.getDefaultInstance(); + } + public static final int SPARK_SQL_JOB_FIELD_NUMBER = 7; /** * .google.cloud.dataproc.v1.SparkSqlJob spark_sql_job = 7; @@ -552,6 +638,54 @@ public com.google.cloud.dataproc.v1.SparkSqlJobOrBuilder getSparkSqlJobOrBuilder return com.google.cloud.dataproc.v1.SparkSqlJob.getDefaultInstance(); } + public static final int PRESTO_JOB_FIELD_NUMBER = 12; + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + * + * @return Whether the prestoJob field is set. + */ + public boolean hasPrestoJob() { + return jobTypeCase_ == 12; + } + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + * + * @return The prestoJob. + */ + public com.google.cloud.dataproc.v1.PrestoJob getPrestoJob() { + if (jobTypeCase_ == 12) { + return (com.google.cloud.dataproc.v1.PrestoJob) jobType_; + } + return com.google.cloud.dataproc.v1.PrestoJob.getDefaultInstance(); + } + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + public com.google.cloud.dataproc.v1.PrestoJobOrBuilder getPrestoJobOrBuilder() { + if (jobTypeCase_ == 12) { + return (com.google.cloud.dataproc.v1.PrestoJob) jobType_; + } + return com.google.cloud.dataproc.v1.PrestoJob.getDefaultInstance(); + } + public static final int LABELS_FIELD_NUMBER = 8; private static final class LabelsDefaultEntryHolder { @@ -833,6 +967,12 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io com.google.protobuf.GeneratedMessageV3.writeString( output, 10, prerequisiteStepIds_.getRaw(i)); } + if (jobTypeCase_ == 11) { + output.writeMessage(11, (com.google.cloud.dataproc.v1.SparkRJob) jobType_); + } + if (jobTypeCase_ == 12) { + output.writeMessage(12, (com.google.cloud.dataproc.v1.PrestoJob) jobType_); + } unknownFields.writeTo(output); } @@ -896,6 +1036,16 @@ public int getSerializedSize() { size += dataSize; size += 1 * getPrerequisiteStepIdsList().size(); } + if (jobTypeCase_ == 11) { + size += + com.google.protobuf.CodedOutputStream.computeMessageSize( + 11, (com.google.cloud.dataproc.v1.SparkRJob) jobType_); + } + if (jobTypeCase_ == 12) { + size += + com.google.protobuf.CodedOutputStream.computeMessageSize( + 12, (com.google.cloud.dataproc.v1.PrestoJob) jobType_); + } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; @@ -935,9 +1085,15 @@ public boolean equals(final java.lang.Object obj) { case 6: if (!getPigJob().equals(other.getPigJob())) return false; break; + case 11: + if (!getSparkRJob().equals(other.getSparkRJob())) return false; + break; case 7: if (!getSparkSqlJob().equals(other.getSparkSqlJob())) return false; break; + case 12: + if (!getPrestoJob().equals(other.getPrestoJob())) return false; + break; case 0: default: } @@ -987,10 +1143,18 @@ public int hashCode() { hash = (37 * hash) + PIG_JOB_FIELD_NUMBER; hash = (53 * hash) + getPigJob().hashCode(); break; + case 11: + hash = (37 * hash) + SPARK_R_JOB_FIELD_NUMBER; + hash = (53 * hash) + getSparkRJob().hashCode(); + break; case 7: hash = (37 * hash) + SPARK_SQL_JOB_FIELD_NUMBER; hash = (53 * hash) + getSparkSqlJob().hashCode(); break; + case 12: + hash = (37 * hash) + PRESTO_JOB_FIELD_NUMBER; + hash = (53 * hash) + getPrestoJob().hashCode(); + break; case 0: default: } @@ -1236,6 +1400,13 @@ public com.google.cloud.dataproc.v1.OrderedJob buildPartial() { result.jobType_ = pigJobBuilder_.build(); } } + if (jobTypeCase_ == 11) { + if (sparkRJobBuilder_ == null) { + result.jobType_ = jobType_; + } else { + result.jobType_ = sparkRJobBuilder_.build(); + } + } if (jobTypeCase_ == 7) { if (sparkSqlJobBuilder_ == null) { result.jobType_ = jobType_; @@ -1243,6 +1414,13 @@ public com.google.cloud.dataproc.v1.OrderedJob buildPartial() { result.jobType_ = sparkSqlJobBuilder_.build(); } } + if (jobTypeCase_ == 12) { + if (prestoJobBuilder_ == null) { + result.jobType_ = jobType_; + } else { + result.jobType_ = prestoJobBuilder_.build(); + } + } result.labels_ = internalGetLabels(); result.labels_.makeImmutable(); if (schedulingBuilder_ == null) { @@ -1349,11 +1527,21 @@ public Builder mergeFrom(com.google.cloud.dataproc.v1.OrderedJob other) { mergePigJob(other.getPigJob()); break; } + case SPARK_R_JOB: + { + mergeSparkRJob(other.getSparkRJob()); + break; + } case SPARK_SQL_JOB: { mergeSparkSqlJob(other.getSparkSqlJob()); break; } + case PRESTO_JOB: + { + mergePrestoJob(other.getPrestoJob()); + break; + } case JOBTYPE_NOT_SET: { break; @@ -2231,6 +2419,211 @@ public com.google.cloud.dataproc.v1.PigJobOrBuilder getPigJobOrBuilder() { return pigJobBuilder_; } + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.SparkRJob, + com.google.cloud.dataproc.v1.SparkRJob.Builder, + com.google.cloud.dataproc.v1.SparkRJobOrBuilder> + sparkRJobBuilder_; + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + * + * @return Whether the sparkRJob field is set. + */ + public boolean hasSparkRJob() { + return jobTypeCase_ == 11; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + * + * @return The sparkRJob. + */ + public com.google.cloud.dataproc.v1.SparkRJob getSparkRJob() { + if (sparkRJobBuilder_ == null) { + if (jobTypeCase_ == 11) { + return (com.google.cloud.dataproc.v1.SparkRJob) jobType_; + } + return com.google.cloud.dataproc.v1.SparkRJob.getDefaultInstance(); + } else { + if (jobTypeCase_ == 11) { + return sparkRJobBuilder_.getMessage(); + } + return com.google.cloud.dataproc.v1.SparkRJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + public Builder setSparkRJob(com.google.cloud.dataproc.v1.SparkRJob value) { + if (sparkRJobBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + jobType_ = value; + onChanged(); + } else { + sparkRJobBuilder_.setMessage(value); + } + jobTypeCase_ = 11; + return this; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + public Builder setSparkRJob(com.google.cloud.dataproc.v1.SparkRJob.Builder builderForValue) { + if (sparkRJobBuilder_ == null) { + jobType_ = builderForValue.build(); + onChanged(); + } else { + sparkRJobBuilder_.setMessage(builderForValue.build()); + } + jobTypeCase_ = 11; + return this; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + public Builder mergeSparkRJob(com.google.cloud.dataproc.v1.SparkRJob value) { + if (sparkRJobBuilder_ == null) { + if (jobTypeCase_ == 11 + && jobType_ != com.google.cloud.dataproc.v1.SparkRJob.getDefaultInstance()) { + jobType_ = + com.google.cloud.dataproc.v1.SparkRJob.newBuilder( + (com.google.cloud.dataproc.v1.SparkRJob) jobType_) + .mergeFrom(value) + .buildPartial(); + } else { + jobType_ = value; + } + onChanged(); + } else { + if (jobTypeCase_ == 11) { + sparkRJobBuilder_.mergeFrom(value); + } + sparkRJobBuilder_.setMessage(value); + } + jobTypeCase_ = 11; + return this; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + public Builder clearSparkRJob() { + if (sparkRJobBuilder_ == null) { + if (jobTypeCase_ == 11) { + jobTypeCase_ = 0; + jobType_ = null; + onChanged(); + } + } else { + if (jobTypeCase_ == 11) { + jobTypeCase_ = 0; + jobType_ = null; + } + sparkRJobBuilder_.clear(); + } + return this; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + public com.google.cloud.dataproc.v1.SparkRJob.Builder getSparkRJobBuilder() { + return getSparkRJobFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + public com.google.cloud.dataproc.v1.SparkRJobOrBuilder getSparkRJobOrBuilder() { + if ((jobTypeCase_ == 11) && (sparkRJobBuilder_ != null)) { + return sparkRJobBuilder_.getMessageOrBuilder(); + } else { + if (jobTypeCase_ == 11) { + return (com.google.cloud.dataproc.v1.SparkRJob) jobType_; + } + return com.google.cloud.dataproc.v1.SparkRJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.SparkRJob, + com.google.cloud.dataproc.v1.SparkRJob.Builder, + com.google.cloud.dataproc.v1.SparkRJobOrBuilder> + getSparkRJobFieldBuilder() { + if (sparkRJobBuilder_ == null) { + if (!(jobTypeCase_ == 11)) { + jobType_ = com.google.cloud.dataproc.v1.SparkRJob.getDefaultInstance(); + } + sparkRJobBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.SparkRJob, + com.google.cloud.dataproc.v1.SparkRJob.Builder, + com.google.cloud.dataproc.v1.SparkRJobOrBuilder>( + (com.google.cloud.dataproc.v1.SparkRJob) jobType_, + getParentForChildren(), + isClean()); + jobType_ = null; + } + jobTypeCase_ = 11; + onChanged(); + ; + return sparkRJobBuilder_; + } + private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1.SparkSqlJob, com.google.cloud.dataproc.v1.SparkSqlJob.Builder, @@ -2369,6 +2762,211 @@ public com.google.cloud.dataproc.v1.SparkSqlJobOrBuilder getSparkSqlJobOrBuilder return sparkSqlJobBuilder_; } + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.PrestoJob, + com.google.cloud.dataproc.v1.PrestoJob.Builder, + com.google.cloud.dataproc.v1.PrestoJobOrBuilder> + prestoJobBuilder_; + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + * + * @return Whether the prestoJob field is set. + */ + public boolean hasPrestoJob() { + return jobTypeCase_ == 12; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + * + * @return The prestoJob. + */ + public com.google.cloud.dataproc.v1.PrestoJob getPrestoJob() { + if (prestoJobBuilder_ == null) { + if (jobTypeCase_ == 12) { + return (com.google.cloud.dataproc.v1.PrestoJob) jobType_; + } + return com.google.cloud.dataproc.v1.PrestoJob.getDefaultInstance(); + } else { + if (jobTypeCase_ == 12) { + return prestoJobBuilder_.getMessage(); + } + return com.google.cloud.dataproc.v1.PrestoJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + public Builder setPrestoJob(com.google.cloud.dataproc.v1.PrestoJob value) { + if (prestoJobBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + jobType_ = value; + onChanged(); + } else { + prestoJobBuilder_.setMessage(value); + } + jobTypeCase_ = 12; + return this; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + public Builder setPrestoJob(com.google.cloud.dataproc.v1.PrestoJob.Builder builderForValue) { + if (prestoJobBuilder_ == null) { + jobType_ = builderForValue.build(); + onChanged(); + } else { + prestoJobBuilder_.setMessage(builderForValue.build()); + } + jobTypeCase_ = 12; + return this; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + public Builder mergePrestoJob(com.google.cloud.dataproc.v1.PrestoJob value) { + if (prestoJobBuilder_ == null) { + if (jobTypeCase_ == 12 + && jobType_ != com.google.cloud.dataproc.v1.PrestoJob.getDefaultInstance()) { + jobType_ = + com.google.cloud.dataproc.v1.PrestoJob.newBuilder( + (com.google.cloud.dataproc.v1.PrestoJob) jobType_) + .mergeFrom(value) + .buildPartial(); + } else { + jobType_ = value; + } + onChanged(); + } else { + if (jobTypeCase_ == 12) { + prestoJobBuilder_.mergeFrom(value); + } + prestoJobBuilder_.setMessage(value); + } + jobTypeCase_ = 12; + return this; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + public Builder clearPrestoJob() { + if (prestoJobBuilder_ == null) { + if (jobTypeCase_ == 12) { + jobTypeCase_ = 0; + jobType_ = null; + onChanged(); + } + } else { + if (jobTypeCase_ == 12) { + jobTypeCase_ = 0; + jobType_ = null; + } + prestoJobBuilder_.clear(); + } + return this; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + public com.google.cloud.dataproc.v1.PrestoJob.Builder getPrestoJobBuilder() { + return getPrestoJobFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + public com.google.cloud.dataproc.v1.PrestoJobOrBuilder getPrestoJobOrBuilder() { + if ((jobTypeCase_ == 12) && (prestoJobBuilder_ != null)) { + return prestoJobBuilder_.getMessageOrBuilder(); + } else { + if (jobTypeCase_ == 12) { + return (com.google.cloud.dataproc.v1.PrestoJob) jobType_; + } + return com.google.cloud.dataproc.v1.PrestoJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.PrestoJob, + com.google.cloud.dataproc.v1.PrestoJob.Builder, + com.google.cloud.dataproc.v1.PrestoJobOrBuilder> + getPrestoJobFieldBuilder() { + if (prestoJobBuilder_ == null) { + if (!(jobTypeCase_ == 12)) { + jobType_ = com.google.cloud.dataproc.v1.PrestoJob.getDefaultInstance(); + } + prestoJobBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1.PrestoJob, + com.google.cloud.dataproc.v1.PrestoJob.Builder, + com.google.cloud.dataproc.v1.PrestoJobOrBuilder>( + (com.google.cloud.dataproc.v1.PrestoJob) jobType_, + getParentForChildren(), + isClean()); + jobType_ = null; + } + jobTypeCase_ = 12; + onChanged(); + ; + return prestoJobBuilder_; + } + private com.google.protobuf.MapField labels_; private com.google.protobuf.MapField internalGetLabels() { diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJobOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJobOrBuilder.java index 00b7799d..27230208 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/OrderedJobOrBuilder.java @@ -139,6 +139,41 @@ public interface OrderedJobOrBuilder /** .google.cloud.dataproc.v1.PigJob pig_job = 6; */ com.google.cloud.dataproc.v1.PigJobOrBuilder getPigJobOrBuilder(); + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + * + * @return Whether the sparkRJob field is set. + */ + boolean hasSparkRJob(); + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + * + * @return The sparkRJob. + */ + com.google.cloud.dataproc.v1.SparkRJob getSparkRJob(); + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1.SparkRJob spark_r_job = 11; + */ + com.google.cloud.dataproc.v1.SparkRJobOrBuilder getSparkRJobOrBuilder(); + /** * .google.cloud.dataproc.v1.SparkSqlJob spark_sql_job = 7; * @@ -154,6 +189,41 @@ public interface OrderedJobOrBuilder /** .google.cloud.dataproc.v1.SparkSqlJob spark_sql_job = 7; */ com.google.cloud.dataproc.v1.SparkSqlJobOrBuilder getSparkSqlJobOrBuilder(); + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + * + * @return Whether the prestoJob field is set. + */ + boolean hasPrestoJob(); + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + * + * @return The prestoJob. + */ + com.google.cloud.dataproc.v1.PrestoJob getPrestoJob(); + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1.PrestoJob presto_job = 12; + */ + com.google.cloud.dataproc.v1.PrestoJobOrBuilder getPrestoJobOrBuilder(); + /** * * diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PrestoJob.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PrestoJob.java index ef1ab155..77d23150 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PrestoJob.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/PrestoJob.java @@ -22,7 +22,10 @@ * * *
- * A Dataproc job for running [Presto](https://prestosql.io/) queries
+ * A Dataproc job for running [Presto](https://prestosql.io/) queries.
+ * **IMPORTANT**: The [Dataproc Presto Optional
+ * Component](/dataproc/docs/concepts/components/presto) must be enabled when
+ * the cluster is created to submit a Presto job to the cluster.
  * 
* * Protobuf type {@code google.cloud.dataproc.v1.PrestoJob} @@ -888,7 +891,10 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * A Dataproc job for running [Presto](https://prestosql.io/) queries
+   * A Dataproc job for running [Presto](https://prestosql.io/) queries.
+   * **IMPORTANT**: The [Dataproc Presto Optional
+   * Component](/dataproc/docs/concepts/components/presto) must be enabled when
+   * the cluster is created to submit a Presto job to the cluster.
    * 
* * Protobuf type {@code google.cloud.dataproc.v1.PrestoJob} diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SharedProto.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SharedProto.java index 1381cafc..923a4367 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SharedProto.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SharedProto.java @@ -37,12 +37,13 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { java.lang.String[] descriptorData = { "\n%google/cloud/dataproc/v1/shared.proto\022" + "\030google.cloud.dataproc.v1\032\034google/api/an" - + "notations.proto*a\n\tComponent\022\031\n\025COMPONEN" + + "notations.proto*|\n\tComponent\022\031\n\025COMPONEN" + "T_UNSPECIFIED\020\000\022\014\n\010ANACONDA\020\005\022\020\n\014HIVE_WE" - + "BHCAT\020\003\022\013\n\007JUPYTER\020\001\022\014\n\010ZEPPELIN\020\004Bo\n\034co" - + "m.google.cloud.dataproc.v1B\013SharedProtoP" - + "\001Z@google.golang.org/genproto/googleapis" - + "/cloud/dataproc/v1;dataprocb\006proto3" + + "BHCAT\020\003\022\013\n\007JUPYTER\020\001\022\n\n\006PRESTO\020\006\022\014\n\010ZEPP" + + "ELIN\020\004\022\r\n\tZOOKEEPER\020\010Bo\n\034com.google.clou" + + "d.dataproc.v1B\013SharedProtoP\001Z@google.gol" + + "ang.org/genproto/googleapis/cloud/datapr" + + "oc/v1;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java index 83999737..fa4495f5 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfig.java @@ -175,10 +175,10 @@ protected com.google.protobuf.MapField internalGetMapField(int number) { *
    * Optional. The version of software inside the cluster. It must be one of the
    * supported [Dataproc
-   * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+   * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
-   * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+   * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
    * If unspecified, it defaults to the latest Debian version.
    * 
* @@ -203,10 +203,10 @@ public java.lang.String getImageVersion() { *
    * Optional. The version of software inside the cluster. It must be one of the
    * supported [Dataproc
-   * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+   * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
-   * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+   * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
    * If unspecified, it defaults to the latest Debian version.
    * 
* @@ -268,8 +268,8 @@ public int getPropertiesCount() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -303,8 +303,8 @@ public java.util.Map getProperties() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -330,8 +330,8 @@ public java.util.Map getPropertiesMap() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -362,8 +362,8 @@ public java.lang.String getPropertiesOrDefault( * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -891,10 +891,10 @@ public Builder mergeFrom( *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -919,10 +919,10 @@ public java.lang.String getImageVersion() { *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -947,10 +947,10 @@ public com.google.protobuf.ByteString getImageVersionBytes() { *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -974,10 +974,10 @@ public Builder setImageVersion(java.lang.String value) { *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -997,10 +997,10 @@ public Builder clearImageVersion() { *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -1065,8 +1065,8 @@ public int getPropertiesCount() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1100,8 +1100,8 @@ public java.util.Map getProperties() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1127,8 +1127,8 @@ public java.util.Map getPropertiesMap() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1159,8 +1159,8 @@ public java.lang.String getPropertiesOrDefault( * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1198,8 +1198,8 @@ public Builder clearProperties() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1234,8 +1234,8 @@ public java.util.Map getMutableProperties() * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1268,8 +1268,8 @@ public Builder putProperties(java.lang.String key, java.lang.String value) { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java index 3f908a47..063de799 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/SoftwareConfigOrBuilder.java @@ -29,10 +29,10 @@ public interface SoftwareConfigOrBuilder *
    * Optional. The version of software inside the cluster. It must be one of the
    * supported [Dataproc
-   * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+   * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
-   * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+   * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
    * If unspecified, it defaults to the latest Debian version.
    * 
* @@ -47,10 +47,10 @@ public interface SoftwareConfigOrBuilder *
    * Optional. The version of software inside the cluster. It must be one of the
    * supported [Dataproc
-   * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+   * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
-   * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+   * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
    * If unspecified, it defaults to the latest Debian version.
    * 
* @@ -77,8 +77,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -102,8 +102,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -130,8 +130,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -155,8 +155,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -180,8 +180,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java index 8aceac67..df69232f 100644 --- a/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java +++ b/proto-google-cloud-dataproc-v1/src/main/java/com/google/cloud/dataproc/v1/WorkflowTemplatesProto.java @@ -186,7 +186,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "\002 \003(\0132<.google.cloud.dataproc.v1.Cluster" + "Selector.ClusterLabelsEntryB\003\340A\002\0324\n\022Clus" + "terLabelsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001" - + "(\t:\0028\001\"\347\004\n\nOrderedJob\022\024\n\007step_id\030\001 \001(\tB\003" + + "(\t:\0028\001\"\336\005\n\nOrderedJob\022\024\n\007step_id\030\001 \001(\tB\003" + "\340A\002\0229\n\nhadoop_job\030\002 \001(\0132#.google.cloud.d" + "ataproc.v1.HadoopJobH\000\0227\n\tspark_job\030\003 \001(" + "\0132\".google.cloud.dataproc.v1.SparkJobH\000\022" @@ -194,142 +194,145 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "proc.v1.PySparkJobH\000\0225\n\010hive_job\030\005 \001(\0132!" + ".google.cloud.dataproc.v1.HiveJobH\000\0223\n\007p" + "ig_job\030\006 \001(\0132 .google.cloud.dataproc.v1." - + "PigJobH\000\022>\n\rspark_sql_job\030\007 \001(\0132%.google" - + ".cloud.dataproc.v1.SparkSqlJobH\000\022E\n\006labe" - + "ls\030\010 \003(\01320.google.cloud.dataproc.v1.Orde" - + "redJob.LabelsEntryB\003\340A\001\022@\n\nscheduling\030\t " - + "\001(\0132\'.google.cloud.dataproc.v1.JobSchedu" - + "lingB\003\340A\001\022\"\n\025prerequisite_step_ids\030\n \003(\t" - + "B\003\340A\001\032-\n\013LabelsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005val" - + "ue\030\002 \001(\t:\0028\001B\n\n\010job_type\"\235\001\n\021TemplatePar" - + "ameter\022\021\n\004name\030\001 \001(\tB\003\340A\002\022\023\n\006fields\030\002 \003(" - + "\tB\003\340A\002\022\030\n\013description\030\003 \001(\tB\003\340A\001\022F\n\nvali" - + "dation\030\004 \001(\0132-.google.cloud.dataproc.v1." - + "ParameterValidationB\003\340A\001\"\241\001\n\023ParameterVa" - + "lidation\022:\n\005regex\030\001 \001(\0132).google.cloud.d" - + "ataproc.v1.RegexValidationH\000\022;\n\006values\030\002" - + " \001(\0132).google.cloud.dataproc.v1.ValueVal" - + "idationH\000B\021\n\017validation_type\"\'\n\017RegexVal" - + "idation\022\024\n\007regexes\030\001 \003(\tB\003\340A\002\"&\n\017ValueVa" - + "lidation\022\023\n\006values\030\001 \003(\tB\003\340A\002\"\257\005\n\020Workfl" - + "owMetadata\022\025\n\010template\030\001 \001(\tB\003\340A\003\022\024\n\007ver" - + "sion\030\002 \001(\005B\003\340A\003\022G\n\016create_cluster\030\003 \001(\0132" - + "*.google.cloud.dataproc.v1.ClusterOperat" - + "ionB\003\340A\003\022;\n\005graph\030\004 \001(\0132\'.google.cloud.d" - + "ataproc.v1.WorkflowGraphB\003\340A\003\022G\n\016delete_" - + "cluster\030\005 \001(\0132*.google.cloud.dataproc.v1" - + ".ClusterOperationB\003\340A\003\022D\n\005state\030\006 \001(\01620." - + "google.cloud.dataproc.v1.WorkflowMetadat" - + "a.StateB\003\340A\003\022\031\n\014cluster_name\030\007 \001(\tB\003\340A\003\022" - + "N\n\nparameters\030\010 \003(\0132:.google.cloud.datap" - + "roc.v1.WorkflowMetadata.ParametersEntry\022" - + "3\n\nstart_time\030\t \001(\0132\032.google.protobuf.Ti" - + "mestampB\003\340A\003\0221\n\010end_time\030\n \001(\0132\032.google." - + "protobuf.TimestampB\003\340A\003\022\031\n\014cluster_uuid\030" - + "\013 \001(\tB\003\340A\003\0321\n\017ParametersEntry\022\013\n\003key\030\001 \001" - + "(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"8\n\005State\022\013\n\007UNKNOW" - + "N\020\000\022\013\n\007PENDING\020\001\022\013\n\007RUNNING\020\002\022\010\n\004DONE\020\003\"" - + "T\n\020ClusterOperation\022\031\n\014operation_id\030\001 \001(" - + "\tB\003\340A\003\022\022\n\005error\030\002 \001(\tB\003\340A\003\022\021\n\004done\030\003 \001(\010" - + "B\003\340A\003\"K\n\rWorkflowGraph\022:\n\005nodes\030\001 \003(\0132&." - + "google.cloud.dataproc.v1.WorkflowNodeB\003\340" - + "A\003\"\243\002\n\014WorkflowNode\022\024\n\007step_id\030\001 \001(\tB\003\340A" - + "\003\022\"\n\025prerequisite_step_ids\030\002 \003(\tB\003\340A\003\022\023\n" - + "\006job_id\030\003 \001(\tB\003\340A\003\022D\n\005state\030\005 \001(\01620.goog" - + "le.cloud.dataproc.v1.WorkflowNode.NodeSt" - + "ateB\003\340A\003\022\022\n\005error\030\006 \001(\tB\003\340A\003\"j\n\tNodeStat" - + "e\022\032\n\026NODE_STATE_UNSPECIFIED\020\000\022\013\n\007BLOCKED" - + "\020\001\022\014\n\010RUNNABLE\020\002\022\013\n\007RUNNING\020\003\022\r\n\tCOMPLET" - + "ED\020\004\022\n\n\006FAILED\020\005\"\244\001\n\035CreateWorkflowTempl" - + "ateRequest\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(data" - + "proc.googleapis.com/WorkflowTemplate\022A\n\010" - + "template\030\002 \001(\0132*.google.cloud.dataproc.v" - + "1.WorkflowTemplateB\003\340A\002\"r\n\032GetWorkflowTe" - + "mplateRequest\022>\n\004name\030\001 \001(\tB0\340A\002\372A*\n(dat" - + "aproc.googleapis.com/WorkflowTemplate\022\024\n" - + "\007version\030\002 \001(\005B\003\340A\001\"\255\002\n\"InstantiateWorkf" + + "PigJobH\000\022:\n\013spark_r_job\030\013 \001(\0132#.google.c" + + "loud.dataproc.v1.SparkRJobH\000\022>\n\rspark_sq" + + "l_job\030\007 \001(\0132%.google.cloud.dataproc.v1.S" + + "parkSqlJobH\000\0229\n\npresto_job\030\014 \001(\0132#.googl" + + "e.cloud.dataproc.v1.PrestoJobH\000\022E\n\006label" + + "s\030\010 \003(\01320.google.cloud.dataproc.v1.Order" + + "edJob.LabelsEntryB\003\340A\001\022@\n\nscheduling\030\t \001" + + "(\0132\'.google.cloud.dataproc.v1.JobSchedul" + + "ingB\003\340A\001\022\"\n\025prerequisite_step_ids\030\n \003(\tB" + + "\003\340A\001\032-\n\013LabelsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005valu" + + "e\030\002 \001(\t:\0028\001B\n\n\010job_type\"\235\001\n\021TemplatePara" + + "meter\022\021\n\004name\030\001 \001(\tB\003\340A\002\022\023\n\006fields\030\002 \003(\t" + + "B\003\340A\002\022\030\n\013description\030\003 \001(\tB\003\340A\001\022F\n\nvalid" + + "ation\030\004 \001(\0132-.google.cloud.dataproc.v1.P" + + "arameterValidationB\003\340A\001\"\241\001\n\023ParameterVal" + + "idation\022:\n\005regex\030\001 \001(\0132).google.cloud.da" + + "taproc.v1.RegexValidationH\000\022;\n\006values\030\002 " + + "\001(\0132).google.cloud.dataproc.v1.ValueVali" + + "dationH\000B\021\n\017validation_type\"\'\n\017RegexVali" + + "dation\022\024\n\007regexes\030\001 \003(\tB\003\340A\002\"&\n\017ValueVal" + + "idation\022\023\n\006values\030\001 \003(\tB\003\340A\002\"\257\005\n\020Workflo" + + "wMetadata\022\025\n\010template\030\001 \001(\tB\003\340A\003\022\024\n\007vers" + + "ion\030\002 \001(\005B\003\340A\003\022G\n\016create_cluster\030\003 \001(\0132*" + + ".google.cloud.dataproc.v1.ClusterOperati" + + "onB\003\340A\003\022;\n\005graph\030\004 \001(\0132\'.google.cloud.da" + + "taproc.v1.WorkflowGraphB\003\340A\003\022G\n\016delete_c" + + "luster\030\005 \001(\0132*.google.cloud.dataproc.v1." + + "ClusterOperationB\003\340A\003\022D\n\005state\030\006 \001(\01620.g" + + "oogle.cloud.dataproc.v1.WorkflowMetadata" + + ".StateB\003\340A\003\022\031\n\014cluster_name\030\007 \001(\tB\003\340A\003\022N" + + "\n\nparameters\030\010 \003(\0132:.google.cloud.datapr" + + "oc.v1.WorkflowMetadata.ParametersEntry\0223" + + "\n\nstart_time\030\t \001(\0132\032.google.protobuf.Tim" + + "estampB\003\340A\003\0221\n\010end_time\030\n \001(\0132\032.google.p" + + "rotobuf.TimestampB\003\340A\003\022\031\n\014cluster_uuid\030\013" + + " \001(\tB\003\340A\003\0321\n\017ParametersEntry\022\013\n\003key\030\001 \001(" + + "\t\022\r\n\005value\030\002 \001(\t:\0028\001\"8\n\005State\022\013\n\007UNKNOWN" + + "\020\000\022\013\n\007PENDING\020\001\022\013\n\007RUNNING\020\002\022\010\n\004DONE\020\003\"T" + + "\n\020ClusterOperation\022\031\n\014operation_id\030\001 \001(\t" + + "B\003\340A\003\022\022\n\005error\030\002 \001(\tB\003\340A\003\022\021\n\004done\030\003 \001(\010B" + + "\003\340A\003\"K\n\rWorkflowGraph\022:\n\005nodes\030\001 \003(\0132&.g" + + "oogle.cloud.dataproc.v1.WorkflowNodeB\003\340A" + + "\003\"\243\002\n\014WorkflowNode\022\024\n\007step_id\030\001 \001(\tB\003\340A\003" + + "\022\"\n\025prerequisite_step_ids\030\002 \003(\tB\003\340A\003\022\023\n\006" + + "job_id\030\003 \001(\tB\003\340A\003\022D\n\005state\030\005 \001(\01620.googl" + + "e.cloud.dataproc.v1.WorkflowNode.NodeSta" + + "teB\003\340A\003\022\022\n\005error\030\006 \001(\tB\003\340A\003\"j\n\tNodeState" + + "\022\032\n\026NODE_STATE_UNSPECIFIED\020\000\022\013\n\007BLOCKED\020" + + "\001\022\014\n\010RUNNABLE\020\002\022\013\n\007RUNNING\020\003\022\r\n\tCOMPLETE" + + "D\020\004\022\n\n\006FAILED\020\005\"\244\001\n\035CreateWorkflowTempla" + + "teRequest\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(datap" + + "roc.googleapis.com/WorkflowTemplate\022A\n\010t" + + "emplate\030\002 \001(\0132*.google.cloud.dataproc.v1" + + ".WorkflowTemplateB\003\340A\002\"r\n\032GetWorkflowTem" + + "plateRequest\022>\n\004name\030\001 \001(\tB0\340A\002\372A*\n(data" + + "proc.googleapis.com/WorkflowTemplate\022\024\n\007" + + "version\030\002 \001(\005B\003\340A\001\"\255\002\n\"InstantiateWorkfl" + + "owTemplateRequest\022>\n\004name\030\001 \001(\tB0\340A\002\372A*\n" + + "(dataproc.googleapis.com/WorkflowTemplat" + + "e\022\024\n\007version\030\002 \001(\005B\003\340A\001\022\027\n\nrequest_id\030\005 " + + "\001(\tB\003\340A\001\022e\n\nparameters\030\006 \003(\0132L.google.cl" + + "oud.dataproc.v1.InstantiateWorkflowTempl" + + "ateRequest.ParametersEntryB\003\340A\001\0321\n\017Param" + + "etersEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\002" + + "8\001\"\310\001\n(InstantiateInlineWorkflowTemplate" + + "Request\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(datapro" + + "c.googleapis.com/WorkflowTemplate\022A\n\010tem" + + "plate\030\002 \001(\0132*.google.cloud.dataproc.v1.W" + + "orkflowTemplateB\003\340A\002\022\027\n\nrequest_id\030\003 \001(\t" + + "B\003\340A\001\"b\n\035UpdateWorkflowTemplateRequest\022A" + + "\n\010template\030\001 \001(\0132*.google.cloud.dataproc" + + ".v1.WorkflowTemplateB\003\340A\002\"\221\001\n\034ListWorkfl" + + "owTemplatesRequest\022@\n\006parent\030\001 \001(\tB0\340A\002\372" + + "A*\022(dataproc.googleapis.com/WorkflowTemp" + + "late\022\026\n\tpage_size\030\002 \001(\005B\003\340A\001\022\027\n\npage_tok" + + "en\030\003 \001(\tB\003\340A\001\"\201\001\n\035ListWorkflowTemplatesR" + + "esponse\022B\n\ttemplates\030\001 \003(\0132*.google.clou" + + "d.dataproc.v1.WorkflowTemplateB\003\340A\003\022\034\n\017n" + + "ext_page_token\030\002 \001(\tB\003\340A\003\"u\n\035DeleteWorkf" + "lowTemplateRequest\022>\n\004name\030\001 \001(\tB0\340A\002\372A*" + "\n(dataproc.googleapis.com/WorkflowTempla" - + "te\022\024\n\007version\030\002 \001(\005B\003\340A\001\022\027\n\nrequest_id\030\005" - + " \001(\tB\003\340A\001\022e\n\nparameters\030\006 \003(\0132L.google.c" - + "loud.dataproc.v1.InstantiateWorkflowTemp" - + "lateRequest.ParametersEntryB\003\340A\001\0321\n\017Para" - + "metersEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:" - + "\0028\001\"\310\001\n(InstantiateInlineWorkflowTemplat" - + "eRequest\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(datapr" - + "oc.googleapis.com/WorkflowTemplate\022A\n\010te" - + "mplate\030\002 \001(\0132*.google.cloud.dataproc.v1." - + "WorkflowTemplateB\003\340A\002\022\027\n\nrequest_id\030\003 \001(" - + "\tB\003\340A\001\"b\n\035UpdateWorkflowTemplateRequest\022" - + "A\n\010template\030\001 \001(\0132*.google.cloud.datapro" - + "c.v1.WorkflowTemplateB\003\340A\002\"\221\001\n\034ListWorkf" - + "lowTemplatesRequest\022@\n\006parent\030\001 \001(\tB0\340A\002" - + "\372A*\022(dataproc.googleapis.com/WorkflowTem" - + "plate\022\026\n\tpage_size\030\002 \001(\005B\003\340A\001\022\027\n\npage_to" - + "ken\030\003 \001(\tB\003\340A\001\"\201\001\n\035ListWorkflowTemplates" - + "Response\022B\n\ttemplates\030\001 \003(\0132*.google.clo" - + "ud.dataproc.v1.WorkflowTemplateB\003\340A\003\022\034\n\017" - + "next_page_token\030\002 \001(\tB\003\340A\003\"u\n\035DeleteWork" - + "flowTemplateRequest\022>\n\004name\030\001 \001(\tB0\340A\002\372A" - + "*\n(dataproc.googleapis.com/WorkflowTempl" - + "ate\022\024\n\007version\030\002 \001(\005B\003\340A\0012\346\020\n\027WorkflowTe" - + "mplateService\022\233\002\n\026CreateWorkflowTemplate" - + "\0227.google.cloud.dataproc.v1.CreateWorkfl" - + "owTemplateRequest\032*.google.cloud.datapro" - + "c.v1.WorkflowTemplate\"\233\001\202\323\344\223\002\202\001\"5/v1/{pa" - + "rent=projects/*/locations/*}/workflowTem" - + "plates:\010templateZ?\"3/v1/{parent=projects" - + "/*/regions/*}/workflowTemplates:\010templat" - + "e\332A\017parent,template\022\364\001\n\023GetWorkflowTempl" - + "ate\0224.google.cloud.dataproc.v1.GetWorkfl" - + "owTemplateRequest\032*.google.cloud.datapro" - + "c.v1.WorkflowTemplate\"{\202\323\344\223\002n\0225/v1/{name" - + "=projects/*/locations/*/workflowTemplate" - + "s/*}Z5\0223/v1/{name=projects/*/regions/*/w" - + "orkflowTemplates/*}\332A\004name\022\325\002\n\033Instantia" - + "teWorkflowTemplate\022<.google.cloud.datapr" - + "oc.v1.InstantiateWorkflowTemplateRequest" - + "\032\035.google.longrunning.Operation\"\330\001\202\323\344\223\002\214" - + "\001\"A/v1/{name=projects/*/locations/*/work" - + "flowTemplates/*}:instantiate:\001*ZD\"?/v1/{" - + "name=projects/*/regions/*/workflowTempla" - + "tes/*}:instantiate:\001*\332A\004name\332A\017name,para" - + "meters\312A)\n\025google.protobuf.Empty\022\020Workfl" - + "owMetadata\022\364\002\n!InstantiateInlineWorkflow" - + "Template\022B.google.cloud.dataproc.v1.Inst" - + "antiateInlineWorkflowTemplateRequest\032\035.g" - + "oogle.longrunning.Operation\"\353\001\202\323\344\223\002\246\001\"G/" - + "v1/{parent=projects/*/locations/*}/workf" - + "lowTemplates:instantiateInline:\010template" - + "ZQ\"E/v1/{parent=projects/*/regions/*}/wo" - + "rkflowTemplates:instantiateInline:\010templ" - + "ate\332A\017parent,template\312A)\n\025google.protobu" - + "f.Empty\022\020WorkflowMetadata\022\246\002\n\026UpdateWork" - + "flowTemplate\0227.google.cloud.dataproc.v1." - + "UpdateWorkflowTemplateRequest\032*.google.c" - + "loud.dataproc.v1.WorkflowTemplate\"\246\001\202\323\344\223" - + "\002\224\001\032>/v1/{template.name=projects/*/locat" - + "ions/*/workflowTemplates/*}:\010templateZH\032" - + "/v1/{template.name=projects/*/locati" + + "ons/*/workflowTemplates/*}:\010templateZH\032<" + + "/v1/{template.name=projects/*/regions/*/" + + "workflowTemplates/*}:\010template\332A\010templat" + + "e\022\207\002\n\025ListWorkflowTemplates\0226.google.clo" + + "ud.dataproc.v1.ListWorkflowTemplatesRequ" + + "est\0327.google.cloud.dataproc.v1.ListWorkf" + + "lowTemplatesResponse\"}\202\323\344\223\002n\0225/v1/{paren" + + "t=projects/*/locations/*}/workflowTempla" + + "tesZ5\0223/v1/{parent=projects/*/regions/*}" + + "/workflowTemplates\332A\006parent\022\346\001\n\026DeleteWo" + + "rkflowTemplate\0227.google.cloud.dataproc.v" + + "1.DeleteWorkflowTemplateRequest\032\026.google" + + ".protobuf.Empty\"{\202\323\344\223\002n*5/v1/{name=proje" + + "cts/*/locations/*/workflowTemplates/*}Z5" + + "*3/v1/{name=projects/*/regions/*/workflo" + + "wTemplates/*}\332A\004name\032K\312A\027dataproc.google" + + "apis.com\322A.https://www.googleapis.com/au" + + "th/cloud-platformBz\n\034com.google.cloud.da" + + "taproc.v1B\026WorkflowTemplatesProtoP\001Z@goo" + + "gle.golang.org/genproto/googleapis/cloud" + + "/dataproc/v1;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( @@ -423,7 +426,9 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "PysparkJob", "HiveJob", "PigJob", + "SparkRJob", "SparkSqlJob", + "PrestoJob", "Labels", "Scheduling", "PrerequisiteStepIds", diff --git a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/autoscaling_policies.proto b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/autoscaling_policies.proto index 8d680bcd..4ce5868d 100644 --- a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/autoscaling_policies.proto +++ b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/autoscaling_policies.proto @@ -36,12 +36,10 @@ option (google.api.resource_definition) = { // Dataproc API. service AutoscalingPolicyService { option (google.api.default_host) = "dataproc.googleapis.com"; - option (google.api.oauth_scopes) = - "https://www.googleapis.com/auth/cloud-platform"; + option (google.api.oauth_scopes) = "https://www.googleapis.com/auth/cloud-platform"; // Creates new autoscaling policy. - rpc CreateAutoscalingPolicy(CreateAutoscalingPolicyRequest) - returns (AutoscalingPolicy) { + rpc CreateAutoscalingPolicy(CreateAutoscalingPolicyRequest) returns (AutoscalingPolicy) { option (google.api.http) = { post: "/v1/{parent=projects/*/locations/*}/autoscalingPolicies" body: "policy" @@ -57,8 +55,7 @@ service AutoscalingPolicyService { // // Disabled check for update_mask, because all updates will be full // replacements. - rpc UpdateAutoscalingPolicy(UpdateAutoscalingPolicyRequest) - returns (AutoscalingPolicy) { + rpc UpdateAutoscalingPolicy(UpdateAutoscalingPolicyRequest) returns (AutoscalingPolicy) { option (google.api.http) = { put: "/v1/{policy.name=projects/*/locations/*/autoscalingPolicies/*}" body: "policy" @@ -71,8 +68,7 @@ service AutoscalingPolicyService { } // Retrieves autoscaling policy. - rpc GetAutoscalingPolicy(GetAutoscalingPolicyRequest) - returns (AutoscalingPolicy) { + rpc GetAutoscalingPolicy(GetAutoscalingPolicyRequest) returns (AutoscalingPolicy) { option (google.api.http) = { get: "/v1/{name=projects/*/locations/*/autoscalingPolicies/*}" additional_bindings { @@ -83,8 +79,7 @@ service AutoscalingPolicyService { } // Lists autoscaling policies in the project. - rpc ListAutoscalingPolicies(ListAutoscalingPoliciesRequest) - returns (ListAutoscalingPoliciesResponse) { + rpc ListAutoscalingPolicies(ListAutoscalingPoliciesRequest) returns (ListAutoscalingPoliciesResponse) { option (google.api.http) = { get: "/v1/{parent=projects/*/locations/*}/autoscalingPolicies" additional_bindings { @@ -96,8 +91,7 @@ service AutoscalingPolicyService { // Deletes an autoscaling policy. It is an error to delete an autoscaling // policy that is in use by one or more clusters. - rpc DeleteAutoscalingPolicy(DeleteAutoscalingPolicyRequest) - returns (google.protobuf.Empty) { + rpc DeleteAutoscalingPolicy(DeleteAutoscalingPolicyRequest) returns (google.protobuf.Empty) { option (google.api.http) = { delete: "/v1/{name=projects/*/locations/*/autoscalingPolicies/*}" additional_bindings { @@ -138,31 +132,26 @@ message AutoscalingPolicy { // Autoscaling algorithm for policy. oneof algorithm { - BasicAutoscalingAlgorithm basic_algorithm = 3 - [(google.api.field_behavior) = REQUIRED]; + BasicAutoscalingAlgorithm basic_algorithm = 3 [(google.api.field_behavior) = REQUIRED]; } // Required. Describes how the autoscaler will operate for primary workers. - InstanceGroupAutoscalingPolicyConfig worker_config = 4 - [(google.api.field_behavior) = REQUIRED]; + InstanceGroupAutoscalingPolicyConfig worker_config = 4 [(google.api.field_behavior) = REQUIRED]; // Optional. Describes how the autoscaler will operate for secondary workers. - InstanceGroupAutoscalingPolicyConfig secondary_worker_config = 5 - [(google.api.field_behavior) = OPTIONAL]; + InstanceGroupAutoscalingPolicyConfig secondary_worker_config = 5 [(google.api.field_behavior) = OPTIONAL]; } // Basic algorithm for autoscaling. message BasicAutoscalingAlgorithm { // Required. YARN autoscaling configuration. - BasicYarnAutoscalingConfig yarn_config = 1 - [(google.api.field_behavior) = REQUIRED]; + BasicYarnAutoscalingConfig yarn_config = 1 [(google.api.field_behavior) = REQUIRED]; // Optional. Duration between scaling events. A scaling period starts after // the update operation from the previous event has completed. // // Bounds: [2m, 1d]. Default: 2m. - google.protobuf.Duration cooldown_period = 2 - [(google.api.field_behavior) = OPTIONAL]; + google.protobuf.Duration cooldown_period = 2 [(google.api.field_behavior) = OPTIONAL]; } // Basic autoscaling configurations for YARN. @@ -173,8 +162,7 @@ message BasicYarnAutoscalingConfig { // downscaling operations. // // Bounds: [0s, 1d]. - google.protobuf.Duration graceful_decommission_timeout = 5 - [(google.api.field_behavior) = REQUIRED]; + google.protobuf.Duration graceful_decommission_timeout = 5 [(google.api.field_behavior) = REQUIRED]; // Required. Fraction of average pending memory in the last cooldown period // for which to add workers. A scale-up factor of 1.0 will result in scaling @@ -201,8 +189,7 @@ message BasicYarnAutoscalingConfig { // on any recommended change. // // Bounds: [0.0, 1.0]. Default: 0.0. - double scale_up_min_worker_fraction = 3 - [(google.api.field_behavior) = OPTIONAL]; + double scale_up_min_worker_fraction = 3 [(google.api.field_behavior) = OPTIONAL]; // Optional. Minimum scale-down threshold as a fraction of total cluster size // before scaling occurs. For example, in a 20-worker cluster, a threshold of @@ -211,8 +198,7 @@ message BasicYarnAutoscalingConfig { // on any recommended change. // // Bounds: [0.0, 1.0]. Default: 0.0. - double scale_down_min_worker_fraction = 4 - [(google.api.field_behavior) = OPTIONAL]; + double scale_down_min_worker_fraction = 4 [(google.api.field_behavior) = OPTIONAL]; } // Configuration for the size bounds of an instance group, including its @@ -272,8 +258,8 @@ message CreateAutoscalingPolicyRequest { } ]; - // The autoscaling policy to create. - AutoscalingPolicy policy = 2; + // Required. The autoscaling policy to create. + AutoscalingPolicy policy = 2 [(google.api.field_behavior) = REQUIRED]; } // A request to fetch an autoscaling policy. @@ -355,8 +341,7 @@ message ListAutoscalingPoliciesRequest { // A response to a request to list autoscaling policies in a project. message ListAutoscalingPoliciesResponse { // Output only. Autoscaling policies list. - repeated AutoscalingPolicy policies = 1 - [(google.api.field_behavior) = OUTPUT_ONLY]; + repeated AutoscalingPolicy policies = 1 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. This token is included in the response if there are more // results to fetch. diff --git a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/clusters.proto b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/clusters.proto index a20b8f31..fbaf9391 100644 --- a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/clusters.proto +++ b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/clusters.proto @@ -88,7 +88,7 @@ service ClusterController { option (google.api.method_signature) = "project_id,region,cluster_name"; } - // Lists all regions/{region}/clusters in a project. + // Lists all regions/{region}/clusters in a project alphabetically. rpc ListClusters(ListClustersRequest) returns (ListClustersResponse) { option (google.api.http) = { get: "/v1/projects/{project_id}/regions/{region}/clusters" @@ -150,11 +150,11 @@ message Cluster { // generates this value when it creates the cluster. string cluster_uuid = 6 [(google.api.field_behavior) = OUTPUT_ONLY]; - // Contains cluster daemon metrics such as HDFS and YARN stats. + // Output only. Contains cluster daemon metrics such as HDFS and YARN stats. // // **Beta Feature**: This report is available for testing purposes only. It // may be changed before final release. - ClusterMetrics metrics = 9; + ClusterMetrics metrics = 9 [(google.api.field_behavior) = OUTPUT_ONLY]; } // The cluster config. @@ -167,7 +167,7 @@ message ClusterConfig { // Compute Engine zone where your cluster is deployed, and then create // and manage this project-level, per-location bucket (see // [Dataproc staging - // bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + // bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). string config_bucket = 1 [(google.api.field_behavior) = OPTIONAL]; // Optional. The shared Compute Engine config settings for @@ -259,7 +259,8 @@ message GceClusterConfig { // communications. Cannot be specified with subnetwork_uri. If neither // `network_uri` nor `subnetwork_uri` is specified, the "default" network of // the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - // [Using Subnetworks](/compute/docs/subnetworks) for more information). + // [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + // more information). // // A full URL, partial URI, or short name are valid. Examples: // @@ -287,15 +288,15 @@ message GceClusterConfig { bool internal_ip_only = 7 [(google.api.field_behavior) = OPTIONAL]; // Optional. The [Dataproc service - // account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc) + // account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc) // (also see [VM Data Plane - // identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity)) + // identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity)) // used by Dataproc cluster VM instances to access Google Cloud Platform // services. // // If not specified, the // [Compute Engine default service - // account](/compute/docs/access/service-accounts#default_service_account) + // account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) // is used. string service_account = 8 [(google.api.field_behavior) = OPTIONAL]; @@ -315,8 +316,8 @@ message GceClusterConfig { // * https://www.googleapis.com/auth/devstorage.full_control repeated string service_account_scopes = 3 [(google.api.field_behavior) = OPTIONAL]; - // The Compute Engine tags to add to all instances (see - // [Tagging instances](/compute/docs/label-or-tag-resources#tags)). + // The Compute Engine tags to add to all instances (see [Tagging + // instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)). repeated string tags = 4; // The Compute Engine metadata entries to add to all instances (see @@ -339,9 +340,24 @@ message InstanceGroupConfig { // from `cluster_name`, `num_instances`, and the instance group. repeated string instance_names = 2 [(google.api.field_behavior) = OUTPUT_ONLY]; - // Optional. The Compute Engine image resource used for cluster - // instances. It can be specified or may be inferred from - // `SoftwareConfig.image_version`. + // Optional. The Compute Engine image resource used for cluster instances. + // + // The URI can represent an image or image family. + // + // Image examples: + // + // * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]` + // * `projects/[project_id]/global/images/[image-id]` + // * `image-id` + // + // Image family examples. Dataproc will use the most recent + // image from the family: + // + // * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]` + // * `projects/[project_id]/global/images/family/[custom-image-family-name]` + // + // If the URI is unspecified, it will be inferred from + // `SoftwareConfig.image_version` or the system default. string image_uri = 3 [(google.api.field_behavior) = OPTIONAL]; // Optional. The Compute Engine machine type used for cluster instances. @@ -354,7 +370,7 @@ message InstanceGroupConfig { // // **Auto Zone Exception**: If you are using the Dataproc // [Auto Zone - // Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + // Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) // feature, you must use the short name of the machine type // resource, for example, `n1-standard-2`. string machine_type_uri = 4 [(google.api.field_behavior) = OPTIONAL]; @@ -362,9 +378,9 @@ message InstanceGroupConfig { // Optional. Disk option config settings. DiskConfig disk_config = 5 [(google.api.field_behavior) = OPTIONAL]; - // Optional. Specifies that this instance group contains preemptible + // Output only. Specifies that this instance group contains preemptible // instances. - bool is_preemptible = 6 [(google.api.field_behavior) = OPTIONAL]; + bool is_preemptible = 6 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The config for Compute Engine Instance Group // Manager that manages this group. @@ -376,8 +392,8 @@ message InstanceGroupConfig { repeated AcceleratorConfig accelerators = 8 [(google.api.field_behavior) = OPTIONAL]; // Optional. Specifies the minimum cpu platform for the Instance Group. - // See [Dataproc→Minimum CPU Platform] - // (/dataproc/docs/concepts/compute/dataproc-min-cpu). + // See [Dataproc -> Minimum CPU + // Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu). string min_cpu_platform = 9 [(google.api.field_behavior) = OPTIONAL]; } @@ -392,12 +408,13 @@ message ManagedGroupConfig { } // Specifies the type and number of accelerator cards attached to the instances -// of an instance. See [GPUs on Compute Engine](/compute/docs/gpus/). +// of an instance. See [GPUs on Compute +// Engine](https://cloud.google.com/compute/docs/gpus/). message AcceleratorConfig { // Full URL, partial URI, or short name of the accelerator type resource to // expose to this instance. See // [Compute Engine - // AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes). + // AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes). // // Examples: // @@ -407,7 +424,7 @@ message AcceleratorConfig { // // **Auto Zone Exception**: If you are using the Dataproc // [Auto Zone - // Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + // Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) // feature, you must use the short name of the accelerator type // resource, for example, `nvidia-tesla-k80`. string accelerator_type_uri = 1; @@ -519,7 +536,8 @@ message SecurityConfig { // Specifies Kerberos related configuration. message KerberosConfig { - // Optional. Flag to indicate whether to Kerberize the cluster. + // Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set + // this field to true to enable Kerberos on a cluster. bool enable_kerberos = 1 [(google.api.field_behavior) = OPTIONAL]; // Required. The Cloud Storage URI of a KMS encrypted file containing the root @@ -590,10 +608,10 @@ message KerberosConfig { message SoftwareConfig { // Optional. The version of software inside the cluster. It must be one of the // supported [Dataproc - // Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions), + // Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions), // such as "1.2" (including a subminor version, such as "1.2.29"), or the // ["preview" - // version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). + // version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). // If unspecified, it defaults to the latest Debian version. string image_version = 1 [(google.api.field_behavior) = OPTIONAL]; @@ -613,8 +631,8 @@ message SoftwareConfig { // * spark: `spark-defaults.conf` // * yarn: `yarn-site.xml` // - // For more information, see - // [Cluster properties](/dataproc/docs/concepts/cluster-properties). + // For more information, see [Cluster + // properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). map properties = 2 [(google.api.field_behavior) = OPTIONAL]; // Optional. The set of components to activate on the cluster. diff --git a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/jobs.proto b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/jobs.proto index 85921dc4..0379c09d 100644 --- a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/jobs.proto +++ b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/jobs.proto @@ -43,6 +43,19 @@ service JobController { option (google.api.method_signature) = "project_id,region,job"; } + // Submits job to a cluster. + rpc SubmitJobAsOperation(SubmitJobRequest) returns (google.longrunning.Operation) { + option (google.api.http) = { + post: "/v1/projects/{project_id}/regions/{region}/jobs:submitAsOperation" + body: "*" + }; + option (google.api.method_signature) = "project_id, region, job"; + option (google.longrunning.operation_info) = { + response_type: "Job" + metadata_type: "JobMetadata" + }; + } + // Gets the resource representation for a job in a project. rpc GetJob(GetJobRequest) returns (Job) { option (google.api.http) = { @@ -420,7 +433,10 @@ message SparkRJob { LoggingConfig logging_config = 6 [(google.api.field_behavior) = OPTIONAL]; } -// A Dataproc job for running [Presto](https://prestosql.io/) queries +// A Dataproc job for running [Presto](https://prestosql.io/) queries. +// **IMPORTANT**: The [Dataproc Presto Optional +// Component](/dataproc/docs/concepts/components/presto) must be enabled when +// the cluster is created to submit a Presto job to the cluster. message PrestoJob { // Required. The sequence of Presto queries to execute, specified as // either an HCFS file URI or as a list of queries. @@ -692,6 +708,12 @@ message Job { // over time. This is in contrast to a user-settable reference.job_id that // may be reused over time. string job_uuid = 22 [(google.api.field_behavior) = OUTPUT_ONLY]; + + // Output only. Indicates whether the job is completed. If the value is `false`, + // the job is still in progress. If `true`, the job is completed, and + // `status.state` field will indicate if it was successful, failed, + // or cancelled. + bool done = 24 [(google.api.field_behavior) = OUTPUT_ONLY]; } // Job scheduling options. @@ -733,6 +755,21 @@ message SubmitJobRequest { string request_id = 4 [(google.api.field_behavior) = OPTIONAL]; } +// Job Operation metadata. +message JobMetadata { + // Output only. The job id. + string job_id = 1 [(google.api.field_behavior) = OUTPUT_ONLY]; + + // Output only. Most recent job status. + JobStatus status = 2 [(google.api.field_behavior) = OUTPUT_ONLY]; + + // Output only. Operation type. + string operation_type = 3 [(google.api.field_behavior) = OUTPUT_ONLY]; + + // Output only. Job submission time. + google.protobuf.Timestamp start_time = 4 [(google.api.field_behavior) = OUTPUT_ONLY]; +} + // A request to get the resource representation for a job in a project. message GetJobRequest { // Required. The ID of the Google Cloud Platform project that the job diff --git a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/shared.proto b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/shared.proto index c6ff8f28..99d6e776 100644 --- a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/shared.proto +++ b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/shared.proto @@ -37,6 +37,12 @@ enum Component { // The Jupyter Notebook. JUPYTER = 1; + // The Presto query engine. + PRESTO = 6; + // The Zeppelin notebook. ZEPPELIN = 4; + + // The Zookeeper service. + ZOOKEEPER = 8; } diff --git a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto index 2db55798..d1cfcc09 100644 --- a/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto +++ b/proto-google-cloud-dataproc-v1/src/main/proto/google/cloud/dataproc/v1/workflow_templates.proto @@ -329,7 +329,13 @@ message OrderedJob { PigJob pig_job = 6; + // Spark R job + SparkRJob spark_r_job = 11; + SparkSqlJob spark_sql_job = 7; + + // Presto job + PrestoJob presto_job = 12; } // Optional. The labels to associate with this job. diff --git a/proto-google-cloud-dataproc-v1beta2/clirr-ignored-differences.xml b/proto-google-cloud-dataproc-v1beta2/clirr-ignored-differences.xml index e1497cef..afdef514 100644 --- a/proto-google-cloud-dataproc-v1beta2/clirr-ignored-differences.xml +++ b/proto-google-cloud-dataproc-v1beta2/clirr-ignored-differences.xml @@ -2,8 +2,18 @@ - 7002 - com/google/cloud/dataproc/v1beta2/AutoscalingPolicyName$Builder - * *etRegion(*) + 7012 + com/google/cloud/dataproc/v1beta2/*OrBuilder + * get*(*) + + + 7012 + com/google/cloud/dataproc/v1beta2/*OrBuilder + boolean contains*(*) + + + 7012 + com/google/cloud/dataproc/v1beta2/*OrBuilder + boolean has*(*) diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfig.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfig.java index b2a9b25f..810d4377 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfig.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfig.java @@ -23,7 +23,8 @@ * *
  * Specifies the type and number of accelerator cards attached to the instances
- * of an instance group (see [GPUs on Compute Engine](/compute/docs/gpus/)).
+ * of an instance group (see [GPUs on Compute
+ * Engine](https://cloud.google.com/compute/docs/gpus/)).
  * 
* * Protobuf type {@code google.cloud.dataproc.v1beta2.AcceleratorConfig} @@ -126,14 +127,14 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * @@ -160,14 +161,14 @@ public java.lang.String getAcceleratorTypeUri() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * @@ -378,7 +379,8 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * *
    * Specifies the type and number of accelerator cards attached to the instances
-   * of an instance group (see [GPUs on Compute Engine](/compute/docs/gpus/)).
+   * of an instance group (see [GPUs on Compute
+   * Engine](https://cloud.google.com/compute/docs/gpus/)).
    * 
* * Protobuf type {@code google.cloud.dataproc.v1beta2.AcceleratorConfig} @@ -547,14 +549,14 @@ public Builder mergeFrom( * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * @@ -581,14 +583,14 @@ public java.lang.String getAcceleratorTypeUri() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * @@ -615,14 +617,14 @@ public com.google.protobuf.ByteString getAcceleratorTypeUriBytes() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * @@ -648,14 +650,14 @@ public Builder setAcceleratorTypeUri(java.lang.String value) { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * @@ -677,14 +679,14 @@ public Builder clearAcceleratorTypeUri() { * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfigOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfigOrBuilder.java index 99bddd89..bac648d0 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/AcceleratorConfigOrBuilder.java @@ -30,14 +30,14 @@ public interface AcceleratorConfigOrBuilder * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * @@ -54,14 +54,14 @@ public interface AcceleratorConfigOrBuilder * Full URL, partial URI, or short name of the accelerator type resource to * expose to this instance. See * [Compute Engine - * AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + * AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) * Examples * * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` * * `nvidia-tesla-k80` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the accelerator type * resource, for example, `nvidia-tesla-k80`. * diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfig.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfig.java index f3b41791..41a5006f 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfig.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfig.java @@ -258,6 +258,23 @@ private ClusterConfig( securityConfig_ = subBuilder.buildPartial(); } + break; + } + case 154: + { + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.Builder subBuilder = null; + if (gkeClusterConfig_ != null) { + subBuilder = gkeClusterConfig_.toBuilder(); + } + gkeClusterConfig_ = + input.readMessage( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.parser(), + extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom(gkeClusterConfig_); + gkeClusterConfig_ = subBuilder.buildPartial(); + } + break; } default: @@ -311,7 +328,7 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -341,7 +358,7 @@ public java.lang.String getConfigBucket() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1023,6 +1040,67 @@ public com.google.cloud.dataproc.v1beta2.SecurityConfigOrBuilder getSecurityConf return getSecurityConfig(); } + public static final int GKE_CLUSTER_CONFIG_FIELD_NUMBER = 19; + private com.google.cloud.dataproc.v1beta2.GkeClusterConfig gkeClusterConfig_; + /** + * + * + *
+   * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+   * Setting this is considered mutually exclusive with Compute Engine-based
+   * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+   * `secondary_worker_config`, and `autoscaling_config`.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the gkeClusterConfig field is set. + */ + public boolean hasGkeClusterConfig() { + return gkeClusterConfig_ != null; + } + /** + * + * + *
+   * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+   * Setting this is considered mutually exclusive with Compute Engine-based
+   * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+   * `secondary_worker_config`, and `autoscaling_config`.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The gkeClusterConfig. + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig getGkeClusterConfig() { + return gkeClusterConfig_ == null + ? com.google.cloud.dataproc.v1beta2.GkeClusterConfig.getDefaultInstance() + : gkeClusterConfig_; + } + /** + * + * + *
+   * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+   * Setting this is considered mutually exclusive with Compute Engine-based
+   * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+   * `secondary_worker_config`, and `autoscaling_config`.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfigOrBuilder + getGkeClusterConfigOrBuilder() { + return getGkeClusterConfig(); + } + private byte memoizedIsInitialized = -1; @java.lang.Override @@ -1073,6 +1151,9 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io if (securityConfig_ != null) { output.writeMessage(18, getSecurityConfig()); } + if (gkeClusterConfig_ != null) { + output.writeMessage(19, getGkeClusterConfig()); + } unknownFields.writeTo(output); } @@ -1121,6 +1202,9 @@ public int getSerializedSize() { if (securityConfig_ != null) { size += com.google.protobuf.CodedOutputStream.computeMessageSize(18, getSecurityConfig()); } + if (gkeClusterConfig_ != null) { + size += com.google.protobuf.CodedOutputStream.computeMessageSize(19, getGkeClusterConfig()); + } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; @@ -1179,6 +1263,10 @@ public boolean equals(final java.lang.Object obj) { if (hasSecurityConfig()) { if (!getSecurityConfig().equals(other.getSecurityConfig())) return false; } + if (hasGkeClusterConfig() != other.hasGkeClusterConfig()) return false; + if (hasGkeClusterConfig()) { + if (!getGkeClusterConfig().equals(other.getGkeClusterConfig())) return false; + } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @@ -1236,6 +1324,10 @@ public int hashCode() { hash = (37 * hash) + SECURITY_CONFIG_FIELD_NUMBER; hash = (53 * hash) + getSecurityConfig().hashCode(); } + if (hasGkeClusterConfig()) { + hash = (37 * hash) + GKE_CLUSTER_CONFIG_FIELD_NUMBER; + hash = (53 * hash) + getGkeClusterConfig().hashCode(); + } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; @@ -1451,6 +1543,12 @@ public Builder clear() { securityConfig_ = null; securityConfigBuilder_ = null; } + if (gkeClusterConfigBuilder_ == null) { + gkeClusterConfig_ = null; + } else { + gkeClusterConfig_ = null; + gkeClusterConfigBuilder_ = null; + } return this; } @@ -1539,6 +1637,11 @@ public com.google.cloud.dataproc.v1beta2.ClusterConfig buildPartial() { } else { result.securityConfig_ = securityConfigBuilder_.build(); } + if (gkeClusterConfigBuilder_ == null) { + result.gkeClusterConfig_ = gkeClusterConfig_; + } else { + result.gkeClusterConfig_ = gkeClusterConfigBuilder_.build(); + } onBuilt(); return result; } @@ -1650,6 +1753,9 @@ public Builder mergeFrom(com.google.cloud.dataproc.v1beta2.ClusterConfig other) if (other.hasSecurityConfig()) { mergeSecurityConfig(other.getSecurityConfig()); } + if (other.hasGkeClusterConfig()) { + mergeGkeClusterConfig(other.getGkeClusterConfig()); + } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; @@ -1694,7 +1800,7 @@ public Builder mergeFrom( * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1724,7 +1830,7 @@ public java.lang.String getConfigBucket() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1754,7 +1860,7 @@ public com.google.protobuf.ByteString getConfigBucketBytes() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1783,7 +1889,7 @@ public Builder setConfigBucket(java.lang.String value) { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1808,7 +1914,7 @@ public Builder clearConfigBucket() { * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -4530,6 +4636,238 @@ public com.google.cloud.dataproc.v1beta2.SecurityConfigOrBuilder getSecurityConf return securityConfigBuilder_; } + private com.google.cloud.dataproc.v1beta2.GkeClusterConfig gkeClusterConfig_; + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.GkeClusterConfig, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.Builder, + com.google.cloud.dataproc.v1beta2.GkeClusterConfigOrBuilder> + gkeClusterConfigBuilder_; + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the gkeClusterConfig field is set. + */ + public boolean hasGkeClusterConfig() { + return gkeClusterConfigBuilder_ != null || gkeClusterConfig_ != null; + } + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The gkeClusterConfig. + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig getGkeClusterConfig() { + if (gkeClusterConfigBuilder_ == null) { + return gkeClusterConfig_ == null + ? com.google.cloud.dataproc.v1beta2.GkeClusterConfig.getDefaultInstance() + : gkeClusterConfig_; + } else { + return gkeClusterConfigBuilder_.getMessage(); + } + } + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setGkeClusterConfig(com.google.cloud.dataproc.v1beta2.GkeClusterConfig value) { + if (gkeClusterConfigBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + gkeClusterConfig_ = value; + onChanged(); + } else { + gkeClusterConfigBuilder_.setMessage(value); + } + + return this; + } + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setGkeClusterConfig( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.Builder builderForValue) { + if (gkeClusterConfigBuilder_ == null) { + gkeClusterConfig_ = builderForValue.build(); + onChanged(); + } else { + gkeClusterConfigBuilder_.setMessage(builderForValue.build()); + } + + return this; + } + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder mergeGkeClusterConfig(com.google.cloud.dataproc.v1beta2.GkeClusterConfig value) { + if (gkeClusterConfigBuilder_ == null) { + if (gkeClusterConfig_ != null) { + gkeClusterConfig_ = + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.newBuilder(gkeClusterConfig_) + .mergeFrom(value) + .buildPartial(); + } else { + gkeClusterConfig_ = value; + } + onChanged(); + } else { + gkeClusterConfigBuilder_.mergeFrom(value); + } + + return this; + } + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder clearGkeClusterConfig() { + if (gkeClusterConfigBuilder_ == null) { + gkeClusterConfig_ = null; + onChanged(); + } else { + gkeClusterConfig_ = null; + gkeClusterConfigBuilder_ = null; + } + + return this; + } + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.Builder getGkeClusterConfigBuilder() { + + onChanged(); + return getGkeClusterConfigFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfigOrBuilder + getGkeClusterConfigOrBuilder() { + if (gkeClusterConfigBuilder_ != null) { + return gkeClusterConfigBuilder_.getMessageOrBuilder(); + } else { + return gkeClusterConfig_ == null + ? com.google.cloud.dataproc.v1beta2.GkeClusterConfig.getDefaultInstance() + : gkeClusterConfig_; + } + } + /** + * + * + *
+     * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+     * Setting this is considered mutually exclusive with Compute Engine-based
+     * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+     * `secondary_worker_config`, and `autoscaling_config`.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.GkeClusterConfig, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.Builder, + com.google.cloud.dataproc.v1beta2.GkeClusterConfigOrBuilder> + getGkeClusterConfigFieldBuilder() { + if (gkeClusterConfigBuilder_ == null) { + gkeClusterConfigBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.GkeClusterConfig, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.Builder, + com.google.cloud.dataproc.v1beta2.GkeClusterConfigOrBuilder>( + getGkeClusterConfig(), getParentForChildren(), isClean()); + gkeClusterConfig_ = null; + } + return gkeClusterConfigBuilder_; + } + @java.lang.Override public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfigOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfigOrBuilder.java index 59237134..cd74b235 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterConfigOrBuilder.java @@ -35,7 +35,7 @@ public interface ClusterConfigOrBuilder * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -55,7 +55,7 @@ public interface ClusterConfigOrBuilder * Compute Engine zone where your cluster is deployed, and then create * and manage this project-level, per-location bucket (see * [Dataproc staging - * bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + * bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). * * * string config_bucket = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -608,4 +608,54 @@ public interface ClusterConfigOrBuilder *
*/ com.google.cloud.dataproc.v1beta2.SecurityConfigOrBuilder getSecurityConfigOrBuilder(); + + /** + * + * + *
+   * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+   * Setting this is considered mutually exclusive with Compute Engine-based
+   * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+   * `secondary_worker_config`, and `autoscaling_config`.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the gkeClusterConfig field is set. + */ + boolean hasGkeClusterConfig(); + /** + * + * + *
+   * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+   * Setting this is considered mutually exclusive with Compute Engine-based
+   * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+   * `secondary_worker_config`, and `autoscaling_config`.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The gkeClusterConfig. + */ + com.google.cloud.dataproc.v1beta2.GkeClusterConfig getGkeClusterConfig(); + /** + * + * + *
+   * Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes.
+   * Setting this is considered mutually exclusive with Compute Engine-based
+   * options such as `gce_cluster_config`, `master_config`, `worker_config`,
+   * `secondary_worker_config`, and `autoscaling_config`.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig gke_cluster_config = 19 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + com.google.cloud.dataproc.v1beta2.GkeClusterConfigOrBuilder getGkeClusterConfigOrBuilder(); } diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterStatus.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterStatus.java index b31e5e36..ada80d9f 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterStatus.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClusterStatus.java @@ -212,6 +212,36 @@ public enum State implements com.google.protobuf.ProtocolMessageEnum { * UPDATING = 5; */ UPDATING(5), + /** + * + * + *
+     * The cluster is being stopped. It cannot be used.
+     * 
+ * + * STOPPING = 6; + */ + STOPPING(6), + /** + * + * + *
+     * The cluster is currently stopped. It is not ready for use.
+     * 
+ * + * STOPPED = 7; + */ + STOPPED(7), + /** + * + * + *
+     * The cluster is being started. It is not ready for use.
+     * 
+ * + * STARTING = 8; + */ + STARTING(8), UNRECOGNIZED(-1), ; @@ -275,6 +305,36 @@ public enum State implements com.google.protobuf.ProtocolMessageEnum { * UPDATING = 5; */ public static final int UPDATING_VALUE = 5; + /** + * + * + *
+     * The cluster is being stopped. It cannot be used.
+     * 
+ * + * STOPPING = 6; + */ + public static final int STOPPING_VALUE = 6; + /** + * + * + *
+     * The cluster is currently stopped. It is not ready for use.
+     * 
+ * + * STOPPED = 7; + */ + public static final int STOPPED_VALUE = 7; + /** + * + * + *
+     * The cluster is being started. It is not ready for use.
+     * 
+ * + * STARTING = 8; + */ + public static final int STARTING_VALUE = 8; public final int getNumber() { if (this == UNRECOGNIZED) { @@ -312,6 +372,12 @@ public static State forNumber(int value) { return DELETING; case 5: return UPDATING; + case 6: + return STOPPING; + case 7: + return STOPPED; + case 8: + return STARTING; default: return null; } diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java index f4ef278e..9f45a968 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/ClustersProto.java @@ -39,6 +39,14 @@ public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry r internal_static_google_cloud_dataproc_v1beta2_ClusterConfig_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_google_cloud_dataproc_v1beta2_ClusterConfig_fieldAccessorTable; + static final com.google.protobuf.Descriptors.Descriptor + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_descriptor; + static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_fieldAccessorTable; + static final com.google.protobuf.Descriptors.Descriptor + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_descriptor; + static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_fieldAccessorTable; static final com.google.protobuf.Descriptors.Descriptor internal_static_google_cloud_dataproc_v1beta2_EndpointConfig_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable @@ -185,7 +193,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "\n\014cluster_uuid\030\006 \001(\tB\003\340A\003\022C\n\007metrics\030\t \001" + "(\0132-.google.cloud.dataproc.v1beta2.Clust" + "erMetricsB\003\340A\003\032-\n\013LabelsEntry\022\013\n\003key\030\001 \001" - + "(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\257\007\n\rClusterConfig\022" + + "(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\201\010\n\rClusterConfig\022" + "\032\n\rconfig_bucket\030\001 \001(\tB\003\340A\001\022P\n\022gce_clust" + "er_config\030\010 \001(\0132/.google.cloud.dataproc." + "v1beta2.GceClusterConfigB\003\340A\001\022N\n\rmaster_" @@ -209,173 +217,183 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + ".google.cloud.dataproc.v1beta2.EndpointC" + "onfigB\003\340A\001\022K\n\017security_config\030\022 \001(\0132-.go" + "ogle.cloud.dataproc.v1beta2.SecurityConf" - + "igB\003\340A\001\"\277\001\n\016EndpointConfig\022U\n\nhttp_ports" - + "\030\001 \003(\0132<.google.cloud.dataproc.v1beta2.E" - + "ndpointConfig.HttpPortsEntryB\003\340A\003\022$\n\027ena" - + "ble_http_port_access\030\002 \001(\010B\003\340A\001\0320\n\016HttpP" - + "ortsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028" - + "\001\",\n\021AutoscalingConfig\022\027\n\npolicy_uri\030\001 \001" - + "(\tB\003\340A\001\"4\n\020EncryptionConfig\022 \n\023gce_pd_km" - + "s_key_name\030\001 \001(\tB\003\340A\001\"\251\003\n\020GceClusterConf" - + "ig\022\025\n\010zone_uri\030\001 \001(\tB\003\340A\001\022\030\n\013network_uri" - + "\030\002 \001(\tB\003\340A\001\022\033\n\016subnetwork_uri\030\006 \001(\tB\003\340A\001" - + "\022\035\n\020internal_ip_only\030\007 \001(\010B\003\340A\001\022\034\n\017servi" - + "ce_account\030\010 \001(\tB\003\340A\001\022#\n\026service_account" - + "_scopes\030\003 \003(\tB\003\340A\001\022\014\n\004tags\030\004 \003(\t\022O\n\010meta" - + "data\030\005 \003(\0132=.google.cloud.dataproc.v1bet" - + "a2.GceClusterConfig.MetadataEntry\022U\n\024res" - + "ervation_affinity\030\013 \001(\01322.google.cloud.d" - + "ataproc.v1beta2.ReservationAffinityB\003\340A\001" - + "\032/\n\rMetadataEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030" - + "\002 \001(\t:\0028\001\"\244\003\n\023InstanceGroupConfig\022\032\n\rnum" - + "_instances\030\001 \001(\005B\003\340A\001\022\033\n\016instance_names\030" - + "\002 \003(\tB\003\340A\003\022\026\n\timage_uri\030\003 \001(\tB\003\340A\001\022\035\n\020ma" - + "chine_type_uri\030\004 \001(\tB\003\340A\001\022C\n\013disk_config" - + "\030\005 \001(\0132).google.cloud.dataproc.v1beta2.D" - + "iskConfigB\003\340A\001\022\033\n\016is_preemptible\030\006 \001(\010B\003" - + "\340A\001\022T\n\024managed_group_config\030\007 \001(\01321.goog" - + "le.cloud.dataproc.v1beta2.ManagedGroupCo" - + "nfigB\003\340A\003\022K\n\014accelerators\030\010 \003(\01320.google" - + ".cloud.dataproc.v1beta2.AcceleratorConfi" - + "gB\003\340A\001\022\030\n\020min_cpu_platform\030\t \001(\t\"c\n\022Mana" - + "gedGroupConfig\022#\n\026instance_template_name" - + "\030\001 \001(\tB\003\340A\003\022(\n\033instance_group_manager_na" - + "me\030\002 \001(\tB\003\340A\003\"L\n\021AcceleratorConfig\022\034\n\024ac" - + "celerator_type_uri\030\001 \001(\t\022\031\n\021accelerator_" - + "count\030\002 \001(\005\"a\n\nDiskConfig\022\033\n\016boot_disk_t" - + "ype\030\003 \001(\tB\003\340A\001\022\036\n\021boot_disk_size_gb\030\001 \001(" - + "\005B\003\340A\001\022\026\n\016num_local_ssds\030\002 \001(\005\"\203\002\n\017Lifec" - + "ycleConfig\0227\n\017idle_delete_ttl\030\001 \001(\0132\031.go" - + "ogle.protobuf.DurationB\003\340A\001\022;\n\020auto_dele" - + "te_time\030\002 \001(\0132\032.google.protobuf.Timestam" - + "pB\003\340A\001H\000\0229\n\017auto_delete_ttl\030\003 \001(\0132\031.goog" - + "le.protobuf.DurationB\003\340A\001H\000\0228\n\017idle_star" - + "t_time\030\004 \001(\0132\032.google.protobuf.Timestamp" - + "B\003\340A\003B\005\n\003ttl\"X\n\016SecurityConfig\022F\n\017kerber" - + "os_config\030\001 \001(\0132-.google.cloud.dataproc." - + "v1beta2.KerberosConfig\"\220\004\n\016KerberosConfi" - + "g\022\034\n\017enable_kerberos\030\001 \001(\010B\003\340A\001\022(\n\033root_" - + "principal_password_uri\030\002 \001(\tB\003\340A\002\022\030\n\013kms" - + "_key_uri\030\003 \001(\tB\003\340A\002\022\031\n\014keystore_uri\030\004 \001(" - + "\tB\003\340A\001\022\033\n\016truststore_uri\030\005 \001(\tB\003\340A\001\022\"\n\025k" - + "eystore_password_uri\030\006 \001(\tB\003\340A\001\022\035\n\020key_p" - + "assword_uri\030\007 \001(\tB\003\340A\001\022$\n\027truststore_pas" - + "sword_uri\030\010 \001(\tB\003\340A\001\022$\n\027cross_realm_trus" - + "t_realm\030\t \001(\tB\003\340A\001\022\"\n\025cross_realm_trust_" - + "kdc\030\n \001(\tB\003\340A\001\022+\n\036cross_realm_trust_admi" - + "n_server\030\013 \001(\tB\003\340A\001\0222\n%cross_realm_trust" - + "_shared_password_uri\030\014 \001(\tB\003\340A\001\022\033\n\016kdc_d" - + "b_key_uri\030\r \001(\tB\003\340A\001\022\037\n\022tgt_lifetime_hou" - + "rs\030\016 \001(\005B\003\340A\001\022\022\n\005realm\030\017 \001(\tB\003\340A\001\"s\n\030Nod" - + "eInitializationAction\022\034\n\017executable_file" - + "\030\001 \001(\tB\003\340A\002\0229\n\021execution_timeout\030\002 \001(\0132\031" - + ".google.protobuf.DurationB\003\340A\001\"\213\003\n\rClust" - + "erStatus\022F\n\005state\030\001 \001(\01622.google.cloud.d" - + "ataproc.v1beta2.ClusterStatus.StateB\003\340A\003" - + "\022\023\n\006detail\030\002 \001(\tB\003\340A\003\0229\n\020state_start_tim" - + "e\030\003 \001(\0132\032.google.protobuf.TimestampB\003\340A\003" - + "\022L\n\010substate\030\004 \001(\01625.google.cloud.datapr" - + "oc.v1beta2.ClusterStatus.SubstateB\003\340A\003\"V" - + "\n\005State\022\013\n\007UNKNOWN\020\000\022\014\n\010CREATING\020\001\022\013\n\007RU" - + "NNING\020\002\022\t\n\005ERROR\020\003\022\014\n\010DELETING\020\004\022\014\n\010UPDA" - + "TING\020\005\"<\n\010Substate\022\017\n\013UNSPECIFIED\020\000\022\r\n\tU" - + "NHEALTHY\020\001\022\020\n\014STALE_STATUS\020\002\"\376\001\n\016Softwar" - + "eConfig\022\032\n\rimage_version\030\001 \001(\tB\003\340A\001\022V\n\np" - + "roperties\030\002 \003(\0132=.google.cloud.dataproc." - + "v1beta2.SoftwareConfig.PropertiesEntryB\003" - + "\340A\001\022E\n\023optional_components\030\003 \003(\0162(.googl" - + "e.cloud.dataproc.v1beta2.Component\0321\n\017Pr" - + "opertiesEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(" - + "\t:\0028\001\"\244\002\n\016ClusterMetrics\022T\n\014hdfs_metrics" - + "\030\001 \003(\0132>.google.cloud.dataproc.v1beta2.C" - + "lusterMetrics.HdfsMetricsEntry\022T\n\014yarn_m" - + "etrics\030\002 \003(\0132>.google.cloud.dataproc.v1b" - + "eta2.ClusterMetrics.YarnMetricsEntry\0322\n\020" - + "HdfsMetricsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002" - + " \001(\003:\0028\001\0322\n\020YarnMetricsEntry\022\013\n\003key\030\001 \001(" - + "\t\022\r\n\005value\030\002 \001(\003:\0028\001\"\233\001\n\024CreateClusterRe" - + "quest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region" - + "\030\003 \001(\tB\003\340A\002\022<\n\007cluster\030\002 \001(\0132&.google.cl" - + "oud.dataproc.v1beta2.ClusterB\003\340A\002\022\027\n\nreq" - + "uest_id\030\004 \001(\tB\003\340A\001\"\263\002\n\024UpdateClusterRequ" - + "est\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\005" - + " \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\022<\n\007" - + "cluster\030\003 \001(\0132&.google.cloud.dataproc.v1" - + "beta2.ClusterB\003\340A\002\022E\n\035graceful_decommiss" - + "ion_timeout\030\006 \001(\0132\031.google.protobuf.Dura" - + "tionB\003\340A\001\0224\n\013update_mask\030\004 \001(\0132\032.google." - + "protobuf.FieldMaskB\003\340A\002\022\027\n\nrequest_id\030\007 " - + "\001(\tB\003\340A\001\"\223\001\n\024DeleteClusterRequest\022\027\n\npro" - + "ject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022" - + "\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\022\031\n\014cluster_uu" - + "id\030\004 \001(\tB\003\340A\001\022\027\n\nrequest_id\030\005 \001(\tB\003\340A\001\"\\" - + "\n\021GetClusterRequest\022\027\n\nproject_id\030\001 \001(\tB" - + "\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_nam" - + "e\030\002 \001(\tB\003\340A\002\"\211\001\n\023ListClustersRequest\022\027\n\n" - + "project_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\004 \001(\tB\003\340" - + "A\002\022\023\n\006filter\030\005 \001(\tB\003\340A\001\022\026\n\tpage_size\030\002 \001" - + "(\005B\003\340A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001\"s\n\024List" - + "ClustersResponse\022=\n\010clusters\030\001 \003(\0132&.goo" - + "gle.cloud.dataproc.v1beta2.ClusterB\003\340A\003\022" - + "\034\n\017next_page_token\030\002 \001(\tB\003\340A\003\"a\n\026Diagnos" - + "eClusterRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002" - + "\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 " - + "\001(\tB\003\340A\002\"1\n\026DiagnoseClusterResults\022\027\n\nou" - + "tput_uri\030\001 \001(\tB\003\340A\003\"\375\001\n\023ReservationAffin" - + "ity\022^\n\030consume_reservation_type\030\001 \001(\01627." - + "google.cloud.dataproc.v1beta2.Reservatio" - + "nAffinity.TypeB\003\340A\001\022\020\n\003key\030\002 \001(\tB\003\340A\001\022\023\n" - + "\006values\030\003 \003(\tB\003\340A\001\"_\n\004Type\022\024\n\020TYPE_UNSPE" - + "CIFIED\020\000\022\022\n\016NO_RESERVATION\020\001\022\023\n\017ANY_RESE" - + "RVATION\020\002\022\030\n\024SPECIFIC_RESERVATION\020\0032\347\r\n\021" - + "ClusterController\022\221\002\n\rCreateCluster\0223.go" - + "ogle.cloud.dataproc.v1beta2.CreateCluste" + + "igB\003\340A\001\022P\n\022gke_cluster_config\030\023 \001(\0132/.go" + + "ogle.cloud.dataproc.v1beta2.GkeClusterCo" + + "nfigB\003\340A\001\"\230\002\n\020GkeClusterConfig\022|\n namesp" + + "aced_gke_deployment_target\030\001 \001(\0132M.googl" + + "e.cloud.dataproc.v1beta2.GkeClusterConfi" + + "g.NamespacedGkeDeploymentTargetB\003\340A\001\032\205\001\n" + + "\035NamespacedGkeDeploymentTarget\022D\n\022target" + + "_gke_cluster\030\001 \001(\tB(\340A\001\372A\"\n container.go" + + "ogleapis.com/Cluster\022\036\n\021cluster_namespac" + + "e\030\002 \001(\tB\003\340A\001\"\277\001\n\016EndpointConfig\022U\n\nhttp_" + + "ports\030\001 \003(\0132<.google.cloud.dataproc.v1be" + + "ta2.EndpointConfig.HttpPortsEntryB\003\340A\003\022$" + + "\n\027enable_http_port_access\030\002 \001(\010B\003\340A\001\0320\n\016" + + "HttpPortsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001" + + "(\t:\0028\001\",\n\021AutoscalingConfig\022\027\n\npolicy_ur" + + "i\030\001 \001(\tB\003\340A\001\"4\n\020EncryptionConfig\022 \n\023gce_" + + "pd_kms_key_name\030\001 \001(\tB\003\340A\001\"\251\003\n\020GceCluste" + + "rConfig\022\025\n\010zone_uri\030\001 \001(\tB\003\340A\001\022\030\n\013networ" + + "k_uri\030\002 \001(\tB\003\340A\001\022\033\n\016subnetwork_uri\030\006 \001(\t" + + "B\003\340A\001\022\035\n\020internal_ip_only\030\007 \001(\010B\003\340A\001\022\034\n\017" + + "service_account\030\010 \001(\tB\003\340A\001\022#\n\026service_ac" + + "count_scopes\030\003 \003(\tB\003\340A\001\022\014\n\004tags\030\004 \003(\t\022O\n" + + "\010metadata\030\005 \003(\0132=.google.cloud.dataproc." + + "v1beta2.GceClusterConfig.MetadataEntry\022U" + + "\n\024reservation_affinity\030\013 \001(\01322.google.cl" + + "oud.dataproc.v1beta2.ReservationAffinity" + + "B\003\340A\001\032/\n\rMetadataEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005v" + + "alue\030\002 \001(\t:\0028\001\"\244\003\n\023InstanceGroupConfig\022\032" + + "\n\rnum_instances\030\001 \001(\005B\003\340A\001\022\033\n\016instance_n" + + "ames\030\002 \003(\tB\003\340A\003\022\026\n\timage_uri\030\003 \001(\tB\003\340A\001\022" + + "\035\n\020machine_type_uri\030\004 \001(\tB\003\340A\001\022C\n\013disk_c" + + "onfig\030\005 \001(\0132).google.cloud.dataproc.v1be" + + "ta2.DiskConfigB\003\340A\001\022\033\n\016is_preemptible\030\006 " + + "\001(\010B\003\340A\003\022T\n\024managed_group_config\030\007 \001(\01321" + + ".google.cloud.dataproc.v1beta2.ManagedGr" + + "oupConfigB\003\340A\003\022K\n\014accelerators\030\010 \003(\01320.g" + + "oogle.cloud.dataproc.v1beta2.Accelerator" + + "ConfigB\003\340A\001\022\030\n\020min_cpu_platform\030\t \001(\t\"c\n" + + "\022ManagedGroupConfig\022#\n\026instance_template" + + "_name\030\001 \001(\tB\003\340A\003\022(\n\033instance_group_manag" + + "er_name\030\002 \001(\tB\003\340A\003\"L\n\021AcceleratorConfig\022" + + "\034\n\024accelerator_type_uri\030\001 \001(\t\022\031\n\021acceler" + + "ator_count\030\002 \001(\005\"a\n\nDiskConfig\022\033\n\016boot_d" + + "isk_type\030\003 \001(\tB\003\340A\001\022\036\n\021boot_disk_size_gb" + + "\030\001 \001(\005B\003\340A\001\022\026\n\016num_local_ssds\030\002 \001(\005\"\203\002\n\017" + + "LifecycleConfig\0227\n\017idle_delete_ttl\030\001 \001(\013" + + "2\031.google.protobuf.DurationB\003\340A\001\022;\n\020auto" + + "_delete_time\030\002 \001(\0132\032.google.protobuf.Tim" + + "estampB\003\340A\001H\000\0229\n\017auto_delete_ttl\030\003 \001(\0132\031" + + ".google.protobuf.DurationB\003\340A\001H\000\0228\n\017idle" + + "_start_time\030\004 \001(\0132\032.google.protobuf.Time" + + "stampB\003\340A\003B\005\n\003ttl\"X\n\016SecurityConfig\022F\n\017k" + + "erberos_config\030\001 \001(\0132-.google.cloud.data" + + "proc.v1beta2.KerberosConfig\"\220\004\n\016Kerberos" + + "Config\022\034\n\017enable_kerberos\030\001 \001(\010B\003\340A\001\022(\n\033" + + "root_principal_password_uri\030\002 \001(\tB\003\340A\002\022\030" + + "\n\013kms_key_uri\030\003 \001(\tB\003\340A\002\022\031\n\014keystore_uri" + + "\030\004 \001(\tB\003\340A\001\022\033\n\016truststore_uri\030\005 \001(\tB\003\340A\001" + + "\022\"\n\025keystore_password_uri\030\006 \001(\tB\003\340A\001\022\035\n\020" + + "key_password_uri\030\007 \001(\tB\003\340A\001\022$\n\027truststor" + + "e_password_uri\030\010 \001(\tB\003\340A\001\022$\n\027cross_realm" + + "_trust_realm\030\t \001(\tB\003\340A\001\022\"\n\025cross_realm_t" + + "rust_kdc\030\n \001(\tB\003\340A\001\022+\n\036cross_realm_trust" + + "_admin_server\030\013 \001(\tB\003\340A\001\0222\n%cross_realm_" + + "trust_shared_password_uri\030\014 \001(\tB\003\340A\001\022\033\n\016" + + "kdc_db_key_uri\030\r \001(\tB\003\340A\001\022\037\n\022tgt_lifetim" + + "e_hours\030\016 \001(\005B\003\340A\001\022\022\n\005realm\030\017 \001(\tB\003\340A\001\"s" + + "\n\030NodeInitializationAction\022\034\n\017executable" + + "_file\030\001 \001(\tB\003\340A\002\0229\n\021execution_timeout\030\002 " + + "\001(\0132\031.google.protobuf.DurationB\003\340A\001\"\264\003\n\r" + + "ClusterStatus\022F\n\005state\030\001 \001(\01622.google.cl" + + "oud.dataproc.v1beta2.ClusterStatus.State" + + "B\003\340A\003\022\023\n\006detail\030\002 \001(\tB\003\340A\003\0229\n\020state_star" + + "t_time\030\003 \001(\0132\032.google.protobuf.Timestamp" + + "B\003\340A\003\022L\n\010substate\030\004 \001(\01625.google.cloud.d" + + "ataproc.v1beta2.ClusterStatus.SubstateB\003" + + "\340A\003\"\177\n\005State\022\013\n\007UNKNOWN\020\000\022\014\n\010CREATING\020\001\022" + + "\013\n\007RUNNING\020\002\022\t\n\005ERROR\020\003\022\014\n\010DELETING\020\004\022\014\n" + + "\010UPDATING\020\005\022\014\n\010STOPPING\020\006\022\013\n\007STOPPED\020\007\022\014" + + "\n\010STARTING\020\010\"<\n\010Substate\022\017\n\013UNSPECIFIED\020" + + "\000\022\r\n\tUNHEALTHY\020\001\022\020\n\014STALE_STATUS\020\002\"\376\001\n\016S" + + "oftwareConfig\022\032\n\rimage_version\030\001 \001(\tB\003\340A" + + "\001\022V\n\nproperties\030\002 \003(\0132=.google.cloud.dat" + + "aproc.v1beta2.SoftwareConfig.PropertiesE" + + "ntryB\003\340A\001\022E\n\023optional_components\030\003 \003(\0162(" + + ".google.cloud.dataproc.v1beta2.Component" + + "\0321\n\017PropertiesEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005valu" + + "e\030\002 \001(\t:\0028\001\"\244\002\n\016ClusterMetrics\022T\n\014hdfs_m" + + "etrics\030\001 \003(\0132>.google.cloud.dataproc.v1b" + + "eta2.ClusterMetrics.HdfsMetricsEntry\022T\n\014" + + "yarn_metrics\030\002 \003(\0132>.google.cloud.datapr" + + "oc.v1beta2.ClusterMetrics.YarnMetricsEnt" + + "ry\0322\n\020HdfsMetricsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005v" + + "alue\030\002 \001(\003:\0028\001\0322\n\020YarnMetricsEntry\022\013\n\003ke" + + "y\030\001 \001(\t\022\r\n\005value\030\002 \001(\003:\0028\001\"\233\001\n\024CreateClu" + + "sterRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006" + + "region\030\003 \001(\tB\003\340A\002\022<\n\007cluster\030\002 \001(\0132&.goo" + + "gle.cloud.dataproc.v1beta2.ClusterB\003\340A\002\022" + + "\027\n\nrequest_id\030\004 \001(\tB\003\340A\001\"\263\002\n\024UpdateClust" + + "erRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006re" + + "gion\030\005 \001(\tB\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340" + + "A\002\022<\n\007cluster\030\003 \001(\0132&.google.cloud.datap" + + "roc.v1beta2.ClusterB\003\340A\002\022E\n\035graceful_dec" + + "ommission_timeout\030\006 \001(\0132\031.google.protobu" + + "f.DurationB\003\340A\001\0224\n\013update_mask\030\004 \001(\0132\032.g" + + "oogle.protobuf.FieldMaskB\003\340A\002\022\027\n\nrequest" + + "_id\030\007 \001(\tB\003\340A\001\"\223\001\n\024DeleteClusterRequest\022" + + "\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\t" + + "B\003\340A\002\022\031\n\014cluster_name\030\002 \001(\tB\003\340A\002\022\031\n\014clus" + + "ter_uuid\030\004 \001(\tB\003\340A\001\022\027\n\nrequest_id\030\005 \001(\tB" + + "\003\340A\001\"\\\n\021GetClusterRequest\022\027\n\nproject_id\030" + + "\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014clust" + + "er_name\030\002 \001(\tB\003\340A\002\"\211\001\n\023ListClustersReque" + + "st\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\004 " + + "\001(\tB\003\340A\002\022\023\n\006filter\030\005 \001(\tB\003\340A\001\022\026\n\tpage_si" + + "ze\030\002 \001(\005B\003\340A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001\"s" + + "\n\024ListClustersResponse\022=\n\010clusters\030\001 \003(\013" + + "2&.google.cloud.dataproc.v1beta2.Cluster" + + "B\003\340A\003\022\034\n\017next_page_token\030\002 \001(\tB\003\340A\003\"a\n\026D" + + "iagnoseClusterRequest\022\027\n\nproject_id\030\001 \001(" + + "\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\031\n\014cluster_n" + + "ame\030\002 \001(\tB\003\340A\002\"1\n\026DiagnoseClusterResults" + + "\022\027\n\noutput_uri\030\001 \001(\tB\003\340A\003\"\375\001\n\023Reservatio" + + "nAffinity\022^\n\030consume_reservation_type\030\001 " + + "\001(\01627.google.cloud.dataproc.v1beta2.Rese" + + "rvationAffinity.TypeB\003\340A\001\022\020\n\003key\030\002 \001(\tB\003" + + "\340A\001\022\023\n\006values\030\003 \003(\tB\003\340A\001\"_\n\004Type\022\024\n\020TYPE" + + "_UNSPECIFIED\020\000\022\022\n\016NO_RESERVATION\020\001\022\023\n\017AN" + + "Y_RESERVATION\020\002\022\030\n\024SPECIFIC_RESERVATION\020" + + "\0032\347\r\n\021ClusterController\022\221\002\n\rCreateCluste" + + "r\0223.google.cloud.dataproc.v1beta2.Create" + + "ClusterRequest\032\035.google.longrunning.Oper" + + "ation\"\253\001\202\323\344\223\002C\"8/v1beta2/projects/{proje" + + "ct_id}/regions/{region}/clusters:\007cluste" + + "r\332A\033project_id, region, cluster\312AA\n\007Clus" + + "ter\0226google.cloud.dataproc.v1beta2.Clust" + + "erOperationMetadata\022\273\002\n\rUpdateCluster\0223." + + "google.cloud.dataproc.v1beta2.UpdateClus" + + "terRequest\032\035.google.longrunning.Operatio" + + "n\"\325\001\202\323\344\223\002R2G/v1beta2/projects/{project_i" + + "d}/regions/{region}/clusters/{cluster_na" + + "me}:\007cluster\332A6project_id, region, clust" + + "er_name, cluster, update_mask\312AA\n\007Cluste" + + "r\0226google.cloud.dataproc.v1beta2.Cluster" + + "OperationMetadata\022\252\002\n\rDeleteCluster\0223.go" + + "ogle.cloud.dataproc.v1beta2.DeleteCluste" + "rRequest\032\035.google.longrunning.Operation\"" - + "\253\001\202\323\344\223\002C\"8/v1beta2/projects/{project_id}" - + "/regions/{region}/clusters:\007cluster\332A\033pr" - + "oject_id, region, cluster\312AA\n\007Cluster\0226g" - + "oogle.cloud.dataproc.v1beta2.ClusterOper" - + "ationMetadata\022\273\002\n\rUpdateCluster\0223.google" - + ".cloud.dataproc.v1beta2.UpdateClusterReq" - + "uest\032\035.google.longrunning.Operation\"\325\001\202\323" - + "\344\223\002R2G/v1beta2/projects/{project_id}/reg" - + "ions/{region}/clusters/{cluster_name}:\007c" - + "luster\332A6project_id, region, cluster_nam" - + "e, cluster, update_mask\312AA\n\007Cluster\0226goo" - + "gle.cloud.dataproc.v1beta2.ClusterOperat" - + "ionMetadata\022\252\002\n\rDeleteCluster\0223.google.c" - + "loud.dataproc.v1beta2.DeleteClusterReque" - + "st\032\035.google.longrunning.Operation\"\304\001\202\323\344\223" - + "\002I*G/v1beta2/projects/{project_id}/regio" - + "ns/{region}/clusters/{cluster_name}\332A pr" - + "oject_id, region, cluster_name\312AO\n\025googl" - + "e.protobuf.Empty\0226google.cloud.dataproc." - + "v1beta2.ClusterOperationMetadata\022\332\001\n\nGet" - + "Cluster\0220.google.cloud.dataproc.v1beta2." - + "GetClusterRequest\032&.google.cloud.datapro" - + "c.v1beta2.Cluster\"r\202\323\344\223\002I\022G/v1beta2/proj" - + "ects/{project_id}/regions/{region}/clust" - + "ers/{cluster_name}\332A project_id, region," - + " cluster_name\022\353\001\n\014ListClusters\0222.google." - + "cloud.dataproc.v1beta2.ListClustersReque" - + "st\0323.google.cloud.dataproc.v1beta2.ListC" - + "lustersResponse\"r\202\323\344\223\002:\0228/v1beta2/projec" - + "ts/{project_id}/regions/{region}/cluster" - + "s\332A\022project_id, region\332A\032project_id, reg" - + "ion, filter\022\272\002\n\017DiagnoseCluster\0225.google" - + ".cloud.dataproc.v1beta2.DiagnoseClusterR" - + "equest\032\035.google.longrunning.Operation\"\320\001" - + "\202\323\344\223\002U\"P/v1beta2/projects/{project_id}/r" - + "egions/{region}/clusters/{cluster_name}:" - + "diagnose:\001*\332A project_id, region, cluste" - + "r_name\312AO\n\025google.protobuf.Empty\0226google" - + ".cloud.dataproc.v1beta2.ClusterOperation" - + "Metadata\032K\312A\027dataproc.googleapis.com\322A.h" - + "ttps://www.googleapis.com/auth/cloud-pla" - + "tformB{\n!com.google.cloud.dataproc.v1bet" - + "a2B\rClustersProtoP\001ZEgoogle.golang.org/g" - + "enproto/googleapis/cloud/dataproc/v1beta" - + "2;dataprocb\006proto3" + + "\304\001\202\323\344\223\002I*G/v1beta2/projects/{project_id}" + + "/regions/{region}/clusters/{cluster_name" + + "}\332A project_id, region, cluster_name\312AO\n" + + "\025google.protobuf.Empty\0226google.cloud.dat" + + "aproc.v1beta2.ClusterOperationMetadata\022\332" + + "\001\n\nGetCluster\0220.google.cloud.dataproc.v1" + + "beta2.GetClusterRequest\032&.google.cloud.d" + + "ataproc.v1beta2.Cluster\"r\202\323\344\223\002I\022G/v1beta" + + "2/projects/{project_id}/regions/{region}" + + "/clusters/{cluster_name}\332A project_id, r" + + "egion, cluster_name\022\353\001\n\014ListClusters\0222.g" + + "oogle.cloud.dataproc.v1beta2.ListCluster" + + "sRequest\0323.google.cloud.dataproc.v1beta2" + + ".ListClustersResponse\"r\202\323\344\223\002:\0228/v1beta2/" + + "projects/{project_id}/regions/{region}/c" + + "lusters\332A\022project_id, region\332A\032project_i" + + "d, region, filter\022\272\002\n\017DiagnoseCluster\0225." + + "google.cloud.dataproc.v1beta2.DiagnoseCl" + + "usterRequest\032\035.google.longrunning.Operat" + + "ion\"\320\001\202\323\344\223\002U\"P/v1beta2/projects/{project" + + "_id}/regions/{region}/clusters/{cluster_" + + "name}:diagnose:\001*\332A project_id, region, " + + "cluster_name\312AO\n\025google.protobuf.Empty\0226" + + "google.cloud.dataproc.v1beta2.ClusterOpe" + + "rationMetadata\032K\312A\027dataproc.googleapis.c" + + "om\322A.https://www.googleapis.com/auth/clo" + + "ud-platformB{\n!com.google.cloud.dataproc" + + ".v1beta2B\rClustersProtoP\001ZEgoogle.golang" + + ".org/genproto/googleapis/cloud/dataproc/" + + "v1beta2;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( @@ -432,9 +450,28 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "AutoscalingConfig", "EndpointConfig", "SecurityConfig", + "GkeClusterConfig", }); - internal_static_google_cloud_dataproc_v1beta2_EndpointConfig_descriptor = + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_descriptor = getDescriptor().getMessageTypes().get(2); + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_fieldAccessorTable = + new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_descriptor, + new java.lang.String[] { + "NamespacedGkeDeploymentTarget", + }); + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_descriptor = + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_descriptor + .getNestedTypes() + .get(0); + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_fieldAccessorTable = + new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_descriptor, + new java.lang.String[] { + "TargetGkeCluster", "ClusterNamespace", + }); + internal_static_google_cloud_dataproc_v1beta2_EndpointConfig_descriptor = + getDescriptor().getMessageTypes().get(3); internal_static_google_cloud_dataproc_v1beta2_EndpointConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_EndpointConfig_descriptor, @@ -452,7 +489,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Key", "Value", }); internal_static_google_cloud_dataproc_v1beta2_AutoscalingConfig_descriptor = - getDescriptor().getMessageTypes().get(3); + getDescriptor().getMessageTypes().get(4); internal_static_google_cloud_dataproc_v1beta2_AutoscalingConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_AutoscalingConfig_descriptor, @@ -460,7 +497,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "PolicyUri", }); internal_static_google_cloud_dataproc_v1beta2_EncryptionConfig_descriptor = - getDescriptor().getMessageTypes().get(4); + getDescriptor().getMessageTypes().get(5); internal_static_google_cloud_dataproc_v1beta2_EncryptionConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_EncryptionConfig_descriptor, @@ -468,7 +505,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "GcePdKmsKeyName", }); internal_static_google_cloud_dataproc_v1beta2_GceClusterConfig_descriptor = - getDescriptor().getMessageTypes().get(5); + getDescriptor().getMessageTypes().get(6); internal_static_google_cloud_dataproc_v1beta2_GceClusterConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_GceClusterConfig_descriptor, @@ -494,7 +531,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Key", "Value", }); internal_static_google_cloud_dataproc_v1beta2_InstanceGroupConfig_descriptor = - getDescriptor().getMessageTypes().get(6); + getDescriptor().getMessageTypes().get(7); internal_static_google_cloud_dataproc_v1beta2_InstanceGroupConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_InstanceGroupConfig_descriptor, @@ -510,7 +547,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "MinCpuPlatform", }); internal_static_google_cloud_dataproc_v1beta2_ManagedGroupConfig_descriptor = - getDescriptor().getMessageTypes().get(7); + getDescriptor().getMessageTypes().get(8); internal_static_google_cloud_dataproc_v1beta2_ManagedGroupConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_ManagedGroupConfig_descriptor, @@ -518,7 +555,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "InstanceTemplateName", "InstanceGroupManagerName", }); internal_static_google_cloud_dataproc_v1beta2_AcceleratorConfig_descriptor = - getDescriptor().getMessageTypes().get(8); + getDescriptor().getMessageTypes().get(9); internal_static_google_cloud_dataproc_v1beta2_AcceleratorConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_AcceleratorConfig_descriptor, @@ -526,7 +563,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "AcceleratorTypeUri", "AcceleratorCount", }); internal_static_google_cloud_dataproc_v1beta2_DiskConfig_descriptor = - getDescriptor().getMessageTypes().get(9); + getDescriptor().getMessageTypes().get(10); internal_static_google_cloud_dataproc_v1beta2_DiskConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_DiskConfig_descriptor, @@ -534,7 +571,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "BootDiskType", "BootDiskSizeGb", "NumLocalSsds", }); internal_static_google_cloud_dataproc_v1beta2_LifecycleConfig_descriptor = - getDescriptor().getMessageTypes().get(10); + getDescriptor().getMessageTypes().get(11); internal_static_google_cloud_dataproc_v1beta2_LifecycleConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_LifecycleConfig_descriptor, @@ -542,7 +579,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "IdleDeleteTtl", "AutoDeleteTime", "AutoDeleteTtl", "IdleStartTime", "Ttl", }); internal_static_google_cloud_dataproc_v1beta2_SecurityConfig_descriptor = - getDescriptor().getMessageTypes().get(11); + getDescriptor().getMessageTypes().get(12); internal_static_google_cloud_dataproc_v1beta2_SecurityConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_SecurityConfig_descriptor, @@ -550,7 +587,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "KerberosConfig", }); internal_static_google_cloud_dataproc_v1beta2_KerberosConfig_descriptor = - getDescriptor().getMessageTypes().get(12); + getDescriptor().getMessageTypes().get(13); internal_static_google_cloud_dataproc_v1beta2_KerberosConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_KerberosConfig_descriptor, @@ -572,7 +609,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Realm", }); internal_static_google_cloud_dataproc_v1beta2_NodeInitializationAction_descriptor = - getDescriptor().getMessageTypes().get(13); + getDescriptor().getMessageTypes().get(14); internal_static_google_cloud_dataproc_v1beta2_NodeInitializationAction_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_NodeInitializationAction_descriptor, @@ -580,7 +617,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ExecutableFile", "ExecutionTimeout", }); internal_static_google_cloud_dataproc_v1beta2_ClusterStatus_descriptor = - getDescriptor().getMessageTypes().get(14); + getDescriptor().getMessageTypes().get(15); internal_static_google_cloud_dataproc_v1beta2_ClusterStatus_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_ClusterStatus_descriptor, @@ -588,7 +625,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "State", "Detail", "StateStartTime", "Substate", }); internal_static_google_cloud_dataproc_v1beta2_SoftwareConfig_descriptor = - getDescriptor().getMessageTypes().get(15); + getDescriptor().getMessageTypes().get(16); internal_static_google_cloud_dataproc_v1beta2_SoftwareConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_SoftwareConfig_descriptor, @@ -606,7 +643,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Key", "Value", }); internal_static_google_cloud_dataproc_v1beta2_ClusterMetrics_descriptor = - getDescriptor().getMessageTypes().get(16); + getDescriptor().getMessageTypes().get(17); internal_static_google_cloud_dataproc_v1beta2_ClusterMetrics_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_ClusterMetrics_descriptor, @@ -634,7 +671,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Key", "Value", }); internal_static_google_cloud_dataproc_v1beta2_CreateClusterRequest_descriptor = - getDescriptor().getMessageTypes().get(17); + getDescriptor().getMessageTypes().get(18); internal_static_google_cloud_dataproc_v1beta2_CreateClusterRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_CreateClusterRequest_descriptor, @@ -642,7 +679,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "Cluster", "RequestId", }); internal_static_google_cloud_dataproc_v1beta2_UpdateClusterRequest_descriptor = - getDescriptor().getMessageTypes().get(18); + getDescriptor().getMessageTypes().get(19); internal_static_google_cloud_dataproc_v1beta2_UpdateClusterRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_UpdateClusterRequest_descriptor, @@ -656,7 +693,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "RequestId", }); internal_static_google_cloud_dataproc_v1beta2_DeleteClusterRequest_descriptor = - getDescriptor().getMessageTypes().get(19); + getDescriptor().getMessageTypes().get(20); internal_static_google_cloud_dataproc_v1beta2_DeleteClusterRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_DeleteClusterRequest_descriptor, @@ -664,7 +701,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "ClusterName", "ClusterUuid", "RequestId", }); internal_static_google_cloud_dataproc_v1beta2_GetClusterRequest_descriptor = - getDescriptor().getMessageTypes().get(20); + getDescriptor().getMessageTypes().get(21); internal_static_google_cloud_dataproc_v1beta2_GetClusterRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_GetClusterRequest_descriptor, @@ -672,7 +709,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "ClusterName", }); internal_static_google_cloud_dataproc_v1beta2_ListClustersRequest_descriptor = - getDescriptor().getMessageTypes().get(21); + getDescriptor().getMessageTypes().get(22); internal_static_google_cloud_dataproc_v1beta2_ListClustersRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_ListClustersRequest_descriptor, @@ -680,7 +717,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "Filter", "PageSize", "PageToken", }); internal_static_google_cloud_dataproc_v1beta2_ListClustersResponse_descriptor = - getDescriptor().getMessageTypes().get(22); + getDescriptor().getMessageTypes().get(23); internal_static_google_cloud_dataproc_v1beta2_ListClustersResponse_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_ListClustersResponse_descriptor, @@ -688,7 +725,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Clusters", "NextPageToken", }); internal_static_google_cloud_dataproc_v1beta2_DiagnoseClusterRequest_descriptor = - getDescriptor().getMessageTypes().get(23); + getDescriptor().getMessageTypes().get(24); internal_static_google_cloud_dataproc_v1beta2_DiagnoseClusterRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_DiagnoseClusterRequest_descriptor, @@ -696,7 +733,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "ClusterName", }); internal_static_google_cloud_dataproc_v1beta2_DiagnoseClusterResults_descriptor = - getDescriptor().getMessageTypes().get(24); + getDescriptor().getMessageTypes().get(25); internal_static_google_cloud_dataproc_v1beta2_DiagnoseClusterResults_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_DiagnoseClusterResults_descriptor, @@ -704,7 +741,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "OutputUri", }); internal_static_google_cloud_dataproc_v1beta2_ReservationAffinity_descriptor = - getDescriptor().getMessageTypes().get(25); + getDescriptor().getMessageTypes().get(26); internal_static_google_cloud_dataproc_v1beta2_ReservationAffinity_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_ReservationAffinity_descriptor, @@ -718,6 +755,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { registry.add(com.google.api.AnnotationsProto.http); registry.add(com.google.api.ClientProto.methodSignature); registry.add(com.google.api.ClientProto.oauthScopes); + registry.add(com.google.api.ResourceProto.resourceReference); registry.add(com.google.longrunning.OperationsProto.operationInfo); com.google.protobuf.Descriptors.FileDescriptor.internalUpdateFileDescriptor( descriptor, registry); diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Component.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Component.java index 2ee3ba76..9fed4256 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Component.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Component.java @@ -48,6 +48,16 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * ANACONDA = 5; */ ANACONDA(5), + /** + * + * + *
+   * Docker
+   * 
+ * + * DOCKER = 13; + */ + DOCKER(13), /** * * @@ -58,6 +68,16 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * DRUID = 9; */ DRUID(9), + /** + * + * + *
+   * Flink
+   * 
+ * + * FLINK = 14; + */ + FLINK(14), /** * * @@ -98,6 +118,26 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * PRESTO = 6; */ PRESTO(6), + /** + * + * + *
+   * The Ranger service.
+   * 
+ * + * RANGER = 12; + */ + RANGER(12), + /** + * + * + *
+   * The Solr service.
+   * 
+ * + * SOLR = 10; + */ + SOLR(10), /** * * @@ -141,6 +181,16 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * ANACONDA = 5; */ public static final int ANACONDA_VALUE = 5; + /** + * + * + *
+   * Docker
+   * 
+ * + * DOCKER = 13; + */ + public static final int DOCKER_VALUE = 13; /** * * @@ -151,6 +201,16 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * DRUID = 9; */ public static final int DRUID_VALUE = 9; + /** + * + * + *
+   * Flink
+   * 
+ * + * FLINK = 14; + */ + public static final int FLINK_VALUE = 14; /** * * @@ -191,6 +251,26 @@ public enum Component implements com.google.protobuf.ProtocolMessageEnum { * PRESTO = 6; */ public static final int PRESTO_VALUE = 6; + /** + * + * + *
+   * The Ranger service.
+   * 
+ * + * RANGER = 12; + */ + public static final int RANGER_VALUE = 12; + /** + * + * + *
+   * The Solr service.
+   * 
+ * + * SOLR = 10; + */ + public static final int SOLR_VALUE = 10; /** * * @@ -240,8 +320,12 @@ public static Component forNumber(int value) { return COMPONENT_UNSPECIFIED; case 5: return ANACONDA; + case 13: + return DOCKER; case 9: return DRUID; + case 14: + return FLINK; case 3: return HIVE_WEBHCAT; case 1: @@ -250,6 +334,10 @@ public static Component forNumber(int value) { return KERBEROS; case 6: return PRESTO; + case 12: + return RANGER; + case 10: + return SOLR; case 4: return ZEPPELIN; case 8: diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfig.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfig.java index f96d4d35..8df448f8 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfig.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfig.java @@ -286,7 +286,8 @@ public com.google.protobuf.ByteString getZoneUriBytes() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -316,7 +317,8 @@ public java.lang.String getNetworkUri() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -425,14 +427,14 @@ public boolean getInternalIpOnly() { * *
    * Optional. The [Dataproc service
-   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
    * (also see [VM Data Plane
-   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
    * used by Dataproc cluster VM instances to access Google Cloud Platform
    * services.
    * If not specified, the
    * [Compute Engine default service
-   * account](/compute/docs/access/service-accounts#default_service_account)
+   * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
    * is used.
    * 
* @@ -456,14 +458,14 @@ public java.lang.String getServiceAccount() { * *
    * Optional. The [Dataproc service
-   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
    * (also see [VM Data Plane
-   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
    * used by Dataproc cluster VM instances to access Google Cloud Platform
    * services.
    * If not specified, the
    * [Compute Engine default service
-   * account](/compute/docs/access/service-accounts#default_service_account)
+   * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
    * is used.
    * 
* @@ -594,8 +596,8 @@ public com.google.protobuf.ByteString getServiceAccountScopesBytes(int index) { * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -609,8 +611,8 @@ public com.google.protobuf.ProtocolStringList getTagsList() { * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -624,8 +626,8 @@ public int getTagsCount() { * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -640,8 +642,8 @@ public java.lang.String getTags(int index) { * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -1484,7 +1486,8 @@ public Builder setZoneUriBytes(com.google.protobuf.ByteString value) { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1514,7 +1517,8 @@ public java.lang.String getNetworkUri() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1544,7 +1548,8 @@ public com.google.protobuf.ByteString getNetworkUriBytes() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1573,7 +1578,8 @@ public Builder setNetworkUri(java.lang.String value) { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1598,7 +1604,8 @@ public Builder clearNetworkUri() { * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -1824,14 +1831,14 @@ public Builder clearInternalIpOnly() { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -1855,14 +1862,14 @@ public java.lang.String getServiceAccount() { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -1886,14 +1893,14 @@ public com.google.protobuf.ByteString getServiceAccountBytes() { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -1916,14 +1923,14 @@ public Builder setServiceAccount(java.lang.String value) { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -1942,14 +1949,14 @@ public Builder clearServiceAccount() { * *
      * Optional. The [Dataproc service
-     * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+     * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
      * (also see [VM Data Plane
-     * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+     * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
      * used by Dataproc cluster VM instances to access Google Cloud Platform
      * services.
      * If not specified, the
      * [Compute Engine default service
-     * account](/compute/docs/access/service-accounts#default_service_account)
+     * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
      * is used.
      * 
* @@ -2249,8 +2256,8 @@ private void ensureTagsIsMutable() { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2264,8 +2271,8 @@ public com.google.protobuf.ProtocolStringList getTagsList() { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2279,8 +2286,8 @@ public int getTagsCount() { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2295,8 +2302,8 @@ public java.lang.String getTags(int index) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2311,8 +2318,8 @@ public com.google.protobuf.ByteString getTagsBytes(int index) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2334,8 +2341,8 @@ public Builder setTags(int index, java.lang.String value) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2356,8 +2363,8 @@ public Builder addTags(java.lang.String value) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2375,8 +2382,8 @@ public Builder addAllTags(java.lang.Iterable values) { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; @@ -2393,8 +2400,8 @@ public Builder clearTags() { * * *
-     * The Compute Engine tags to add to all instances (see
-     * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+     * The Compute Engine tags to add to all instances (see [Tagging
+     * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
      * 
* * repeated string tags = 4; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfigOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfigOrBuilder.java index 188e5c20..1c431378 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GceClusterConfigOrBuilder.java @@ -72,7 +72,8 @@ public interface GceClusterConfigOrBuilder * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -92,7 +93,8 @@ public interface GceClusterConfigOrBuilder * communications. Cannot be specified with subnetwork_uri. If neither * `network_uri` nor `subnetwork_uri` is specified, the "default" network of * the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - * [Using Subnetworks](/compute/docs/subnetworks) for more information). + * [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + * more information). * A full URL, partial URI, or short name are valid. Examples: * * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` * * `projects/[project_id]/regions/global/default` @@ -163,14 +165,14 @@ public interface GceClusterConfigOrBuilder * *
    * Optional. The [Dataproc service
-   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
    * (also see [VM Data Plane
-   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
    * used by Dataproc cluster VM instances to access Google Cloud Platform
    * services.
    * If not specified, the
    * [Compute Engine default service
-   * account](/compute/docs/access/service-accounts#default_service_account)
+   * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
    * is used.
    * 
* @@ -184,14 +186,14 @@ public interface GceClusterConfigOrBuilder * *
    * Optional. The [Dataproc service
-   * account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
+   * account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
    * (also see [VM Data Plane
-   * identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
+   * identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
    * used by Dataproc cluster VM instances to access Google Cloud Platform
    * services.
    * If not specified, the
    * [Compute Engine default service
-   * account](/compute/docs/access/service-accounts#default_service_account)
+   * account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account)
    * is used.
    * 
* @@ -300,8 +302,8 @@ public interface GceClusterConfigOrBuilder * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -313,8 +315,8 @@ public interface GceClusterConfigOrBuilder * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -326,8 +328,8 @@ public interface GceClusterConfigOrBuilder * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; @@ -340,8 +342,8 @@ public interface GceClusterConfigOrBuilder * * *
-   * The Compute Engine tags to add to all instances (see
-   * [Tagging instances](/compute/docs/label-or-tag-resources#tags)).
+   * The Compute Engine tags to add to all instances (see [Tagging
+   * instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).
    * 
* * repeated string tags = 4; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GkeClusterConfig.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GkeClusterConfig.java new file mode 100644 index 00000000..a4a16616 --- /dev/null +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GkeClusterConfig.java @@ -0,0 +1,1701 @@ +/* + * Copyright 2020 Google LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dataproc/v1beta2/clusters.proto + +package com.google.cloud.dataproc.v1beta2; + +/** + * + * + *
+ * The GKE config for this cluster.
+ * 
+ * + * Protobuf type {@code google.cloud.dataproc.v1beta2.GkeClusterConfig} + */ +public final class GkeClusterConfig extends com.google.protobuf.GeneratedMessageV3 + implements + // @@protoc_insertion_point(message_implements:google.cloud.dataproc.v1beta2.GkeClusterConfig) + GkeClusterConfigOrBuilder { + private static final long serialVersionUID = 0L; + // Use GkeClusterConfig.newBuilder() to construct. + private GkeClusterConfig(com.google.protobuf.GeneratedMessageV3.Builder builder) { + super(builder); + } + + private GkeClusterConfig() {} + + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance(UnusedPrivateParameter unused) { + return new GkeClusterConfig(); + } + + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet getUnknownFields() { + return this.unknownFields; + } + + private GkeClusterConfig( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + case 10: + { + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .Builder + subBuilder = null; + if (namespacedGkeDeploymentTarget_ != null) { + subBuilder = namespacedGkeDeploymentTarget_.toBuilder(); + } + namespacedGkeDeploymentTarget_ = + input.readMessage( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig + .NamespacedGkeDeploymentTarget.parser(), + extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom(namespacedGkeDeploymentTarget_); + namespacedGkeDeploymentTarget_ = subBuilder.buildPartial(); + } + + break; + } + default: + { + if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException(e).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_descriptor; + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.class, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.Builder.class); + } + + public interface NamespacedGkeDeploymentTargetOrBuilder + extends + // @@protoc_insertion_point(interface_extends:google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) + com.google.protobuf.MessageOrBuilder { + + /** + * + * + *
+     * Optional. The target GKE cluster to deploy to.
+     * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+     * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @return The targetGkeCluster. + */ + java.lang.String getTargetGkeCluster(); + /** + * + * + *
+     * Optional. The target GKE cluster to deploy to.
+     * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+     * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @return The bytes for targetGkeCluster. + */ + com.google.protobuf.ByteString getTargetGkeClusterBytes(); + + /** + * + * + *
+     * Optional. A namespace within the GKE cluster to deploy into.
+     * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The clusterNamespace. + */ + java.lang.String getClusterNamespace(); + /** + * + * + *
+     * Optional. A namespace within the GKE cluster to deploy into.
+     * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The bytes for clusterNamespace. + */ + com.google.protobuf.ByteString getClusterNamespaceBytes(); + } + /** + * + * + *
+   * A full, namespace-isolated deployment target for an existing GKE cluster.
+   * 
+ * + * Protobuf type {@code + * google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget} + */ + public static final class NamespacedGkeDeploymentTarget + extends com.google.protobuf.GeneratedMessageV3 + implements + // @@protoc_insertion_point(message_implements:google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) + NamespacedGkeDeploymentTargetOrBuilder { + private static final long serialVersionUID = 0L; + // Use NamespacedGkeDeploymentTarget.newBuilder() to construct. + private NamespacedGkeDeploymentTarget( + com.google.protobuf.GeneratedMessageV3.Builder builder) { + super(builder); + } + + private NamespacedGkeDeploymentTarget() { + targetGkeCluster_ = ""; + clusterNamespace_ = ""; + } + + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance(UnusedPrivateParameter unused) { + return new NamespacedGkeDeploymentTarget(); + } + + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet getUnknownFields() { + return this.unknownFields; + } + + private NamespacedGkeDeploymentTarget( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + case 10: + { + java.lang.String s = input.readStringRequireUtf8(); + + targetGkeCluster_ = s; + break; + } + case 18: + { + java.lang.String s = input.readStringRequireUtf8(); + + clusterNamespace_ = s; + break; + } + default: + { + if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException(e).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_descriptor; + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .class, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .Builder.class); + } + + public static final int TARGET_GKE_CLUSTER_FIELD_NUMBER = 1; + private volatile java.lang.Object targetGkeCluster_; + /** + * + * + *
+     * Optional. The target GKE cluster to deploy to.
+     * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+     * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @return The targetGkeCluster. + */ + public java.lang.String getTargetGkeCluster() { + java.lang.Object ref = targetGkeCluster_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + targetGkeCluster_ = s; + return s; + } + } + /** + * + * + *
+     * Optional. The target GKE cluster to deploy to.
+     * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+     * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @return The bytes for targetGkeCluster. + */ + public com.google.protobuf.ByteString getTargetGkeClusterBytes() { + java.lang.Object ref = targetGkeCluster_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + targetGkeCluster_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + public static final int CLUSTER_NAMESPACE_FIELD_NUMBER = 2; + private volatile java.lang.Object clusterNamespace_; + /** + * + * + *
+     * Optional. A namespace within the GKE cluster to deploy into.
+     * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The clusterNamespace. + */ + public java.lang.String getClusterNamespace() { + java.lang.Object ref = clusterNamespace_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + clusterNamespace_ = s; + return s; + } + } + /** + * + * + *
+     * Optional. A namespace within the GKE cluster to deploy into.
+     * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The bytes for clusterNamespace. + */ + public com.google.protobuf.ByteString getClusterNamespaceBytes() { + java.lang.Object ref = clusterNamespace_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + clusterNamespace_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + private byte memoizedIsInitialized = -1; + + @java.lang.Override + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; + + memoizedIsInitialized = 1; + return true; + } + + @java.lang.Override + public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { + if (!getTargetGkeClusterBytes().isEmpty()) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, targetGkeCluster_); + } + if (!getClusterNamespaceBytes().isEmpty()) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, clusterNamespace_); + } + unknownFields.writeTo(output); + } + + @java.lang.Override + public int getSerializedSize() { + int size = memoizedSize; + if (size != -1) return size; + + size = 0; + if (!getTargetGkeClusterBytes().isEmpty()) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, targetGkeCluster_); + } + if (!getClusterNamespaceBytes().isEmpty()) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, clusterNamespace_); + } + size += unknownFields.getSerializedSize(); + memoizedSize = size; + return size; + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj + instanceof + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget)) { + return super.equals(obj); + } + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget other = + (com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) obj; + + if (!getTargetGkeCluster().equals(other.getTargetGkeCluster())) return false; + if (!getClusterNamespace().equals(other.getClusterNamespace())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (37 * hash) + TARGET_GKE_CLUSTER_FIELD_NUMBER; + hash = (53 * hash) + getTargetGkeCluster().hashCode(); + hash = (37 * hash) + CLUSTER_NAMESPACE_FIELD_NUMBER; + hash = (53 * hash) + getClusterNamespace().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom(java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom( + java.nio.ByteBuffer data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom(com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom(byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom(java.io.InputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseDelimitedFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom(com.google.protobuf.CodedInputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + @java.lang.Override + public Builder newBuilderForType() { + return newBuilder(); + } + + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } + + public static Builder newBuilder( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + prototype) { + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); + } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * + * + *
+     * A full, namespace-isolated deployment target for an existing GKE cluster.
+     * 
+ * + * Protobuf type {@code + * google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget} + */ + public static final class Builder + extends com.google.protobuf.GeneratedMessageV3.Builder + implements + // @@protoc_insertion_point(builder_implements:google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTargetOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_descriptor; + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .class, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .Builder.class); + } + + // Construct using + // com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders) {} + } + + @java.lang.Override + public Builder clear() { + super.clear(); + targetGkeCluster_ = ""; + + clusterNamespace_ = ""; + + return this; + } + + @java.lang.Override + public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_NamespacedGkeDeploymentTarget_descriptor; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + getDefaultInstanceForType() { + return com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .getDefaultInstance(); + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + build() { + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget result = + buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + buildPartial() { + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget result = + new com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget( + this); + result.targetGkeCluster_ = targetGkeCluster_; + result.clusterNamespace_ = clusterNamespace_; + onBuilt(); + return result; + } + + @java.lang.Override + public Builder clone() { + return super.clone(); + } + + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.setField(field, value); + } + + @java.lang.Override + public Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + + @java.lang.Override + public Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, + java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.addRepeatedField(field, value); + } + + @java.lang.Override + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other + instanceof + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) { + return mergeFrom( + (com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) + other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget other) { + if (other + == com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .getDefaultInstance()) return this; + if (!other.getTargetGkeCluster().isEmpty()) { + targetGkeCluster_ = other.targetGkeCluster_; + onChanged(); + } + if (!other.getClusterNamespace().isEmpty()) { + clusterNamespace_ = other.clusterNamespace_; + onChanged(); + } + this.mergeUnknownFields(other.unknownFields); + onChanged(); + return this; + } + + @java.lang.Override + public final boolean isInitialized() { + return true; + } + + @java.lang.Override + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = + (com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) + e.getUnfinishedMessage(); + throw e.unwrapIOException(); + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + + private java.lang.Object targetGkeCluster_ = ""; + /** + * + * + *
+       * Optional. The target GKE cluster to deploy to.
+       * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+       * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @return The targetGkeCluster. + */ + public java.lang.String getTargetGkeCluster() { + java.lang.Object ref = targetGkeCluster_; + if (!(ref instanceof java.lang.String)) { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + targetGkeCluster_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * + * + *
+       * Optional. The target GKE cluster to deploy to.
+       * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+       * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @return The bytes for targetGkeCluster. + */ + public com.google.protobuf.ByteString getTargetGkeClusterBytes() { + java.lang.Object ref = targetGkeCluster_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + targetGkeCluster_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * + * + *
+       * Optional. The target GKE cluster to deploy to.
+       * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+       * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @param value The targetGkeCluster to set. + * @return This builder for chaining. + */ + public Builder setTargetGkeCluster(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + + targetGkeCluster_ = value; + onChanged(); + return this; + } + /** + * + * + *
+       * Optional. The target GKE cluster to deploy to.
+       * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+       * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @return This builder for chaining. + */ + public Builder clearTargetGkeCluster() { + + targetGkeCluster_ = getDefaultInstance().getTargetGkeCluster(); + onChanged(); + return this; + } + /** + * + * + *
+       * Optional. The target GKE cluster to deploy to.
+       * Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}'
+       * 
+ * + * + * string target_gke_cluster = 1 [(.google.api.field_behavior) = OPTIONAL, (.google.api.resource_reference) = { ... } + * + * + * @param value The bytes for targetGkeCluster to set. + * @return This builder for chaining. + */ + public Builder setTargetGkeClusterBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + + targetGkeCluster_ = value; + onChanged(); + return this; + } + + private java.lang.Object clusterNamespace_ = ""; + /** + * + * + *
+       * Optional. A namespace within the GKE cluster to deploy into.
+       * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The clusterNamespace. + */ + public java.lang.String getClusterNamespace() { + java.lang.Object ref = clusterNamespace_; + if (!(ref instanceof java.lang.String)) { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + clusterNamespace_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * + * + *
+       * Optional. A namespace within the GKE cluster to deploy into.
+       * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The bytes for clusterNamespace. + */ + public com.google.protobuf.ByteString getClusterNamespaceBytes() { + java.lang.Object ref = clusterNamespace_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + clusterNamespace_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * + * + *
+       * Optional. A namespace within the GKE cluster to deploy into.
+       * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param value The clusterNamespace to set. + * @return This builder for chaining. + */ + public Builder setClusterNamespace(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + + clusterNamespace_ = value; + onChanged(); + return this; + } + /** + * + * + *
+       * Optional. A namespace within the GKE cluster to deploy into.
+       * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return This builder for chaining. + */ + public Builder clearClusterNamespace() { + + clusterNamespace_ = getDefaultInstance().getClusterNamespace(); + onChanged(); + return this; + } + /** + * + * + *
+       * Optional. A namespace within the GKE cluster to deploy into.
+       * 
+ * + * string cluster_namespace = 2 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param value The bytes for clusterNamespace to set. + * @return This builder for chaining. + */ + public Builder setClusterNamespaceBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + + clusterNamespace_ = value; + onChanged(); + return this; + } + + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + + // @@protoc_insertion_point(builder_scope:google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) + } + + // @@protoc_insertion_point(class_scope:google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget) + private static final com.google.cloud.dataproc.v1beta2.GkeClusterConfig + .NamespacedGkeDeploymentTarget + DEFAULT_INSTANCE; + + static { + DEFAULT_INSTANCE = + new com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget(); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + private static final com.google.protobuf.Parser PARSER = + new com.google.protobuf.AbstractParser() { + @java.lang.Override + public NamespacedGkeDeploymentTarget parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new NamespacedGkeDeploymentTarget(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } + } + + public static final int NAMESPACED_GKE_DEPLOYMENT_TARGET_FIELD_NUMBER = 1; + private com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + namespacedGkeDeploymentTarget_; + /** + * + * + *
+   * Optional. A target for the deployment.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the namespacedGkeDeploymentTarget field is set. + */ + public boolean hasNamespacedGkeDeploymentTarget() { + return namespacedGkeDeploymentTarget_ != null; + } + /** + * + * + *
+   * Optional. A target for the deployment.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The namespacedGkeDeploymentTarget. + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + getNamespacedGkeDeploymentTarget() { + return namespacedGkeDeploymentTarget_ == null + ? com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .getDefaultInstance() + : namespacedGkeDeploymentTarget_; + } + /** + * + * + *
+   * Optional. A target for the deployment.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTargetOrBuilder + getNamespacedGkeDeploymentTargetOrBuilder() { + return getNamespacedGkeDeploymentTarget(); + } + + private byte memoizedIsInitialized = -1; + + @java.lang.Override + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; + + memoizedIsInitialized = 1; + return true; + } + + @java.lang.Override + public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { + if (namespacedGkeDeploymentTarget_ != null) { + output.writeMessage(1, getNamespacedGkeDeploymentTarget()); + } + unknownFields.writeTo(output); + } + + @java.lang.Override + public int getSerializedSize() { + int size = memoizedSize; + if (size != -1) return size; + + size = 0; + if (namespacedGkeDeploymentTarget_ != null) { + size += + com.google.protobuf.CodedOutputStream.computeMessageSize( + 1, getNamespacedGkeDeploymentTarget()); + } + size += unknownFields.getSerializedSize(); + memoizedSize = size; + return size; + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof com.google.cloud.dataproc.v1beta2.GkeClusterConfig)) { + return super.equals(obj); + } + com.google.cloud.dataproc.v1beta2.GkeClusterConfig other = + (com.google.cloud.dataproc.v1beta2.GkeClusterConfig) obj; + + if (hasNamespacedGkeDeploymentTarget() != other.hasNamespacedGkeDeploymentTarget()) + return false; + if (hasNamespacedGkeDeploymentTarget()) { + if (!getNamespacedGkeDeploymentTarget().equals(other.getNamespacedGkeDeploymentTarget())) + return false; + } + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + if (hasNamespacedGkeDeploymentTarget()) { + hash = (37 * hash) + NAMESPACED_GKE_DEPLOYMENT_TARGET_FIELD_NUMBER; + hash = (53 * hash) + getNamespacedGkeDeploymentTarget().hashCode(); + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + java.nio.ByteBuffer data) throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + java.nio.ByteBuffer data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + java.io.InputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseDelimitedFrom( + java.io.InputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseDelimitedFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + com.google.protobuf.CodedInputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + @java.lang.Override + public Builder newBuilderForType() { + return newBuilder(); + } + + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } + + public static Builder newBuilder(com.google.cloud.dataproc.v1beta2.GkeClusterConfig prototype) { + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); + } + + @java.lang.Override + protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * + * + *
+   * The GKE config for this cluster.
+   * 
+ * + * Protobuf type {@code google.cloud.dataproc.v1beta2.GkeClusterConfig} + */ + public static final class Builder extends com.google.protobuf.GeneratedMessageV3.Builder + implements + // @@protoc_insertion_point(builder_implements:google.cloud.dataproc.v1beta2.GkeClusterConfig) + com.google.cloud.dataproc.v1beta2.GkeClusterConfigOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_descriptor; + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.class, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.Builder.class); + } + + // Construct using com.google.cloud.dataproc.v1beta2.GkeClusterConfig.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders) {} + } + + @java.lang.Override + public Builder clear() { + super.clear(); + if (namespacedGkeDeploymentTargetBuilder_ == null) { + namespacedGkeDeploymentTarget_ = null; + } else { + namespacedGkeDeploymentTarget_ = null; + namespacedGkeDeploymentTargetBuilder_ = null; + } + return this; + } + + @java.lang.Override + public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { + return com.google.cloud.dataproc.v1beta2.ClustersProto + .internal_static_google_cloud_dataproc_v1beta2_GkeClusterConfig_descriptor; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig getDefaultInstanceForType() { + return com.google.cloud.dataproc.v1beta2.GkeClusterConfig.getDefaultInstance(); + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig build() { + com.google.cloud.dataproc.v1beta2.GkeClusterConfig result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig buildPartial() { + com.google.cloud.dataproc.v1beta2.GkeClusterConfig result = + new com.google.cloud.dataproc.v1beta2.GkeClusterConfig(this); + if (namespacedGkeDeploymentTargetBuilder_ == null) { + result.namespacedGkeDeploymentTarget_ = namespacedGkeDeploymentTarget_; + } else { + result.namespacedGkeDeploymentTarget_ = namespacedGkeDeploymentTargetBuilder_.build(); + } + onBuilt(); + return result; + } + + @java.lang.Override + public Builder clone() { + return super.clone(); + } + + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.setField(field, value); + } + + @java.lang.Override + public Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + + @java.lang.Override + public Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.addRepeatedField(field, value); + } + + @java.lang.Override + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof com.google.cloud.dataproc.v1beta2.GkeClusterConfig) { + return mergeFrom((com.google.cloud.dataproc.v1beta2.GkeClusterConfig) other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(com.google.cloud.dataproc.v1beta2.GkeClusterConfig other) { + if (other == com.google.cloud.dataproc.v1beta2.GkeClusterConfig.getDefaultInstance()) + return this; + if (other.hasNamespacedGkeDeploymentTarget()) { + mergeNamespacedGkeDeploymentTarget(other.getNamespacedGkeDeploymentTarget()); + } + this.mergeUnknownFields(other.unknownFields); + onChanged(); + return this; + } + + @java.lang.Override + public final boolean isInitialized() { + return true; + } + + @java.lang.Override + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.cloud.dataproc.v1beta2.GkeClusterConfig parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = + (com.google.cloud.dataproc.v1beta2.GkeClusterConfig) e.getUnfinishedMessage(); + throw e.unwrapIOException(); + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + + private com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + namespacedGkeDeploymentTarget_; + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .Builder, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig + .NamespacedGkeDeploymentTargetOrBuilder> + namespacedGkeDeploymentTargetBuilder_; + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the namespacedGkeDeploymentTarget field is set. + */ + public boolean hasNamespacedGkeDeploymentTarget() { + return namespacedGkeDeploymentTargetBuilder_ != null + || namespacedGkeDeploymentTarget_ != null; + } + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The namespacedGkeDeploymentTarget. + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + getNamespacedGkeDeploymentTarget() { + if (namespacedGkeDeploymentTargetBuilder_ == null) { + return namespacedGkeDeploymentTarget_ == null + ? com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .getDefaultInstance() + : namespacedGkeDeploymentTarget_; + } else { + return namespacedGkeDeploymentTargetBuilder_.getMessage(); + } + } + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setNamespacedGkeDeploymentTarget( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget value) { + if (namespacedGkeDeploymentTargetBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + namespacedGkeDeploymentTarget_ = value; + onChanged(); + } else { + namespacedGkeDeploymentTargetBuilder_.setMessage(value); + } + + return this; + } + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setNamespacedGkeDeploymentTarget( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget.Builder + builderForValue) { + if (namespacedGkeDeploymentTargetBuilder_ == null) { + namespacedGkeDeploymentTarget_ = builderForValue.build(); + onChanged(); + } else { + namespacedGkeDeploymentTargetBuilder_.setMessage(builderForValue.build()); + } + + return this; + } + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder mergeNamespacedGkeDeploymentTarget( + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget value) { + if (namespacedGkeDeploymentTargetBuilder_ == null) { + if (namespacedGkeDeploymentTarget_ != null) { + namespacedGkeDeploymentTarget_ = + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .newBuilder(namespacedGkeDeploymentTarget_) + .mergeFrom(value) + .buildPartial(); + } else { + namespacedGkeDeploymentTarget_ = value; + } + onChanged(); + } else { + namespacedGkeDeploymentTargetBuilder_.mergeFrom(value); + } + + return this; + } + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder clearNamespacedGkeDeploymentTarget() { + if (namespacedGkeDeploymentTargetBuilder_ == null) { + namespacedGkeDeploymentTarget_ = null; + onChanged(); + } else { + namespacedGkeDeploymentTarget_ = null; + namespacedGkeDeploymentTargetBuilder_ = null; + } + + return this; + } + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget.Builder + getNamespacedGkeDeploymentTargetBuilder() { + + onChanged(); + return getNamespacedGkeDeploymentTargetFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTargetOrBuilder + getNamespacedGkeDeploymentTargetOrBuilder() { + if (namespacedGkeDeploymentTargetBuilder_ != null) { + return namespacedGkeDeploymentTargetBuilder_.getMessageOrBuilder(); + } else { + return namespacedGkeDeploymentTarget_ == null + ? com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .getDefaultInstance() + : namespacedGkeDeploymentTarget_; + } + } + /** + * + * + *
+     * Optional. A target for the deployment.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .Builder, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig + .NamespacedGkeDeploymentTargetOrBuilder> + getNamespacedGkeDeploymentTargetFieldBuilder() { + if (namespacedGkeDeploymentTargetBuilder_ == null) { + namespacedGkeDeploymentTargetBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + .Builder, + com.google.cloud.dataproc.v1beta2.GkeClusterConfig + .NamespacedGkeDeploymentTargetOrBuilder>( + getNamespacedGkeDeploymentTarget(), getParentForChildren(), isClean()); + namespacedGkeDeploymentTarget_ = null; + } + return namespacedGkeDeploymentTargetBuilder_; + } + + @java.lang.Override + public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + + // @@protoc_insertion_point(builder_scope:google.cloud.dataproc.v1beta2.GkeClusterConfig) + } + + // @@protoc_insertion_point(class_scope:google.cloud.dataproc.v1beta2.GkeClusterConfig) + private static final com.google.cloud.dataproc.v1beta2.GkeClusterConfig DEFAULT_INSTANCE; + + static { + DEFAULT_INSTANCE = new com.google.cloud.dataproc.v1beta2.GkeClusterConfig(); + } + + public static com.google.cloud.dataproc.v1beta2.GkeClusterConfig getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + private static final com.google.protobuf.Parser PARSER = + new com.google.protobuf.AbstractParser() { + @java.lang.Override + public GkeClusterConfig parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new GkeClusterConfig(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.GkeClusterConfig getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } +} diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GkeClusterConfigOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GkeClusterConfigOrBuilder.java new file mode 100644 index 00000000..231d5731 --- /dev/null +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/GkeClusterConfigOrBuilder.java @@ -0,0 +1,68 @@ +/* + * Copyright 2020 Google LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dataproc/v1beta2/clusters.proto + +package com.google.cloud.dataproc.v1beta2; + +public interface GkeClusterConfigOrBuilder + extends + // @@protoc_insertion_point(interface_extends:google.cloud.dataproc.v1beta2.GkeClusterConfig) + com.google.protobuf.MessageOrBuilder { + + /** + * + * + *
+   * Optional. A target for the deployment.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the namespacedGkeDeploymentTarget field is set. + */ + boolean hasNamespacedGkeDeploymentTarget(); + /** + * + * + *
+   * Optional. A target for the deployment.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The namespacedGkeDeploymentTarget. + */ + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget + getNamespacedGkeDeploymentTarget(); + /** + * + * + *
+   * Optional. A target for the deployment.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + com.google.cloud.dataproc.v1beta2.GkeClusterConfig.NamespacedGkeDeploymentTargetOrBuilder + getNamespacedGkeDeploymentTargetOrBuilder(); +} diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfig.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfig.java index 244559b4..30b6cd2d 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfig.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfig.java @@ -292,9 +292,18 @@ public com.google.protobuf.ByteString getInstanceNamesBytes(int index) { * * *
-   * Optional. The Compute Engine image resource used for cluster
-   * instances. It can be specified or may be inferred from
-   * `SoftwareConfig.image_version`.
+   * Optional. The Compute Engine image resource used for cluster instances.
+   * The URI can represent an image or image family.
+   * Image examples:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+   * * `projects/[project_id]/global/images/[image-id]`
+   * * `image-id`
+   * Image family examples. Dataproc will use the most recent
+   * image from the family:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * If the URI is unspecified, it will be inferred from
+   * `SoftwareConfig.image_version` or the system default.
    * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -316,9 +325,18 @@ public java.lang.String getImageUri() { * * *
-   * Optional. The Compute Engine image resource used for cluster
-   * instances. It can be specified or may be inferred from
-   * `SoftwareConfig.image_version`.
+   * Optional. The Compute Engine image resource used for cluster instances.
+   * The URI can represent an image or image family.
+   * Image examples:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+   * * `projects/[project_id]/global/images/[image-id]`
+   * * `image-id`
+   * Image family examples. Dataproc will use the most recent
+   * image from the family:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * If the URI is unspecified, it will be inferred from
+   * `SoftwareConfig.image_version` or the system default.
    * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -350,7 +368,7 @@ public com.google.protobuf.ByteString getImageUriBytes() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -381,7 +399,7 @@ public java.lang.String getMachineTypeUri() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -459,11 +477,11 @@ public com.google.cloud.dataproc.v1beta2.DiskConfigOrBuilder getDiskConfigOrBuil * * *
-   * Optional. Specifies that this instance group contains preemptible
+   * Output only. Specifies that this instance group contains preemptible
    * instances.
    * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @return The isPreemptible. */ @@ -616,8 +634,8 @@ public com.google.cloud.dataproc.v1beta2.AcceleratorConfigOrBuilder getAccelerat * *
    * Specifies the minimum cpu platform for the Instance Group.
-   * See [Dataproc&rarr;Minimum CPU Platform]
-   * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+   * See [Dataproc -> Minimum CPU
+   * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* * string min_cpu_platform = 9; @@ -640,8 +658,8 @@ public java.lang.String getMinCpuPlatform() { * *
    * Specifies the minimum cpu platform for the Instance Group.
-   * See [Dataproc&rarr;Minimum CPU Platform]
-   * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+   * See [Dataproc -> Minimum CPU
+   * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* * string min_cpu_platform = 9; @@ -1425,9 +1443,18 @@ public Builder addInstanceNamesBytes(com.google.protobuf.ByteString value) { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1449,9 +1476,18 @@ public java.lang.String getImageUri() { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1473,9 +1509,18 @@ public com.google.protobuf.ByteString getImageUriBytes() { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1496,9 +1541,18 @@ public Builder setImageUri(java.lang.String value) { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1515,9 +1569,18 @@ public Builder clearImageUri() { * * *
-     * Optional. The Compute Engine image resource used for cluster
-     * instances. It can be specified or may be inferred from
-     * `SoftwareConfig.image_version`.
+     * Optional. The Compute Engine image resource used for cluster instances.
+     * The URI can represent an image or image family.
+     * Image examples:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+     * * `projects/[project_id]/global/images/[image-id]`
+     * * `image-id`
+     * Image family examples. Dataproc will use the most recent
+     * image from the family:
+     * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+     * If the URI is unspecified, it will be inferred from
+     * `SoftwareConfig.image_version` or the system default.
      * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -1548,7 +1611,7 @@ public Builder setImageUriBytes(com.google.protobuf.ByteString value) { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1579,7 +1642,7 @@ public java.lang.String getMachineTypeUri() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1610,7 +1673,7 @@ public com.google.protobuf.ByteString getMachineTypeUriBytes() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1640,7 +1703,7 @@ public Builder setMachineTypeUri(java.lang.String value) { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1666,7 +1729,7 @@ public Builder clearMachineTypeUri() { * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -1896,11 +1959,11 @@ public com.google.cloud.dataproc.v1beta2.DiskConfigOrBuilder getDiskConfigOrBuil * * *
-     * Optional. Specifies that this instance group contains preemptible
+     * Output only. Specifies that this instance group contains preemptible
      * instances.
      * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @return The isPreemptible. */ @@ -1911,11 +1974,11 @@ public boolean getIsPreemptible() { * * *
-     * Optional. Specifies that this instance group contains preemptible
+     * Output only. Specifies that this instance group contains preemptible
      * instances.
      * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @param value The isPreemptible to set. * @return This builder for chaining. @@ -1930,11 +1993,11 @@ public Builder setIsPreemptible(boolean value) { * * *
-     * Optional. Specifies that this instance group contains preemptible
+     * Output only. Specifies that this instance group contains preemptible
      * instances.
      * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @return This builder for chaining. */ @@ -2590,8 +2653,8 @@ public com.google.cloud.dataproc.v1beta2.AcceleratorConfig.Builder addAccelerato * *
      * Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9; @@ -2614,8 +2677,8 @@ public java.lang.String getMinCpuPlatform() { * *
      * Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9; @@ -2638,8 +2701,8 @@ public com.google.protobuf.ByteString getMinCpuPlatformBytes() { * *
      * Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9; @@ -2661,8 +2724,8 @@ public Builder setMinCpuPlatform(java.lang.String value) { * *
      * Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9; @@ -2680,8 +2743,8 @@ public Builder clearMinCpuPlatform() { * *
      * Specifies the minimum cpu platform for the Instance Group.
-     * See [Dataproc&rarr;Minimum CPU Platform]
-     * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+     * See [Dataproc -> Minimum CPU
+     * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
      * 
* * string min_cpu_platform = 9; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfigOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfigOrBuilder.java index 202e9b2c..ae533db6 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/InstanceGroupConfigOrBuilder.java @@ -96,9 +96,18 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Optional. The Compute Engine image resource used for cluster
-   * instances. It can be specified or may be inferred from
-   * `SoftwareConfig.image_version`.
+   * Optional. The Compute Engine image resource used for cluster instances.
+   * The URI can represent an image or image family.
+   * Image examples:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+   * * `projects/[project_id]/global/images/[image-id]`
+   * * `image-id`
+   * Image family examples. Dataproc will use the most recent
+   * image from the family:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * If the URI is unspecified, it will be inferred from
+   * `SoftwareConfig.image_version` or the system default.
    * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -110,9 +119,18 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Optional. The Compute Engine image resource used for cluster
-   * instances. It can be specified or may be inferred from
-   * `SoftwareConfig.image_version`.
+   * Optional. The Compute Engine image resource used for cluster instances.
+   * The URI can represent an image or image family.
+   * Image examples:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]`
+   * * `projects/[project_id]/global/images/[image-id]`
+   * * `image-id`
+   * Image family examples. Dataproc will use the most recent
+   * image from the family:
+   * * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * * `projects/[project_id]/global/images/family/[custom-image-family-name]`
+   * If the URI is unspecified, it will be inferred from
+   * `SoftwareConfig.image_version` or the system default.
    * 
* * string image_uri = 3 [(.google.api.field_behavior) = OPTIONAL]; @@ -132,7 +150,7 @@ public interface InstanceGroupConfigOrBuilder * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -153,7 +171,7 @@ public interface InstanceGroupConfigOrBuilder * * `n1-standard-2` * **Auto Zone Exception**: If you are using the Dataproc * [Auto Zone - * Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + * Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) * feature, you must use the short name of the machine type * resource, for example, `n1-standard-2`. * @@ -209,11 +227,11 @@ public interface InstanceGroupConfigOrBuilder * * *
-   * Optional. Specifies that this instance group contains preemptible
+   * Output only. Specifies that this instance group contains preemptible
    * instances.
    * 
* - * bool is_preemptible = 6 [(.google.api.field_behavior) = OPTIONAL]; + * bool is_preemptible = 6 [(.google.api.field_behavior) = OUTPUT_ONLY]; * * @return The isPreemptible. */ @@ -338,8 +356,8 @@ public interface InstanceGroupConfigOrBuilder * *
    * Specifies the minimum cpu platform for the Instance Group.
-   * See [Dataproc&rarr;Minimum CPU Platform]
-   * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+   * See [Dataproc -> Minimum CPU
+   * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* * string min_cpu_platform = 9; @@ -352,8 +370,8 @@ public interface InstanceGroupConfigOrBuilder * *
    * Specifies the minimum cpu platform for the Instance Group.
-   * See [Dataproc&rarr;Minimum CPU Platform]
-   * (/dataproc/docs/concepts/compute/dataproc-min-cpu).
+   * See [Dataproc -> Minimum CPU
+   * Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu).
    * 
* * string min_cpu_platform = 9; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Job.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Job.java index d3d27256..d85cbfe3 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Job.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/Job.java @@ -318,6 +318,27 @@ private Job( jobUuid_ = s; break; } + case 186: + { + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder subBuilder = null; + if (typeJobCase_ == 23) { + subBuilder = ((com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_).toBuilder(); + } + typeJob_ = + input.readMessage( + com.google.cloud.dataproc.v1beta2.PrestoJob.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom((com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_); + typeJob_ = subBuilder.buildPartial(); + } + typeJobCase_ = 23; + break; + } + case 192: + { + done_ = input.readBool(); + break; + } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { @@ -383,6 +404,7 @@ public enum TypeJobCase PIG_JOB(7), SPARK_R_JOB(21), SPARK_SQL_JOB(12), + PRESTO_JOB(23), TYPEJOB_NOT_SET(0); private final int value; @@ -415,6 +437,8 @@ public static TypeJobCase forNumber(int value) { return SPARK_R_JOB; case 12: return SPARK_SQL_JOB; + case 23: + return PRESTO_JOB; case 0: return TYPEJOB_NOT_SET; default: @@ -547,7 +571,15 @@ public com.google.cloud.dataproc.v1beta2.JobPlacementOrBuilder getPlacementOrBui public static final int HADOOP_JOB_FIELD_NUMBER = 3; /** - * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; + * + * + *
+   * Optional. Job is a Hadoop job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the hadoopJob field is set. */ @@ -555,7 +587,15 @@ public boolean hasHadoopJob() { return typeJobCase_ == 3; } /** - * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; + * + * + *
+   * Optional. Job is a Hadoop job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The hadoopJob. */ @@ -565,7 +605,17 @@ public com.google.cloud.dataproc.v1beta2.HadoopJob getHadoopJob() { } return com.google.cloud.dataproc.v1beta2.HadoopJob.getDefaultInstance(); } - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ + /** + * + * + *
+   * Optional. Job is a Hadoop job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.HadoopJobOrBuilder getHadoopJobOrBuilder() { if (typeJobCase_ == 3) { return (com.google.cloud.dataproc.v1beta2.HadoopJob) typeJob_; @@ -575,7 +625,15 @@ public com.google.cloud.dataproc.v1beta2.HadoopJobOrBuilder getHadoopJobOrBuilde public static final int SPARK_JOB_FIELD_NUMBER = 4; /** - * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; + * + * + *
+   * Optional. Job is a Spark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkJob field is set. */ @@ -583,7 +641,15 @@ public boolean hasSparkJob() { return typeJobCase_ == 4; } /** - * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; + * + * + *
+   * Optional. Job is a Spark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkJob. */ @@ -593,7 +659,17 @@ public com.google.cloud.dataproc.v1beta2.SparkJob getSparkJob() { } return com.google.cloud.dataproc.v1beta2.SparkJob.getDefaultInstance(); } - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+   * Optional. Job is a Spark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkJobOrBuilder getSparkJobOrBuilder() { if (typeJobCase_ == 4) { return (com.google.cloud.dataproc.v1beta2.SparkJob) typeJob_; @@ -603,7 +679,15 @@ public com.google.cloud.dataproc.v1beta2.SparkJobOrBuilder getSparkJobOrBuilder( public static final int PYSPARK_JOB_FIELD_NUMBER = 5; /** - * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; + * + * + *
+   * Optional. Job is a PySpark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the pysparkJob field is set. */ @@ -611,7 +695,15 @@ public boolean hasPysparkJob() { return typeJobCase_ == 5; } /** - * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; + * + * + *
+   * Optional. Job is a PySpark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The pysparkJob. */ @@ -621,7 +713,17 @@ public com.google.cloud.dataproc.v1beta2.PySparkJob getPysparkJob() { } return com.google.cloud.dataproc.v1beta2.PySparkJob.getDefaultInstance(); } - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+   * Optional. Job is a PySpark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.PySparkJobOrBuilder getPysparkJobOrBuilder() { if (typeJobCase_ == 5) { return (com.google.cloud.dataproc.v1beta2.PySparkJob) typeJob_; @@ -631,7 +733,15 @@ public com.google.cloud.dataproc.v1beta2.PySparkJobOrBuilder getPysparkJobOrBuil public static final int HIVE_JOB_FIELD_NUMBER = 6; /** - * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; + * + * + *
+   * Optional. Job is a Hive job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the hiveJob field is set. */ @@ -639,7 +749,15 @@ public boolean hasHiveJob() { return typeJobCase_ == 6; } /** - * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; + * + * + *
+   * Optional. Job is a Hive job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The hiveJob. */ @@ -649,7 +767,17 @@ public com.google.cloud.dataproc.v1beta2.HiveJob getHiveJob() { } return com.google.cloud.dataproc.v1beta2.HiveJob.getDefaultInstance(); } - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+   * Optional. Job is a Hive job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.HiveJobOrBuilder getHiveJobOrBuilder() { if (typeJobCase_ == 6) { return (com.google.cloud.dataproc.v1beta2.HiveJob) typeJob_; @@ -659,7 +787,15 @@ public com.google.cloud.dataproc.v1beta2.HiveJobOrBuilder getHiveJobOrBuilder() public static final int PIG_JOB_FIELD_NUMBER = 7; /** - * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; + * + * + *
+   * Optional. Job is a Pig job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the pigJob field is set. */ @@ -667,7 +803,15 @@ public boolean hasPigJob() { return typeJobCase_ == 7; } /** - * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; + * + * + *
+   * Optional. Job is a Pig job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The pigJob. */ @@ -677,7 +821,17 @@ public com.google.cloud.dataproc.v1beta2.PigJob getPigJob() { } return com.google.cloud.dataproc.v1beta2.PigJob.getDefaultInstance(); } - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+   * Optional. Job is a Pig job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder() { if (typeJobCase_ == 7) { return (com.google.cloud.dataproc.v1beta2.PigJob) typeJob_; @@ -687,7 +841,15 @@ public com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder() { public static final int SPARK_R_JOB_FIELD_NUMBER = 21; /** - * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; + * + * + *
+   * Optional. Job is a SparkR job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkRJob field is set. */ @@ -695,7 +857,15 @@ public boolean hasSparkRJob() { return typeJobCase_ == 21; } /** - * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; + * + * + *
+   * Optional. Job is a SparkR job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkRJob. */ @@ -705,7 +875,17 @@ public com.google.cloud.dataproc.v1beta2.SparkRJob getSparkRJob() { } return com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); } - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+   * Optional. Job is a SparkR job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilder() { if (typeJobCase_ == 21) { return (com.google.cloud.dataproc.v1beta2.SparkRJob) typeJob_; @@ -715,7 +895,15 @@ public com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilde public static final int SPARK_SQL_JOB_FIELD_NUMBER = 12; /** - * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; + * + * + *
+   * Optional. Job is a SparkSql job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkSqlJob field is set. */ @@ -723,7 +911,15 @@ public boolean hasSparkSqlJob() { return typeJobCase_ == 12; } /** - * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; + * + * + *
+   * Optional. Job is a SparkSql job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkSqlJob. */ @@ -733,7 +929,17 @@ public com.google.cloud.dataproc.v1beta2.SparkSqlJob getSparkSqlJob() { } return com.google.cloud.dataproc.v1beta2.SparkSqlJob.getDefaultInstance(); } - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+   * Optional. Job is a SparkSql job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBuilder() { if (typeJobCase_ == 12) { return (com.google.cloud.dataproc.v1beta2.SparkSqlJob) typeJob_; @@ -741,6 +947,60 @@ public com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBu return com.google.cloud.dataproc.v1beta2.SparkSqlJob.getDefaultInstance(); } + public static final int PRESTO_JOB_FIELD_NUMBER = 23; + /** + * + * + *
+   * Optional. Job is a Presto job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the prestoJob field is set. + */ + public boolean hasPrestoJob() { + return typeJobCase_ == 23; + } + /** + * + * + *
+   * Optional. Job is a Presto job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The prestoJob. + */ + public com.google.cloud.dataproc.v1beta2.PrestoJob getPrestoJob() { + if (typeJobCase_ == 23) { + return (com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_; + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + /** + * + * + *
+   * Optional. Job is a Presto job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder getPrestoJobOrBuilder() { + if (typeJobCase_ == 23) { + return (com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_; + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + public static final int STATUS_FIELD_NUMBER = 8; private com.google.cloud.dataproc.v1beta2.JobStatus status_; /** @@ -1328,6 +1588,26 @@ public com.google.protobuf.ByteString getJobUuidBytes() { } } + public static final int DONE_FIELD_NUMBER = 24; + private boolean done_; + /** + * + * + *
+   * Output only. Indicates whether the job is completed. If the value is `false`,
+   * the job is still in progress. If `true`, the job is completed, and
+   * `status.state` field will indicate if it was successful, failed,
+   * or cancelled.
+   * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The done. + */ + public boolean getDone() { + return done_; + } + private byte memoizedIsInitialized = -1; @java.lang.Override @@ -1395,6 +1675,12 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io if (!getJobUuidBytes().isEmpty()) { com.google.protobuf.GeneratedMessageV3.writeString(output, 22, jobUuid_); } + if (typeJobCase_ == 23) { + output.writeMessage(23, (com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_); + } + if (done_ != false) { + output.writeBool(24, done_); + } unknownFields.writeTo(output); } @@ -1480,6 +1766,14 @@ public int getSerializedSize() { if (!getJobUuidBytes().isEmpty()) { size += com.google.protobuf.GeneratedMessageV3.computeStringSize(22, jobUuid_); } + if (typeJobCase_ == 23) { + size += + com.google.protobuf.CodedOutputStream.computeMessageSize( + 23, (com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_); + } + if (done_ != false) { + size += com.google.protobuf.CodedOutputStream.computeBoolSize(24, done_); + } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; @@ -1518,6 +1812,7 @@ public boolean equals(final java.lang.Object obj) { if (!getScheduling().equals(other.getScheduling())) return false; } if (!getJobUuid().equals(other.getJobUuid())) return false; + if (getDone() != other.getDone()) return false; if (!getTypeJobCase().equals(other.getTypeJobCase())) return false; switch (typeJobCase_) { case 3: @@ -1541,6 +1836,9 @@ public boolean equals(final java.lang.Object obj) { case 12: if (!getSparkSqlJob().equals(other.getSparkSqlJob())) return false; break; + case 23: + if (!getPrestoJob().equals(other.getPrestoJob())) return false; + break; case 0: default: } @@ -1591,6 +1889,8 @@ public int hashCode() { } hash = (37 * hash) + JOB_UUID_FIELD_NUMBER; hash = (53 * hash) + getJobUuid().hashCode(); + hash = (37 * hash) + DONE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(getDone()); switch (typeJobCase_) { case 3: hash = (37 * hash) + HADOOP_JOB_FIELD_NUMBER; @@ -1620,6 +1920,10 @@ public int hashCode() { hash = (37 * hash) + SPARK_SQL_JOB_FIELD_NUMBER; hash = (53 * hash) + getSparkSqlJob().hashCode(); break; + case 23: + hash = (37 * hash) + PRESTO_JOB_FIELD_NUMBER; + hash = (53 * hash) + getPrestoJob().hashCode(); + break; case 0: default: } @@ -1835,6 +2139,8 @@ public Builder clear() { } jobUuid_ = ""; + done_ = false; + typeJobCase_ = 0; typeJob_ = null; return this; @@ -1924,6 +2230,13 @@ public com.google.cloud.dataproc.v1beta2.Job buildPartial() { result.typeJob_ = sparkSqlJobBuilder_.build(); } } + if (typeJobCase_ == 23) { + if (prestoJobBuilder_ == null) { + result.typeJob_ = typeJob_; + } else { + result.typeJob_ = prestoJobBuilder_.build(); + } + } if (statusBuilder_ == null) { result.status_ = status_; } else { @@ -1958,6 +2271,7 @@ public com.google.cloud.dataproc.v1beta2.Job buildPartial() { result.scheduling_ = schedulingBuilder_.build(); } result.jobUuid_ = jobUuid_; + result.done_ = done_; result.typeJobCase_ = typeJobCase_; onBuilt(); return result; @@ -2091,6 +2405,9 @@ public Builder mergeFrom(com.google.cloud.dataproc.v1beta2.Job other) { jobUuid_ = other.jobUuid_; onChanged(); } + if (other.getDone() != false) { + setDone(other.getDone()); + } switch (other.getTypeJobCase()) { case HADOOP_JOB: { @@ -2127,6 +2444,11 @@ public Builder mergeFrom(com.google.cloud.dataproc.v1beta2.Job other) { mergeSparkSqlJob(other.getSparkSqlJob()); break; } + case PRESTO_JOB: + { + mergePrestoJob(other.getPrestoJob()); + break; + } case TYPEJOB_NOT_SET: { break; @@ -2627,7 +2949,15 @@ public com.google.cloud.dataproc.v1beta2.JobPlacementOrBuilder getPlacementOrBui com.google.cloud.dataproc.v1beta2.HadoopJobOrBuilder> hadoopJobBuilder_; /** - * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the hadoopJob field is set. */ @@ -2635,7 +2965,15 @@ public boolean hasHadoopJob() { return typeJobCase_ == 3; } /** - * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The hadoopJob. */ @@ -2652,21 +2990,41 @@ public com.google.cloud.dataproc.v1beta2.HadoopJob getHadoopJob() { return com.google.cloud.dataproc.v1beta2.HadoopJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ - public Builder setHadoopJob(com.google.cloud.dataproc.v1beta2.HadoopJob value) { - if (hadoopJobBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - typeJob_ = value; - onChanged(); + /** + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setHadoopJob(com.google.cloud.dataproc.v1beta2.HadoopJob value) { + if (hadoopJobBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + typeJob_ = value; + onChanged(); } else { hadoopJobBuilder_.setMessage(value); } typeJobCase_ = 3; return this; } - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ + /** + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setHadoopJob( com.google.cloud.dataproc.v1beta2.HadoopJob.Builder builderForValue) { if (hadoopJobBuilder_ == null) { @@ -2678,7 +3036,17 @@ public Builder setHadoopJob( typeJobCase_ = 3; return this; } - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ + /** + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder mergeHadoopJob(com.google.cloud.dataproc.v1beta2.HadoopJob value) { if (hadoopJobBuilder_ == null) { if (typeJobCase_ == 3 @@ -2701,7 +3069,17 @@ public Builder mergeHadoopJob(com.google.cloud.dataproc.v1beta2.HadoopJob value) typeJobCase_ = 3; return this; } - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ + /** + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder clearHadoopJob() { if (hadoopJobBuilder_ == null) { if (typeJobCase_ == 3) { @@ -2718,11 +3096,31 @@ public Builder clearHadoopJob() { } return this; } - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ + /** + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.HadoopJob.Builder getHadoopJobBuilder() { return getHadoopJobFieldBuilder().getBuilder(); } - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ + /** + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.HadoopJobOrBuilder getHadoopJobOrBuilder() { if ((typeJobCase_ == 3) && (hadoopJobBuilder_ != null)) { return hadoopJobBuilder_.getMessageOrBuilder(); @@ -2733,7 +3131,17 @@ public com.google.cloud.dataproc.v1beta2.HadoopJobOrBuilder getHadoopJobOrBuilde return com.google.cloud.dataproc.v1beta2.HadoopJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ + /** + * + * + *
+     * Optional. Job is a Hadoop job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.HadoopJob, com.google.cloud.dataproc.v1beta2.HadoopJob.Builder, @@ -2765,7 +3173,15 @@ public com.google.cloud.dataproc.v1beta2.HadoopJobOrBuilder getHadoopJobOrBuilde com.google.cloud.dataproc.v1beta2.SparkJobOrBuilder> sparkJobBuilder_; /** - * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkJob field is set. */ @@ -2773,7 +3189,15 @@ public boolean hasSparkJob() { return typeJobCase_ == 4; } /** - * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkJob. */ @@ -2790,7 +3214,17 @@ public com.google.cloud.dataproc.v1beta2.SparkJob getSparkJob() { return com.google.cloud.dataproc.v1beta2.SparkJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setSparkJob(com.google.cloud.dataproc.v1beta2.SparkJob value) { if (sparkJobBuilder_ == null) { if (value == null) { @@ -2804,7 +3238,17 @@ public Builder setSparkJob(com.google.cloud.dataproc.v1beta2.SparkJob value) { typeJobCase_ = 4; return this; } - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setSparkJob(com.google.cloud.dataproc.v1beta2.SparkJob.Builder builderForValue) { if (sparkJobBuilder_ == null) { typeJob_ = builderForValue.build(); @@ -2815,7 +3259,17 @@ public Builder setSparkJob(com.google.cloud.dataproc.v1beta2.SparkJob.Builder bu typeJobCase_ = 4; return this; } - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder mergeSparkJob(com.google.cloud.dataproc.v1beta2.SparkJob value) { if (sparkJobBuilder_ == null) { if (typeJobCase_ == 4 @@ -2838,7 +3292,17 @@ public Builder mergeSparkJob(com.google.cloud.dataproc.v1beta2.SparkJob value) { typeJobCase_ = 4; return this; } - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder clearSparkJob() { if (sparkJobBuilder_ == null) { if (typeJobCase_ == 4) { @@ -2855,11 +3319,31 @@ public Builder clearSparkJob() { } return this; } - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkJob.Builder getSparkJobBuilder() { return getSparkJobFieldBuilder().getBuilder(); } - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkJobOrBuilder getSparkJobOrBuilder() { if ((typeJobCase_ == 4) && (sparkJobBuilder_ != null)) { return sparkJobBuilder_.getMessageOrBuilder(); @@ -2870,7 +3354,17 @@ public com.google.cloud.dataproc.v1beta2.SparkJobOrBuilder getSparkJobOrBuilder( return com.google.cloud.dataproc.v1beta2.SparkJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+     * Optional. Job is a Spark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.SparkJob, com.google.cloud.dataproc.v1beta2.SparkJob.Builder, @@ -2902,7 +3396,15 @@ public com.google.cloud.dataproc.v1beta2.SparkJobOrBuilder getSparkJobOrBuilder( com.google.cloud.dataproc.v1beta2.PySparkJobOrBuilder> pysparkJobBuilder_; /** - * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the pysparkJob field is set. */ @@ -2910,7 +3412,15 @@ public boolean hasPysparkJob() { return typeJobCase_ == 5; } /** - * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The pysparkJob. */ @@ -2927,7 +3437,17 @@ public com.google.cloud.dataproc.v1beta2.PySparkJob getPysparkJob() { return com.google.cloud.dataproc.v1beta2.PySparkJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setPysparkJob(com.google.cloud.dataproc.v1beta2.PySparkJob value) { if (pysparkJobBuilder_ == null) { if (value == null) { @@ -2941,7 +3461,17 @@ public Builder setPysparkJob(com.google.cloud.dataproc.v1beta2.PySparkJob value) typeJobCase_ = 5; return this; } - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setPysparkJob( com.google.cloud.dataproc.v1beta2.PySparkJob.Builder builderForValue) { if (pysparkJobBuilder_ == null) { @@ -2953,7 +3483,17 @@ public Builder setPysparkJob( typeJobCase_ = 5; return this; } - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder mergePysparkJob(com.google.cloud.dataproc.v1beta2.PySparkJob value) { if (pysparkJobBuilder_ == null) { if (typeJobCase_ == 5 @@ -2976,7 +3516,17 @@ public Builder mergePysparkJob(com.google.cloud.dataproc.v1beta2.PySparkJob valu typeJobCase_ = 5; return this; } - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder clearPysparkJob() { if (pysparkJobBuilder_ == null) { if (typeJobCase_ == 5) { @@ -2993,11 +3543,31 @@ public Builder clearPysparkJob() { } return this; } - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.PySparkJob.Builder getPysparkJobBuilder() { return getPysparkJobFieldBuilder().getBuilder(); } - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.PySparkJobOrBuilder getPysparkJobOrBuilder() { if ((typeJobCase_ == 5) && (pysparkJobBuilder_ != null)) { return pysparkJobBuilder_.getMessageOrBuilder(); @@ -3008,7 +3578,17 @@ public com.google.cloud.dataproc.v1beta2.PySparkJobOrBuilder getPysparkJobOrBuil return com.google.cloud.dataproc.v1beta2.PySparkJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+     * Optional. Job is a PySpark job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.PySparkJob, com.google.cloud.dataproc.v1beta2.PySparkJob.Builder, @@ -3040,7 +3620,15 @@ public com.google.cloud.dataproc.v1beta2.PySparkJobOrBuilder getPysparkJobOrBuil com.google.cloud.dataproc.v1beta2.HiveJobOrBuilder> hiveJobBuilder_; /** - * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the hiveJob field is set. */ @@ -3048,7 +3636,15 @@ public boolean hasHiveJob() { return typeJobCase_ == 6; } /** - * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The hiveJob. */ @@ -3065,7 +3661,17 @@ public com.google.cloud.dataproc.v1beta2.HiveJob getHiveJob() { return com.google.cloud.dataproc.v1beta2.HiveJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setHiveJob(com.google.cloud.dataproc.v1beta2.HiveJob value) { if (hiveJobBuilder_ == null) { if (value == null) { @@ -3079,7 +3685,17 @@ public Builder setHiveJob(com.google.cloud.dataproc.v1beta2.HiveJob value) { typeJobCase_ = 6; return this; } - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setHiveJob(com.google.cloud.dataproc.v1beta2.HiveJob.Builder builderForValue) { if (hiveJobBuilder_ == null) { typeJob_ = builderForValue.build(); @@ -3090,7 +3706,17 @@ public Builder setHiveJob(com.google.cloud.dataproc.v1beta2.HiveJob.Builder buil typeJobCase_ = 6; return this; } - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder mergeHiveJob(com.google.cloud.dataproc.v1beta2.HiveJob value) { if (hiveJobBuilder_ == null) { if (typeJobCase_ == 6 @@ -3113,7 +3739,17 @@ public Builder mergeHiveJob(com.google.cloud.dataproc.v1beta2.HiveJob value) { typeJobCase_ = 6; return this; } - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder clearHiveJob() { if (hiveJobBuilder_ == null) { if (typeJobCase_ == 6) { @@ -3130,11 +3766,31 @@ public Builder clearHiveJob() { } return this; } - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.HiveJob.Builder getHiveJobBuilder() { return getHiveJobFieldBuilder().getBuilder(); } - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.HiveJobOrBuilder getHiveJobOrBuilder() { if ((typeJobCase_ == 6) && (hiveJobBuilder_ != null)) { return hiveJobBuilder_.getMessageOrBuilder(); @@ -3145,7 +3801,17 @@ public com.google.cloud.dataproc.v1beta2.HiveJobOrBuilder getHiveJobOrBuilder() return com.google.cloud.dataproc.v1beta2.HiveJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+     * Optional. Job is a Hive job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.HiveJob, com.google.cloud.dataproc.v1beta2.HiveJob.Builder, @@ -3177,7 +3843,15 @@ public com.google.cloud.dataproc.v1beta2.HiveJobOrBuilder getHiveJobOrBuilder() com.google.cloud.dataproc.v1beta2.PigJobOrBuilder> pigJobBuilder_; /** - * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the pigJob field is set. */ @@ -3185,7 +3859,15 @@ public boolean hasPigJob() { return typeJobCase_ == 7; } /** - * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The pigJob. */ @@ -3202,7 +3884,17 @@ public com.google.cloud.dataproc.v1beta2.PigJob getPigJob() { return com.google.cloud.dataproc.v1beta2.PigJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setPigJob(com.google.cloud.dataproc.v1beta2.PigJob value) { if (pigJobBuilder_ == null) { if (value == null) { @@ -3216,7 +3908,17 @@ public Builder setPigJob(com.google.cloud.dataproc.v1beta2.PigJob value) { typeJobCase_ = 7; return this; } - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setPigJob(com.google.cloud.dataproc.v1beta2.PigJob.Builder builderForValue) { if (pigJobBuilder_ == null) { typeJob_ = builderForValue.build(); @@ -3227,7 +3929,17 @@ public Builder setPigJob(com.google.cloud.dataproc.v1beta2.PigJob.Builder builde typeJobCase_ = 7; return this; } - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder mergePigJob(com.google.cloud.dataproc.v1beta2.PigJob value) { if (pigJobBuilder_ == null) { if (typeJobCase_ == 7 @@ -3250,7 +3962,17 @@ public Builder mergePigJob(com.google.cloud.dataproc.v1beta2.PigJob value) { typeJobCase_ = 7; return this; } - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder clearPigJob() { if (pigJobBuilder_ == null) { if (typeJobCase_ == 7) { @@ -3267,11 +3989,31 @@ public Builder clearPigJob() { } return this; } - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.PigJob.Builder getPigJobBuilder() { return getPigJobFieldBuilder().getBuilder(); } - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder() { if ((typeJobCase_ == 7) && (pigJobBuilder_ != null)) { return pigJobBuilder_.getMessageOrBuilder(); @@ -3282,7 +4024,17 @@ public com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder() { return com.google.cloud.dataproc.v1beta2.PigJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+     * Optional. Job is a Pig job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.PigJob, com.google.cloud.dataproc.v1beta2.PigJob.Builder, @@ -3314,7 +4066,15 @@ public com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder() { com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder> sparkRJobBuilder_; /** - * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkRJob field is set. */ @@ -3322,7 +4082,15 @@ public boolean hasSparkRJob() { return typeJobCase_ == 21; } /** - * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkRJob. */ @@ -3339,7 +4107,17 @@ public com.google.cloud.dataproc.v1beta2.SparkRJob getSparkRJob() { return com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setSparkRJob(com.google.cloud.dataproc.v1beta2.SparkRJob value) { if (sparkRJobBuilder_ == null) { if (value == null) { @@ -3353,7 +4131,17 @@ public Builder setSparkRJob(com.google.cloud.dataproc.v1beta2.SparkRJob value) { typeJobCase_ = 21; return this; } - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setSparkRJob( com.google.cloud.dataproc.v1beta2.SparkRJob.Builder builderForValue) { if (sparkRJobBuilder_ == null) { @@ -3365,7 +4153,17 @@ public Builder setSparkRJob( typeJobCase_ = 21; return this; } - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder mergeSparkRJob(com.google.cloud.dataproc.v1beta2.SparkRJob value) { if (sparkRJobBuilder_ == null) { if (typeJobCase_ == 21 @@ -3388,7 +4186,17 @@ public Builder mergeSparkRJob(com.google.cloud.dataproc.v1beta2.SparkRJob value) typeJobCase_ = 21; return this; } - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder clearSparkRJob() { if (sparkRJobBuilder_ == null) { if (typeJobCase_ == 21) { @@ -3405,11 +4213,31 @@ public Builder clearSparkRJob() { } return this; } - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkRJob.Builder getSparkRJobBuilder() { return getSparkRJobFieldBuilder().getBuilder(); } - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilder() { if ((typeJobCase_ == 21) && (sparkRJobBuilder_ != null)) { return sparkRJobBuilder_.getMessageOrBuilder(); @@ -3420,7 +4248,17 @@ public com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilde return com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+     * Optional. Job is a SparkR job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.SparkRJob, com.google.cloud.dataproc.v1beta2.SparkRJob.Builder, @@ -3452,7 +4290,15 @@ public com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilde com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder> sparkSqlJobBuilder_; /** - * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkSqlJob field is set. */ @@ -3460,7 +4306,15 @@ public boolean hasSparkSqlJob() { return typeJobCase_ == 12; } /** - * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkSqlJob. */ @@ -3477,7 +4331,17 @@ public com.google.cloud.dataproc.v1beta2.SparkSqlJob getSparkSqlJob() { return com.google.cloud.dataproc.v1beta2.SparkSqlJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setSparkSqlJob(com.google.cloud.dataproc.v1beta2.SparkSqlJob value) { if (sparkSqlJobBuilder_ == null) { if (value == null) { @@ -3491,7 +4355,17 @@ public Builder setSparkSqlJob(com.google.cloud.dataproc.v1beta2.SparkSqlJob valu typeJobCase_ = 12; return this; } - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder setSparkSqlJob( com.google.cloud.dataproc.v1beta2.SparkSqlJob.Builder builderForValue) { if (sparkSqlJobBuilder_ == null) { @@ -3503,7 +4377,17 @@ public Builder setSparkSqlJob( typeJobCase_ = 12; return this; } - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder mergeSparkSqlJob(com.google.cloud.dataproc.v1beta2.SparkSqlJob value) { if (sparkSqlJobBuilder_ == null) { if (typeJobCase_ == 12 @@ -3526,7 +4410,17 @@ public Builder mergeSparkSqlJob(com.google.cloud.dataproc.v1beta2.SparkSqlJob va typeJobCase_ = 12; return this; } - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public Builder clearSparkSqlJob() { if (sparkSqlJobBuilder_ == null) { if (typeJobCase_ == 12) { @@ -3543,11 +4437,31 @@ public Builder clearSparkSqlJob() { } return this; } - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkSqlJob.Builder getSparkSqlJobBuilder() { return getSparkSqlJobFieldBuilder().getBuilder(); } - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ public com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBuilder() { if ((typeJobCase_ == 12) && (sparkSqlJobBuilder_ != null)) { return sparkSqlJobBuilder_.getMessageOrBuilder(); @@ -3558,7 +4472,17 @@ public com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBu return com.google.cloud.dataproc.v1beta2.SparkSqlJob.getDefaultInstance(); } } - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+     * Optional. Job is a SparkSql job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.SparkSqlJob, com.google.cloud.dataproc.v1beta2.SparkSqlJob.Builder, @@ -3584,6 +4508,230 @@ public com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBu return sparkSqlJobBuilder_; } + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.PrestoJob, + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder, + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder> + prestoJobBuilder_; + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the prestoJob field is set. + */ + public boolean hasPrestoJob() { + return typeJobCase_ == 23; + } + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The prestoJob. + */ + public com.google.cloud.dataproc.v1beta2.PrestoJob getPrestoJob() { + if (prestoJobBuilder_ == null) { + if (typeJobCase_ == 23) { + return (com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_; + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } else { + if (typeJobCase_ == 23) { + return prestoJobBuilder_.getMessage(); + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setPrestoJob(com.google.cloud.dataproc.v1beta2.PrestoJob value) { + if (prestoJobBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + typeJob_ = value; + onChanged(); + } else { + prestoJobBuilder_.setMessage(value); + } + typeJobCase_ = 23; + return this; + } + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setPrestoJob( + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder builderForValue) { + if (prestoJobBuilder_ == null) { + typeJob_ = builderForValue.build(); + onChanged(); + } else { + prestoJobBuilder_.setMessage(builderForValue.build()); + } + typeJobCase_ = 23; + return this; + } + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder mergePrestoJob(com.google.cloud.dataproc.v1beta2.PrestoJob value) { + if (prestoJobBuilder_ == null) { + if (typeJobCase_ == 23 + && typeJob_ != com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance()) { + typeJob_ = + com.google.cloud.dataproc.v1beta2.PrestoJob.newBuilder( + (com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_) + .mergeFrom(value) + .buildPartial(); + } else { + typeJob_ = value; + } + onChanged(); + } else { + if (typeJobCase_ == 23) { + prestoJobBuilder_.mergeFrom(value); + } + prestoJobBuilder_.setMessage(value); + } + typeJobCase_ = 23; + return this; + } + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder clearPrestoJob() { + if (prestoJobBuilder_ == null) { + if (typeJobCase_ == 23) { + typeJobCase_ = 0; + typeJob_ = null; + onChanged(); + } + } else { + if (typeJobCase_ == 23) { + typeJobCase_ = 0; + typeJob_ = null; + } + prestoJobBuilder_.clear(); + } + return this; + } + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.PrestoJob.Builder getPrestoJobBuilder() { + return getPrestoJobFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder getPrestoJobOrBuilder() { + if ((typeJobCase_ == 23) && (prestoJobBuilder_ != null)) { + return prestoJobBuilder_.getMessageOrBuilder(); + } else { + if (typeJobCase_ == 23) { + return (com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_; + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Optional. Job is a Presto job.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.PrestoJob, + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder, + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder> + getPrestoJobFieldBuilder() { + if (prestoJobBuilder_ == null) { + if (!(typeJobCase_ == 23)) { + typeJob_ = com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + prestoJobBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.PrestoJob, + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder, + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder>( + (com.google.cloud.dataproc.v1beta2.PrestoJob) typeJob_, + getParentForChildren(), + isClean()); + typeJob_ = null; + } + typeJobCase_ = 23; + onChanged(); + ; + return prestoJobBuilder_; + } + private com.google.cloud.dataproc.v1beta2.JobStatus status_; private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.JobStatus, @@ -5492,6 +6640,66 @@ public Builder setJobUuidBytes(com.google.protobuf.ByteString value) { return this; } + private boolean done_; + /** + * + * + *
+     * Output only. Indicates whether the job is completed. If the value is `false`,
+     * the job is still in progress. If `true`, the job is completed, and
+     * `status.state` field will indicate if it was successful, failed,
+     * or cancelled.
+     * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The done. + */ + public boolean getDone() { + return done_; + } + /** + * + * + *
+     * Output only. Indicates whether the job is completed. If the value is `false`,
+     * the job is still in progress. If `true`, the job is completed, and
+     * `status.state` field will indicate if it was successful, failed,
+     * or cancelled.
+     * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The done to set. + * @return This builder for chaining. + */ + public Builder setDone(boolean value) { + + done_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. Indicates whether the job is completed. If the value is `false`,
+     * the job is still in progress. If `true`, the job is completed, and
+     * `status.state` field will indicate if it was successful, failed,
+     * or cancelled.
+     * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return This builder for chaining. + */ + public Builder clearDone() { + + done_ = false; + onChanged(); + return this; + } + @java.lang.Override public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobMetadata.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobMetadata.java new file mode 100644 index 00000000..659d6a6b --- /dev/null +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobMetadata.java @@ -0,0 +1,1384 @@ +/* + * Copyright 2020 Google LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dataproc/v1beta2/jobs.proto + +package com.google.cloud.dataproc.v1beta2; + +/** + * + * + *
+ * Job Operation metadata.
+ * 
+ * + * Protobuf type {@code google.cloud.dataproc.v1beta2.JobMetadata} + */ +public final class JobMetadata extends com.google.protobuf.GeneratedMessageV3 + implements + // @@protoc_insertion_point(message_implements:google.cloud.dataproc.v1beta2.JobMetadata) + JobMetadataOrBuilder { + private static final long serialVersionUID = 0L; + // Use JobMetadata.newBuilder() to construct. + private JobMetadata(com.google.protobuf.GeneratedMessageV3.Builder builder) { + super(builder); + } + + private JobMetadata() { + jobId_ = ""; + operationType_ = ""; + } + + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance(UnusedPrivateParameter unused) { + return new JobMetadata(); + } + + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet getUnknownFields() { + return this.unknownFields; + } + + private JobMetadata( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + case 10: + { + java.lang.String s = input.readStringRequireUtf8(); + + jobId_ = s; + break; + } + case 18: + { + com.google.cloud.dataproc.v1beta2.JobStatus.Builder subBuilder = null; + if (status_ != null) { + subBuilder = status_.toBuilder(); + } + status_ = + input.readMessage( + com.google.cloud.dataproc.v1beta2.JobStatus.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom(status_); + status_ = subBuilder.buildPartial(); + } + + break; + } + case 26: + { + java.lang.String s = input.readStringRequireUtf8(); + + operationType_ = s; + break; + } + case 34: + { + com.google.protobuf.Timestamp.Builder subBuilder = null; + if (startTime_ != null) { + subBuilder = startTime_.toBuilder(); + } + startTime_ = + input.readMessage(com.google.protobuf.Timestamp.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom(startTime_); + startTime_ = subBuilder.buildPartial(); + } + + break; + } + default: + { + if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException(e).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_JobMetadata_descriptor; + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_JobMetadata_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1beta2.JobMetadata.class, + com.google.cloud.dataproc.v1beta2.JobMetadata.Builder.class); + } + + public static final int JOB_ID_FIELD_NUMBER = 1; + private volatile java.lang.Object jobId_; + /** + * + * + *
+   * Output only. The job id.
+   * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The jobId. + */ + public java.lang.String getJobId() { + java.lang.Object ref = jobId_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + jobId_ = s; + return s; + } + } + /** + * + * + *
+   * Output only. The job id.
+   * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for jobId. + */ + public com.google.protobuf.ByteString getJobIdBytes() { + java.lang.Object ref = jobId_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + jobId_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + public static final int STATUS_FIELD_NUMBER = 2; + private com.google.cloud.dataproc.v1beta2.JobStatus status_; + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the status field is set. + */ + public boolean hasStatus() { + return status_ != null; + } + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The status. + */ + public com.google.cloud.dataproc.v1beta2.JobStatus getStatus() { + return status_ == null + ? com.google.cloud.dataproc.v1beta2.JobStatus.getDefaultInstance() + : status_; + } + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.cloud.dataproc.v1beta2.JobStatusOrBuilder getStatusOrBuilder() { + return getStatus(); + } + + public static final int OPERATION_TYPE_FIELD_NUMBER = 3; + private volatile java.lang.Object operationType_; + /** + * + * + *
+   * Output only. Operation type.
+   * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The operationType. + */ + public java.lang.String getOperationType() { + java.lang.Object ref = operationType_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + operationType_ = s; + return s; + } + } + /** + * + * + *
+   * Output only. Operation type.
+   * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for operationType. + */ + public com.google.protobuf.ByteString getOperationTypeBytes() { + java.lang.Object ref = operationType_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + operationType_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + public static final int START_TIME_FIELD_NUMBER = 4; + private com.google.protobuf.Timestamp startTime_; + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the startTime field is set. + */ + public boolean hasStartTime() { + return startTime_ != null; + } + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The startTime. + */ + public com.google.protobuf.Timestamp getStartTime() { + return startTime_ == null ? com.google.protobuf.Timestamp.getDefaultInstance() : startTime_; + } + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.protobuf.TimestampOrBuilder getStartTimeOrBuilder() { + return getStartTime(); + } + + private byte memoizedIsInitialized = -1; + + @java.lang.Override + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; + + memoizedIsInitialized = 1; + return true; + } + + @java.lang.Override + public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { + if (!getJobIdBytes().isEmpty()) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, jobId_); + } + if (status_ != null) { + output.writeMessage(2, getStatus()); + } + if (!getOperationTypeBytes().isEmpty()) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 3, operationType_); + } + if (startTime_ != null) { + output.writeMessage(4, getStartTime()); + } + unknownFields.writeTo(output); + } + + @java.lang.Override + public int getSerializedSize() { + int size = memoizedSize; + if (size != -1) return size; + + size = 0; + if (!getJobIdBytes().isEmpty()) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, jobId_); + } + if (status_ != null) { + size += com.google.protobuf.CodedOutputStream.computeMessageSize(2, getStatus()); + } + if (!getOperationTypeBytes().isEmpty()) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, operationType_); + } + if (startTime_ != null) { + size += com.google.protobuf.CodedOutputStream.computeMessageSize(4, getStartTime()); + } + size += unknownFields.getSerializedSize(); + memoizedSize = size; + return size; + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof com.google.cloud.dataproc.v1beta2.JobMetadata)) { + return super.equals(obj); + } + com.google.cloud.dataproc.v1beta2.JobMetadata other = + (com.google.cloud.dataproc.v1beta2.JobMetadata) obj; + + if (!getJobId().equals(other.getJobId())) return false; + if (hasStatus() != other.hasStatus()) return false; + if (hasStatus()) { + if (!getStatus().equals(other.getStatus())) return false; + } + if (!getOperationType().equals(other.getOperationType())) return false; + if (hasStartTime() != other.hasStartTime()) return false; + if (hasStartTime()) { + if (!getStartTime().equals(other.getStartTime())) return false; + } + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (37 * hash) + JOB_ID_FIELD_NUMBER; + hash = (53 * hash) + getJobId().hashCode(); + if (hasStatus()) { + hash = (37 * hash) + STATUS_FIELD_NUMBER; + hash = (53 * hash) + getStatus().hashCode(); + } + hash = (37 * hash) + OPERATION_TYPE_FIELD_NUMBER; + hash = (53 * hash) + getOperationType().hashCode(); + if (hasStartTime()) { + hash = (37 * hash) + START_TIME_FIELD_NUMBER; + hash = (53 * hash) + getStartTime().hashCode(); + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom(java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom( + java.nio.ByteBuffer data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom( + byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom(java.io.InputStream input) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseDelimitedFrom( + java.io.InputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseDelimitedFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom( + com.google.protobuf.CodedInputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + @java.lang.Override + public Builder newBuilderForType() { + return newBuilder(); + } + + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } + + public static Builder newBuilder(com.google.cloud.dataproc.v1beta2.JobMetadata prototype) { + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); + } + + @java.lang.Override + protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * + * + *
+   * Job Operation metadata.
+   * 
+ * + * Protobuf type {@code google.cloud.dataproc.v1beta2.JobMetadata} + */ + public static final class Builder extends com.google.protobuf.GeneratedMessageV3.Builder + implements + // @@protoc_insertion_point(builder_implements:google.cloud.dataproc.v1beta2.JobMetadata) + com.google.cloud.dataproc.v1beta2.JobMetadataOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_JobMetadata_descriptor; + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_JobMetadata_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1beta2.JobMetadata.class, + com.google.cloud.dataproc.v1beta2.JobMetadata.Builder.class); + } + + // Construct using com.google.cloud.dataproc.v1beta2.JobMetadata.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders) {} + } + + @java.lang.Override + public Builder clear() { + super.clear(); + jobId_ = ""; + + if (statusBuilder_ == null) { + status_ = null; + } else { + status_ = null; + statusBuilder_ = null; + } + operationType_ = ""; + + if (startTimeBuilder_ == null) { + startTime_ = null; + } else { + startTime_ = null; + startTimeBuilder_ = null; + } + return this; + } + + @java.lang.Override + public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_JobMetadata_descriptor; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.JobMetadata getDefaultInstanceForType() { + return com.google.cloud.dataproc.v1beta2.JobMetadata.getDefaultInstance(); + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.JobMetadata build() { + com.google.cloud.dataproc.v1beta2.JobMetadata result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.JobMetadata buildPartial() { + com.google.cloud.dataproc.v1beta2.JobMetadata result = + new com.google.cloud.dataproc.v1beta2.JobMetadata(this); + result.jobId_ = jobId_; + if (statusBuilder_ == null) { + result.status_ = status_; + } else { + result.status_ = statusBuilder_.build(); + } + result.operationType_ = operationType_; + if (startTimeBuilder_ == null) { + result.startTime_ = startTime_; + } else { + result.startTime_ = startTimeBuilder_.build(); + } + onBuilt(); + return result; + } + + @java.lang.Override + public Builder clone() { + return super.clone(); + } + + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.setField(field, value); + } + + @java.lang.Override + public Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + + @java.lang.Override + public Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.addRepeatedField(field, value); + } + + @java.lang.Override + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof com.google.cloud.dataproc.v1beta2.JobMetadata) { + return mergeFrom((com.google.cloud.dataproc.v1beta2.JobMetadata) other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(com.google.cloud.dataproc.v1beta2.JobMetadata other) { + if (other == com.google.cloud.dataproc.v1beta2.JobMetadata.getDefaultInstance()) return this; + if (!other.getJobId().isEmpty()) { + jobId_ = other.jobId_; + onChanged(); + } + if (other.hasStatus()) { + mergeStatus(other.getStatus()); + } + if (!other.getOperationType().isEmpty()) { + operationType_ = other.operationType_; + onChanged(); + } + if (other.hasStartTime()) { + mergeStartTime(other.getStartTime()); + } + this.mergeUnknownFields(other.unknownFields); + onChanged(); + return this; + } + + @java.lang.Override + public final boolean isInitialized() { + return true; + } + + @java.lang.Override + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.cloud.dataproc.v1beta2.JobMetadata parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (com.google.cloud.dataproc.v1beta2.JobMetadata) e.getUnfinishedMessage(); + throw e.unwrapIOException(); + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + + private java.lang.Object jobId_ = ""; + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The jobId. + */ + public java.lang.String getJobId() { + java.lang.Object ref = jobId_; + if (!(ref instanceof java.lang.String)) { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + jobId_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for jobId. + */ + public com.google.protobuf.ByteString getJobIdBytes() { + java.lang.Object ref = jobId_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + jobId_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The jobId to set. + * @return This builder for chaining. + */ + public Builder setJobId(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + + jobId_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return This builder for chaining. + */ + public Builder clearJobId() { + + jobId_ = getDefaultInstance().getJobId(); + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. The job id.
+     * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The bytes for jobId to set. + * @return This builder for chaining. + */ + public Builder setJobIdBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + + jobId_ = value; + onChanged(); + return this; + } + + private com.google.cloud.dataproc.v1beta2.JobStatus status_; + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.JobStatus, + com.google.cloud.dataproc.v1beta2.JobStatus.Builder, + com.google.cloud.dataproc.v1beta2.JobStatusOrBuilder> + statusBuilder_; + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the status field is set. + */ + public boolean hasStatus() { + return statusBuilder_ != null || status_ != null; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The status. + */ + public com.google.cloud.dataproc.v1beta2.JobStatus getStatus() { + if (statusBuilder_ == null) { + return status_ == null + ? com.google.cloud.dataproc.v1beta2.JobStatus.getDefaultInstance() + : status_; + } else { + return statusBuilder_.getMessage(); + } + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder setStatus(com.google.cloud.dataproc.v1beta2.JobStatus value) { + if (statusBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + status_ = value; + onChanged(); + } else { + statusBuilder_.setMessage(value); + } + + return this; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder setStatus(com.google.cloud.dataproc.v1beta2.JobStatus.Builder builderForValue) { + if (statusBuilder_ == null) { + status_ = builderForValue.build(); + onChanged(); + } else { + statusBuilder_.setMessage(builderForValue.build()); + } + + return this; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder mergeStatus(com.google.cloud.dataproc.v1beta2.JobStatus value) { + if (statusBuilder_ == null) { + if (status_ != null) { + status_ = + com.google.cloud.dataproc.v1beta2.JobStatus.newBuilder(status_) + .mergeFrom(value) + .buildPartial(); + } else { + status_ = value; + } + onChanged(); + } else { + statusBuilder_.mergeFrom(value); + } + + return this; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder clearStatus() { + if (statusBuilder_ == null) { + status_ = null; + onChanged(); + } else { + status_ = null; + statusBuilder_ = null; + } + + return this; + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.cloud.dataproc.v1beta2.JobStatus.Builder getStatusBuilder() { + + onChanged(); + return getStatusFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.cloud.dataproc.v1beta2.JobStatusOrBuilder getStatusOrBuilder() { + if (statusBuilder_ != null) { + return statusBuilder_.getMessageOrBuilder(); + } else { + return status_ == null + ? com.google.cloud.dataproc.v1beta2.JobStatus.getDefaultInstance() + : status_; + } + } + /** + * + * + *
+     * Output only. Most recent job status.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.JobStatus, + com.google.cloud.dataproc.v1beta2.JobStatus.Builder, + com.google.cloud.dataproc.v1beta2.JobStatusOrBuilder> + getStatusFieldBuilder() { + if (statusBuilder_ == null) { + statusBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.JobStatus, + com.google.cloud.dataproc.v1beta2.JobStatus.Builder, + com.google.cloud.dataproc.v1beta2.JobStatusOrBuilder>( + getStatus(), getParentForChildren(), isClean()); + status_ = null; + } + return statusBuilder_; + } + + private java.lang.Object operationType_ = ""; + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The operationType. + */ + public java.lang.String getOperationType() { + java.lang.Object ref = operationType_; + if (!(ref instanceof java.lang.String)) { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + operationType_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for operationType. + */ + public com.google.protobuf.ByteString getOperationTypeBytes() { + java.lang.Object ref = operationType_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + operationType_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The operationType to set. + * @return This builder for chaining. + */ + public Builder setOperationType(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + + operationType_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return This builder for chaining. + */ + public Builder clearOperationType() { + + operationType_ = getDefaultInstance().getOperationType(); + onChanged(); + return this; + } + /** + * + * + *
+     * Output only. Operation type.
+     * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @param value The bytes for operationType to set. + * @return This builder for chaining. + */ + public Builder setOperationTypeBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + + operationType_ = value; + onChanged(); + return this; + } + + private com.google.protobuf.Timestamp startTime_; + private com.google.protobuf.SingleFieldBuilderV3< + com.google.protobuf.Timestamp, + com.google.protobuf.Timestamp.Builder, + com.google.protobuf.TimestampOrBuilder> + startTimeBuilder_; + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the startTime field is set. + */ + public boolean hasStartTime() { + return startTimeBuilder_ != null || startTime_ != null; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The startTime. + */ + public com.google.protobuf.Timestamp getStartTime() { + if (startTimeBuilder_ == null) { + return startTime_ == null ? com.google.protobuf.Timestamp.getDefaultInstance() : startTime_; + } else { + return startTimeBuilder_.getMessage(); + } + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder setStartTime(com.google.protobuf.Timestamp value) { + if (startTimeBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + startTime_ = value; + onChanged(); + } else { + startTimeBuilder_.setMessage(value); + } + + return this; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder setStartTime(com.google.protobuf.Timestamp.Builder builderForValue) { + if (startTimeBuilder_ == null) { + startTime_ = builderForValue.build(); + onChanged(); + } else { + startTimeBuilder_.setMessage(builderForValue.build()); + } + + return this; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder mergeStartTime(com.google.protobuf.Timestamp value) { + if (startTimeBuilder_ == null) { + if (startTime_ != null) { + startTime_ = + com.google.protobuf.Timestamp.newBuilder(startTime_).mergeFrom(value).buildPartial(); + } else { + startTime_ = value; + } + onChanged(); + } else { + startTimeBuilder_.mergeFrom(value); + } + + return this; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public Builder clearStartTime() { + if (startTimeBuilder_ == null) { + startTime_ = null; + onChanged(); + } else { + startTime_ = null; + startTimeBuilder_ = null; + } + + return this; + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.protobuf.Timestamp.Builder getStartTimeBuilder() { + + onChanged(); + return getStartTimeFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + public com.google.protobuf.TimestampOrBuilder getStartTimeOrBuilder() { + if (startTimeBuilder_ != null) { + return startTimeBuilder_.getMessageOrBuilder(); + } else { + return startTime_ == null ? com.google.protobuf.Timestamp.getDefaultInstance() : startTime_; + } + } + /** + * + * + *
+     * Output only. Job submission time.
+     * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.protobuf.Timestamp, + com.google.protobuf.Timestamp.Builder, + com.google.protobuf.TimestampOrBuilder> + getStartTimeFieldBuilder() { + if (startTimeBuilder_ == null) { + startTimeBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.protobuf.Timestamp, + com.google.protobuf.Timestamp.Builder, + com.google.protobuf.TimestampOrBuilder>( + getStartTime(), getParentForChildren(), isClean()); + startTime_ = null; + } + return startTimeBuilder_; + } + + @java.lang.Override + public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + + // @@protoc_insertion_point(builder_scope:google.cloud.dataproc.v1beta2.JobMetadata) + } + + // @@protoc_insertion_point(class_scope:google.cloud.dataproc.v1beta2.JobMetadata) + private static final com.google.cloud.dataproc.v1beta2.JobMetadata DEFAULT_INSTANCE; + + static { + DEFAULT_INSTANCE = new com.google.cloud.dataproc.v1beta2.JobMetadata(); + } + + public static com.google.cloud.dataproc.v1beta2.JobMetadata getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + private static final com.google.protobuf.Parser PARSER = + new com.google.protobuf.AbstractParser() { + @java.lang.Override + public JobMetadata parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new JobMetadata(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.JobMetadata getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } +} diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobMetadataOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobMetadataOrBuilder.java new file mode 100644 index 00000000..f5183d06 --- /dev/null +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobMetadataOrBuilder.java @@ -0,0 +1,154 @@ +/* + * Copyright 2020 Google LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dataproc/v1beta2/jobs.proto + +package com.google.cloud.dataproc.v1beta2; + +public interface JobMetadataOrBuilder + extends + // @@protoc_insertion_point(interface_extends:google.cloud.dataproc.v1beta2.JobMetadata) + com.google.protobuf.MessageOrBuilder { + + /** + * + * + *
+   * Output only. The job id.
+   * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The jobId. + */ + java.lang.String getJobId(); + /** + * + * + *
+   * Output only. The job id.
+   * 
+ * + * string job_id = 1 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for jobId. + */ + com.google.protobuf.ByteString getJobIdBytes(); + + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the status field is set. + */ + boolean hasStatus(); + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The status. + */ + com.google.cloud.dataproc.v1beta2.JobStatus getStatus(); + /** + * + * + *
+   * Output only. Most recent job status.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.JobStatus status = 2 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + com.google.cloud.dataproc.v1beta2.JobStatusOrBuilder getStatusOrBuilder(); + + /** + * + * + *
+   * Output only. Operation type.
+   * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The operationType. + */ + java.lang.String getOperationType(); + /** + * + * + *
+   * Output only. Operation type.
+   * 
+ * + * string operation_type = 3 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The bytes for operationType. + */ + com.google.protobuf.ByteString getOperationTypeBytes(); + + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return Whether the startTime field is set. + */ + boolean hasStartTime(); + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * + * @return The startTime. + */ + com.google.protobuf.Timestamp getStartTime(); + /** + * + * + *
+   * Output only. Job submission time.
+   * 
+ * + * .google.protobuf.Timestamp start_time = 4 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + */ + com.google.protobuf.TimestampOrBuilder getStartTimeOrBuilder(); +} diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobOrBuilder.java index d07f69b6..302177cc 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobOrBuilder.java @@ -118,110 +118,333 @@ public interface JobOrBuilder com.google.cloud.dataproc.v1beta2.JobPlacementOrBuilder getPlacementOrBuilder(); /** - * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; + * + * + *
+   * Optional. Job is a Hadoop job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the hadoopJob field is set. */ boolean hasHadoopJob(); /** - * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; + * + * + *
+   * Optional. Job is a Hadoop job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The hadoopJob. */ com.google.cloud.dataproc.v1beta2.HadoopJob getHadoopJob(); - /** .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3; */ + /** + * + * + *
+   * Optional. Job is a Hadoop job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HadoopJob hadoop_job = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + */ com.google.cloud.dataproc.v1beta2.HadoopJobOrBuilder getHadoopJobOrBuilder(); /** - * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; + * + * + *
+   * Optional. Job is a Spark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkJob field is set. */ boolean hasSparkJob(); /** - * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; + * + * + *
+   * Optional. Job is a Spark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkJob. */ com.google.cloud.dataproc.v1beta2.SparkJob getSparkJob(); - /** .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4; */ + /** + * + * + *
+   * Optional. Job is a Spark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkJob spark_job = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + */ com.google.cloud.dataproc.v1beta2.SparkJobOrBuilder getSparkJobOrBuilder(); /** - * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; + * + * + *
+   * Optional. Job is a PySpark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the pysparkJob field is set. */ boolean hasPysparkJob(); /** - * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; + * + * + *
+   * Optional. Job is a PySpark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The pysparkJob. */ com.google.cloud.dataproc.v1beta2.PySparkJob getPysparkJob(); - /** .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5; */ + /** + * + * + *
+   * Optional. Job is a PySpark job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PySparkJob pyspark_job = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + */ com.google.cloud.dataproc.v1beta2.PySparkJobOrBuilder getPysparkJobOrBuilder(); /** - * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; + * + * + *
+   * Optional. Job is a Hive job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the hiveJob field is set. */ boolean hasHiveJob(); /** - * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; + * + * + *
+   * Optional. Job is a Hive job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The hiveJob. */ com.google.cloud.dataproc.v1beta2.HiveJob getHiveJob(); - /** .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6; */ + /** + * + * + *
+   * Optional. Job is a Hive job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.HiveJob hive_job = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ com.google.cloud.dataproc.v1beta2.HiveJobOrBuilder getHiveJobOrBuilder(); /** - * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; + * + * + *
+   * Optional. Job is a Pig job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the pigJob field is set. */ boolean hasPigJob(); /** - * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; + * + * + *
+   * Optional. Job is a Pig job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The pigJob. */ com.google.cloud.dataproc.v1beta2.PigJob getPigJob(); - /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 7; */ + /** + * + * + *
+   * Optional. Job is a Pig job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PigJob pig_job = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder(); /** - * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; + * + * + *
+   * Optional. Job is a SparkR job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkRJob field is set. */ boolean hasSparkRJob(); /** - * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; + * + * + *
+   * Optional. Job is a SparkR job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkRJob. */ com.google.cloud.dataproc.v1beta2.SparkRJob getSparkRJob(); - /** .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21; */ + /** + * + * + *
+   * Optional. Job is a SparkR job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 21 [(.google.api.field_behavior) = OPTIONAL]; + * + */ com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilder(); /** - * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; + * + * + *
+   * Optional. Job is a SparkSql job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the sparkSqlJob field is set. */ boolean hasSparkSqlJob(); /** - * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; + * + * + *
+   * Optional. Job is a SparkSql job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The sparkSqlJob. */ com.google.cloud.dataproc.v1beta2.SparkSqlJob getSparkSqlJob(); - /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12; */ + /** + * + * + *
+   * Optional. Job is a SparkSql job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 12 [(.google.api.field_behavior) = OPTIONAL]; + * + */ com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBuilder(); + /** + * + * + *
+   * Optional. Job is a Presto job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the prestoJob field is set. + */ + boolean hasPrestoJob(); + /** + * + * + *
+   * Optional. Job is a Presto job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The prestoJob. + */ + com.google.cloud.dataproc.v1beta2.PrestoJob getPrestoJob(); + /** + * + * + *
+   * Optional. Job is a Presto job.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 23 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder getPrestoJobOrBuilder(); + /** * * @@ -643,5 +866,21 @@ com.google.cloud.dataproc.v1beta2.YarnApplicationOrBuilder getYarnApplicationsOr */ com.google.protobuf.ByteString getJobUuidBytes(); + /** + * + * + *
+   * Output only. Indicates whether the job is completed. If the value is `false`,
+   * the job is still in progress. If `true`, the job is completed, and
+   * `status.state` field will indicate if it was successful, failed,
+   * or cancelled.
+   * 
+ * + * bool done = 24 [(.google.api.field_behavior) = OUTPUT_ONLY]; + * + * @return The done. + */ + boolean getDone(); + public com.google.cloud.dataproc.v1beta2.Job.TypeJobCase getTypeJobCase(); } diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobsProto.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobsProto.java index 808fd83f..5060ff89 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobsProto.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/JobsProto.java @@ -107,6 +107,14 @@ public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry r internal_static_google_cloud_dataproc_v1beta2_SparkRJob_PropertiesEntry_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_google_cloud_dataproc_v1beta2_SparkRJob_PropertiesEntry_fieldAccessorTable; + static final com.google.protobuf.Descriptors.Descriptor + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_descriptor; + static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_fieldAccessorTable; + static final com.google.protobuf.Descriptors.Descriptor + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_PropertiesEntry_descriptor; + static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_PropertiesEntry_fieldAccessorTable; static final com.google.protobuf.Descriptors.Descriptor internal_static_google_cloud_dataproc_v1beta2_JobPlacement_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable @@ -135,6 +143,10 @@ public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry r internal_static_google_cloud_dataproc_v1beta2_JobScheduling_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_google_cloud_dataproc_v1beta2_JobScheduling_fieldAccessorTable; + static final com.google.protobuf.Descriptors.Descriptor + internal_static_google_cloud_dataproc_v1beta2_JobMetadata_descriptor; + static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_google_cloud_dataproc_v1beta2_JobMetadata_fieldAccessorTable; static final com.google.protobuf.Descriptors.Descriptor internal_static_google_cloud_dataproc_v1beta2_SubmitJobRequest_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable @@ -262,120 +274,144 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + ".PropertiesEntryB\003\340A\001\022I\n\016logging_config\030" + "\006 \001(\0132,.google.cloud.dataproc.v1beta2.Lo" + "ggingConfigB\003\340A\001\0321\n\017PropertiesEntry\022\013\n\003k" - + "ey\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"D\n\014JobPlacem" - + "ent\022\031\n\014cluster_name\030\001 \001(\tB\003\340A\002\022\031\n\014cluste" - + "r_uuid\030\002 \001(\tB\003\340A\003\"\340\003\n\tJobStatus\022B\n\005state" - + "\030\001 \001(\0162..google.cloud.dataproc.v1beta2.J" - + "obStatus.StateB\003\340A\003\022\024\n\007details\030\002 \001(\tB\003\340A" - + "\003\0229\n\020state_start_time\030\006 \001(\0132\032.google.pro" - + "tobuf.TimestampB\003\340A\003\022H\n\010substate\030\007 \001(\01621" - + ".google.cloud.dataproc.v1beta2.JobStatus" - + ".SubstateB\003\340A\003\"\251\001\n\005State\022\025\n\021STATE_UNSPEC" - + "IFIED\020\000\022\013\n\007PENDING\020\001\022\016\n\nSETUP_DONE\020\010\022\013\n\007" - + "RUNNING\020\002\022\022\n\016CANCEL_PENDING\020\003\022\022\n\016CANCEL_" - + "STARTED\020\007\022\r\n\tCANCELLED\020\004\022\010\n\004DONE\020\005\022\t\n\005ER" - + "ROR\020\006\022\023\n\017ATTEMPT_FAILURE\020\t\"H\n\010Substate\022\017" - + "\n\013UNSPECIFIED\020\000\022\r\n\tSUBMITTED\020\001\022\n\n\006QUEUED" - + "\020\002\022\020\n\014STALE_STATUS\020\003\"<\n\014JobReference\022\027\n\n" - + "project_id\030\001 \001(\tB\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340" - + "A\001\"\252\002\n\017YarnApplication\022\021\n\004name\030\001 \001(\tB\003\340A" - + "\003\022H\n\005state\030\002 \001(\01624.google.cloud.dataproc" - + ".v1beta2.YarnApplication.StateB\003\340A\003\022\025\n\010p" - + "rogress\030\003 \001(\002B\003\340A\003\022\031\n\014tracking_url\030\004 \001(\t" - + "B\003\340A\003\"\207\001\n\005State\022\025\n\021STATE_UNSPECIFIED\020\000\022\007" - + "\n\003NEW\020\001\022\016\n\nNEW_SAVING\020\002\022\r\n\tSUBMITTED\020\003\022\014" - + "\n\010ACCEPTED\020\004\022\013\n\007RUNNING\020\005\022\014\n\010FINISHED\020\006\022" - + "\n\n\006FAILED\020\007\022\n\n\006KILLED\020\010\"\352\010\n\003Job\022C\n\trefer" - + "ence\030\001 \001(\0132+.google.cloud.dataproc.v1bet" - + "a2.JobReferenceB\003\340A\001\022C\n\tplacement\030\002 \001(\0132" - + "+.google.cloud.dataproc.v1beta2.JobPlace" - + "mentB\003\340A\002\022>\n\nhadoop_job\030\003 \001(\0132(.google.c" - + "loud.dataproc.v1beta2.HadoopJobH\000\022<\n\tspa" - + "rk_job\030\004 \001(\0132\'.google.cloud.dataproc.v1b" - + "eta2.SparkJobH\000\022@\n\013pyspark_job\030\005 \001(\0132).g" - + "oogle.cloud.dataproc.v1beta2.PySparkJobH" - + "\000\022:\n\010hive_job\030\006 \001(\0132&.google.cloud.datap" - + "roc.v1beta2.HiveJobH\000\0228\n\007pig_job\030\007 \001(\0132%" - + ".google.cloud.dataproc.v1beta2.PigJobH\000\022" - + "?\n\013spark_r_job\030\025 \001(\0132(.google.cloud.data" - + "proc.v1beta2.SparkRJobH\000\022C\n\rspark_sql_jo" - + "b\030\014 \001(\0132*.google.cloud.dataproc.v1beta2." - + "SparkSqlJobH\000\022=\n\006status\030\010 \001(\0132(.google.c" - + "loud.dataproc.v1beta2.JobStatusB\003\340A\003\022E\n\016" - + "status_history\030\r \003(\0132(.google.cloud.data" - + "proc.v1beta2.JobStatusB\003\340A\003\022N\n\021yarn_appl" - + "ications\030\t \003(\0132..google.cloud.dataproc.v" - + "1beta2.YarnApplicationB\003\340A\003\022\031\n\014submitted" - + "_by\030\n \001(\tB\003\340A\003\022\'\n\032driver_output_resource" - + "_uri\030\021 \001(\tB\003\340A\003\022%\n\030driver_control_files_" - + "uri\030\017 \001(\tB\003\340A\003\022C\n\006labels\030\022 \003(\0132..google." - + "cloud.dataproc.v1beta2.Job.LabelsEntryB\003" - + "\340A\001\022E\n\nscheduling\030\024 \001(\0132,.google.cloud.d" - + "ataproc.v1beta2.JobSchedulingB\003\340A\001\022\025\n\010jo" - + "b_uuid\030\026 \001(\tB\003\340A\003\032-\n\013LabelsEntry\022\013\n\003key\030" - + "\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001B\n\n\010type_job\"3\n\r" - + "JobScheduling\022\"\n\025max_failures_per_hour\030\001" - + " \001(\005B\003\340A\001\"\217\001\n\020SubmitJobRequest\022\027\n\nprojec" - + "t_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\0224\n\003" - + "job\030\002 \001(\0132\".google.cloud.dataproc.v1beta" - + "2.JobB\003\340A\002\022\027\n\nrequest_id\030\004 \001(\tB\003\340A\001\"R\n\rG" - + "etJobRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n" - + "\006region\030\003 \001(\tB\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\002\"" - + "\270\002\n\017ListJobsRequest\022\027\n\nproject_id\030\001 \001(\tB" - + "\003\340A\002\022\023\n\006region\030\006 \001(\tB\003\340A\002\022\026\n\tpage_size\030\002" - + " \001(\005B\003\340A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001\022\031\n\014cl" - + "uster_name\030\004 \001(\tB\003\340A\001\022^\n\021job_state_match" - + "er\030\005 \001(\0162>.google.cloud.dataproc.v1beta2" - + ".ListJobsRequest.JobStateMatcherB\003\340A\001\022\023\n" - + "\006filter\030\007 \001(\tB\003\340A\001\"6\n\017JobStateMatcher\022\007\n" - + "\003ALL\020\000\022\n\n\006ACTIVE\020\001\022\016\n\nNON_ACTIVE\020\002\"\301\001\n\020U" - + "pdateJobRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002" - + "\022\023\n\006region\030\002 \001(\tB\003\340A\002\022\023\n\006job_id\030\003 \001(\tB\003\340" - + "A\002\0224\n\003job\030\004 \001(\0132\".google.cloud.dataproc." - + "v1beta2.JobB\003\340A\002\0224\n\013update_mask\030\005 \001(\0132\032." - + "google.protobuf.FieldMaskB\003\340A\002\"g\n\020ListJo" - + "bsResponse\0225\n\004jobs\030\001 \003(\0132\".google.cloud." - + "dataproc.v1beta2.JobB\003\340A\003\022\034\n\017next_page_t" - + "oken\030\002 \001(\tB\003\340A\001\"U\n\020CancelJobRequest\022\027\n\np" - + "roject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A" - + "\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\002\"U\n\020DeleteJobReque" - + "st\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 " - + "\001(\tB\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\0022\373\t\n\rJobCon" - + "troller\022\302\001\n\tSubmitJob\022/.google.cloud.dat" - + "aproc.v1beta2.SubmitJobRequest\032\".google." - + "cloud.dataproc.v1beta2.Job\"`\202\323\344\223\002@\";/v1b" - + "eta2/projects/{project_id}/regions/{regi" - + "on}/jobs:submit:\001*\332A\027project_id, region," - + " job\022\276\001\n\006GetJob\022,.google.cloud.dataproc." - + "v1beta2.GetJobRequest\032\".google.cloud.dat" - + "aproc.v1beta2.Job\"b\202\323\344\223\002?\022=/v1beta2/proj" - + "ects/{project_id}/regions/{region}/jobs/" - + "{job_id}\332A\032project_id, region, job_id\022\333\001" - + "\n\010ListJobs\022..google.cloud.dataproc.v1bet" - + "a2.ListJobsRequest\032/.google.cloud.datapr" - + "oc.v1beta2.ListJobsResponse\"n\202\323\344\223\0026\0224/v1" + + "ey\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\231\003\n\tPrestoJo" + + "b\022\030\n\016query_file_uri\030\001 \001(\tH\000\022>\n\nquery_lis" + + "t\030\002 \001(\0132(.google.cloud.dataproc.v1beta2." + + "QueryListH\000\022 \n\023continue_on_failure\030\003 \001(\010" + + "B\003\340A\001\022\032\n\routput_format\030\004 \001(\tB\003\340A\001\022\030\n\013cli" + + "ent_tags\030\005 \003(\tB\003\340A\001\022Q\n\nproperties\030\006 \003(\0132" + + "8.google.cloud.dataproc.v1beta2.PrestoJo" + + "b.PropertiesEntryB\003\340A\001\022I\n\016logging_config" + + "\030\007 \001(\0132,.google.cloud.dataproc.v1beta2.L" + + "oggingConfigB\003\340A\001\0321\n\017PropertiesEntry\022\013\n\003" + + "key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001B\t\n\007queries\"" + + "D\n\014JobPlacement\022\031\n\014cluster_name\030\001 \001(\tB\003\340" + + "A\002\022\031\n\014cluster_uuid\030\002 \001(\tB\003\340A\003\"\340\003\n\tJobSta" + + "tus\022B\n\005state\030\001 \001(\0162..google.cloud.datapr" + + "oc.v1beta2.JobStatus.StateB\003\340A\003\022\024\n\007detai" + + "ls\030\002 \001(\tB\003\340A\003\0229\n\020state_start_time\030\006 \001(\0132" + + "\032.google.protobuf.TimestampB\003\340A\003\022H\n\010subs" + + "tate\030\007 \001(\01621.google.cloud.dataproc.v1bet" + + "a2.JobStatus.SubstateB\003\340A\003\"\251\001\n\005State\022\025\n\021" + + "STATE_UNSPECIFIED\020\000\022\013\n\007PENDING\020\001\022\016\n\nSETU" + + "P_DONE\020\010\022\013\n\007RUNNING\020\002\022\022\n\016CANCEL_PENDING\020" + + "\003\022\022\n\016CANCEL_STARTED\020\007\022\r\n\tCANCELLED\020\004\022\010\n\004" + + "DONE\020\005\022\t\n\005ERROR\020\006\022\023\n\017ATTEMPT_FAILURE\020\t\"H" + + "\n\010Substate\022\017\n\013UNSPECIFIED\020\000\022\r\n\tSUBMITTED" + + "\020\001\022\n\n\006QUEUED\020\002\022\020\n\014STALE_STATUS\020\003\"<\n\014JobR" + + "eference\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006job" + + "_id\030\002 \001(\tB\003\340A\001\"\252\002\n\017YarnApplication\022\021\n\004na" + + "me\030\001 \001(\tB\003\340A\003\022H\n\005state\030\002 \001(\01624.google.cl" + + "oud.dataproc.v1beta2.YarnApplication.Sta" + + "teB\003\340A\003\022\025\n\010progress\030\003 \001(\002B\003\340A\003\022\031\n\014tracki" + + "ng_url\030\004 \001(\tB\003\340A\003\"\207\001\n\005State\022\025\n\021STATE_UNS" + + "PECIFIED\020\000\022\007\n\003NEW\020\001\022\016\n\nNEW_SAVING\020\002\022\r\n\tS" + + "UBMITTED\020\003\022\014\n\010ACCEPTED\020\004\022\013\n\007RUNNING\020\005\022\014\n" + + "\010FINISHED\020\006\022\n\n\006FAILED\020\007\022\n\n\006KILLED\020\010\"\345\t\n\003" + + "Job\022C\n\treference\030\001 \001(\0132+.google.cloud.da" + + "taproc.v1beta2.JobReferenceB\003\340A\001\022C\n\tplac" + + "ement\030\002 \001(\0132+.google.cloud.dataproc.v1be" + + "ta2.JobPlacementB\003\340A\002\022C\n\nhadoop_job\030\003 \001(" + + "\0132(.google.cloud.dataproc.v1beta2.Hadoop" + + "JobB\003\340A\001H\000\022A\n\tspark_job\030\004 \001(\0132\'.google.c" + + "loud.dataproc.v1beta2.SparkJobB\003\340A\001H\000\022E\n" + + "\013pyspark_job\030\005 \001(\0132).google.cloud.datapr" + + "oc.v1beta2.PySparkJobB\003\340A\001H\000\022?\n\010hive_job" + + "\030\006 \001(\0132&.google.cloud.dataproc.v1beta2.H" + + "iveJobB\003\340A\001H\000\022=\n\007pig_job\030\007 \001(\0132%.google." + + "cloud.dataproc.v1beta2.PigJobB\003\340A\001H\000\022D\n\013" + + "spark_r_job\030\025 \001(\0132(.google.cloud.datapro" + + "c.v1beta2.SparkRJobB\003\340A\001H\000\022H\n\rspark_sql_" + + "job\030\014 \001(\0132*.google.cloud.dataproc.v1beta" + + "2.SparkSqlJobB\003\340A\001H\000\022C\n\npresto_job\030\027 \001(\013" + + "2(.google.cloud.dataproc.v1beta2.PrestoJ" + + "obB\003\340A\001H\000\022=\n\006status\030\010 \001(\0132(.google.cloud" + + ".dataproc.v1beta2.JobStatusB\003\340A\003\022E\n\016stat" + + "us_history\030\r \003(\0132(.google.cloud.dataproc" + + ".v1beta2.JobStatusB\003\340A\003\022N\n\021yarn_applicat" + + "ions\030\t \003(\0132..google.cloud.dataproc.v1bet" + + "a2.YarnApplicationB\003\340A\003\022\031\n\014submitted_by\030" + + "\n \001(\tB\003\340A\003\022\'\n\032driver_output_resource_uri" + + "\030\021 \001(\tB\003\340A\003\022%\n\030driver_control_files_uri\030" + + "\017 \001(\tB\003\340A\003\022C\n\006labels\030\022 \003(\0132..google.clou" + + "d.dataproc.v1beta2.Job.LabelsEntryB\003\340A\001\022" + + "E\n\nscheduling\030\024 \001(\0132,.google.cloud.datap" + + "roc.v1beta2.JobSchedulingB\003\340A\001\022\025\n\010job_uu" + + "id\030\026 \001(\tB\003\340A\003\022\021\n\004done\030\030 \001(\010B\003\340A\003\032-\n\013Labe" + + "lsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001B" + + "\n\n\010type_job\"3\n\rJobScheduling\022\"\n\025max_fail" + + "ures_per_hour\030\001 \001(\005B\003\340A\001\"\263\001\n\013JobMetadata" + + "\022\023\n\006job_id\030\001 \001(\tB\003\340A\003\022=\n\006status\030\002 \001(\0132(." + + "google.cloud.dataproc.v1beta2.JobStatusB" + + "\003\340A\003\022\033\n\016operation_type\030\003 \001(\tB\003\340A\003\0223\n\nsta" + + "rt_time\030\004 \001(\0132\032.google.protobuf.Timestam" + + "pB\003\340A\003\"\217\001\n\020SubmitJobRequest\022\027\n\nproject_i" + + "d\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\0224\n\003job" + + "\030\002 \001(\0132\".google.cloud.dataproc.v1beta2.J" + + "obB\003\340A\002\022\027\n\nrequest_id\030\004 \001(\tB\003\340A\001\"R\n\rGetJ" + + "obRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006re" + + "gion\030\003 \001(\tB\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\002\"\270\002\n" + + "\017ListJobsRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A" + + "\002\022\023\n\006region\030\006 \001(\tB\003\340A\002\022\026\n\tpage_size\030\002 \001(" + + "\005B\003\340A\001\022\027\n\npage_token\030\003 \001(\tB\003\340A\001\022\031\n\014clust" + + "er_name\030\004 \001(\tB\003\340A\001\022^\n\021job_state_matcher\030" + + "\005 \001(\0162>.google.cloud.dataproc.v1beta2.Li" + + "stJobsRequest.JobStateMatcherB\003\340A\001\022\023\n\006fi" + + "lter\030\007 \001(\tB\003\340A\001\"6\n\017JobStateMatcher\022\007\n\003AL" + + "L\020\000\022\n\n\006ACTIVE\020\001\022\016\n\nNON_ACTIVE\020\002\"\301\001\n\020Upda" + + "teJobRequest\022\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n" + + "\006region\030\002 \001(\tB\003\340A\002\022\023\n\006job_id\030\003 \001(\tB\003\340A\002\022" + + "4\n\003job\030\004 \001(\0132\".google.cloud.dataproc.v1b" + + "eta2.JobB\003\340A\002\0224\n\013update_mask\030\005 \001(\0132\032.goo" + + "gle.protobuf.FieldMaskB\003\340A\002\"g\n\020ListJobsR" + + "esponse\0225\n\004jobs\030\001 \003(\0132\".google.cloud.dat" + + "aproc.v1beta2.JobB\003\340A\003\022\034\n\017next_page_toke" + + "n\030\002 \001(\tB\003\340A\001\"U\n\020CancelJobRequest\022\027\n\nproj" + + "ect_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\tB\003\340A\002\022\023" + + "\n\006job_id\030\002 \001(\tB\003\340A\002\"U\n\020DeleteJobRequest\022" + + "\027\n\nproject_id\030\001 \001(\tB\003\340A\002\022\023\n\006region\030\003 \001(\t" + + "B\003\340A\002\022\023\n\006job_id\030\002 \001(\tB\003\340A\0022\347\013\n\rJobContro" + + "ller\022\302\001\n\tSubmitJob\022/.google.cloud.datapr" + + "oc.v1beta2.SubmitJobRequest\032\".google.clo" + + "ud.dataproc.v1beta2.Job\"`\202\323\344\223\002@\";/v1beta" + + "2/projects/{project_id}/regions/{region}" + + "/jobs:submit:\001*\332A\027project_id, region, jo" + + "b\022\351\001\n\024SubmitJobAsOperation\022/.google.clou" + + "d.dataproc.v1beta2.SubmitJobRequest\032\035.go" + + "ogle.longrunning.Operation\"\200\001\202\323\344\223\002K\"F/v1" + "beta2/projects/{project_id}/regions/{reg" - + "ion}/jobs\332A\022project_id, region\332A\032project" - + "_id, region, filter\022\254\001\n\tUpdateJob\022/.goog" - + "le.cloud.dataproc.v1beta2.UpdateJobReque" - + "st\032\".google.cloud.dataproc.v1beta2.Job\"J" - + "\202\323\344\223\002D2=/v1beta2/projects/{project_id}/r" - + "egions/{region}/jobs/{job_id}:\003job\022\316\001\n\tC" - + "ancelJob\022/.google.cloud.dataproc.v1beta2" - + ".CancelJobRequest\032\".google.cloud.datapro" - + "c.v1beta2.Job\"l\202\323\344\223\002I\"D/v1beta2/projects" - + "/{project_id}/regions/{region}/jobs/{job" - + "_id}:cancel:\001*\332A\032project_id, region, job" - + "_id\022\270\001\n\tDeleteJob\022/.google.cloud.datapro" - + "c.v1beta2.DeleteJobRequest\032\026.google.prot" - + "obuf.Empty\"b\202\323\344\223\002?*=/v1beta2/projects/{p" + + "ion}/jobs:submitAsOperation:\001*\332A\027project" + + "_id, region, job\312A\022\n\003Job\022\013JobMetadata\022\276\001" + + "\n\006GetJob\022,.google.cloud.dataproc.v1beta2" + + ".GetJobRequest\032\".google.cloud.dataproc.v" + + "1beta2.Job\"b\202\323\344\223\002?\022=/v1beta2/projects/{p" + "roject_id}/regions/{region}/jobs/{job_id" - + "}\332A\032project_id, region, job_id\032K\312A\027datap" - + "roc.googleapis.com\322A.https://www.googlea" - + "pis.com/auth/cloud-platformBw\n!com.googl" - + "e.cloud.dataproc.v1beta2B\tJobsProtoP\001ZEg" - + "oogle.golang.org/genproto/googleapis/clo" - + "ud/dataproc/v1beta2;dataprocb\006proto3" + + "}\332A\032project_id, region, job_id\022\333\001\n\010ListJ" + + "obs\022..google.cloud.dataproc.v1beta2.List" + + "JobsRequest\032/.google.cloud.dataproc.v1be" + + "ta2.ListJobsResponse\"n\202\323\344\223\0026\0224/v1beta2/p" + + "rojects/{project_id}/regions/{region}/jo" + + "bs\332A\022project_id, region\332A\032project_id, re" + + "gion, filter\022\254\001\n\tUpdateJob\022/.google.clou" + + "d.dataproc.v1beta2.UpdateJobRequest\032\".go" + + "ogle.cloud.dataproc.v1beta2.Job\"J\202\323\344\223\002D2" + + "=/v1beta2/projects/{project_id}/regions/" + + "{region}/jobs/{job_id}:\003job\022\316\001\n\tCancelJo" + + "b\022/.google.cloud.dataproc.v1beta2.Cancel" + + "JobRequest\032\".google.cloud.dataproc.v1bet" + + "a2.Job\"l\202\323\344\223\002I\"D/v1beta2/projects/{proje" + + "ct_id}/regions/{region}/jobs/{job_id}:ca" + + "ncel:\001*\332A\032project_id, region, job_id\022\270\001\n" + + "\tDeleteJob\022/.google.cloud.dataproc.v1bet" + + "a2.DeleteJobRequest\032\026.google.protobuf.Em" + + "pty\"b\202\323\344\223\002?*=/v1beta2/projects/{project_" + + "id}/regions/{region}/jobs/{job_id}\332A\032pro" + + "ject_id, region, job_id\032K\312A\027dataproc.goo" + + "gleapis.com\322A.https://www.googleapis.com" + + "/auth/cloud-platformBw\n!com.google.cloud" + + ".dataproc.v1beta2B\tJobsProtoP\001ZEgoogle.g" + + "olang.org/genproto/googleapis/cloud/data" + + "proc/v1beta2;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( @@ -597,8 +633,31 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { new java.lang.String[] { "Key", "Value", }); - internal_static_google_cloud_dataproc_v1beta2_JobPlacement_descriptor = + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_descriptor = getDescriptor().getMessageTypes().get(9); + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_fieldAccessorTable = + new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_descriptor, + new java.lang.String[] { + "QueryFileUri", + "QueryList", + "ContinueOnFailure", + "OutputFormat", + "ClientTags", + "Properties", + "LoggingConfig", + "Queries", + }); + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_PropertiesEntry_descriptor = + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_descriptor.getNestedTypes().get(0); + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_PropertiesEntry_fieldAccessorTable = + new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_google_cloud_dataproc_v1beta2_PrestoJob_PropertiesEntry_descriptor, + new java.lang.String[] { + "Key", "Value", + }); + internal_static_google_cloud_dataproc_v1beta2_JobPlacement_descriptor = + getDescriptor().getMessageTypes().get(10); internal_static_google_cloud_dataproc_v1beta2_JobPlacement_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_JobPlacement_descriptor, @@ -606,7 +665,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ClusterName", "ClusterUuid", }); internal_static_google_cloud_dataproc_v1beta2_JobStatus_descriptor = - getDescriptor().getMessageTypes().get(10); + getDescriptor().getMessageTypes().get(11); internal_static_google_cloud_dataproc_v1beta2_JobStatus_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_JobStatus_descriptor, @@ -614,7 +673,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "State", "Details", "StateStartTime", "Substate", }); internal_static_google_cloud_dataproc_v1beta2_JobReference_descriptor = - getDescriptor().getMessageTypes().get(11); + getDescriptor().getMessageTypes().get(12); internal_static_google_cloud_dataproc_v1beta2_JobReference_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_JobReference_descriptor, @@ -622,7 +681,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "JobId", }); internal_static_google_cloud_dataproc_v1beta2_YarnApplication_descriptor = - getDescriptor().getMessageTypes().get(12); + getDescriptor().getMessageTypes().get(13); internal_static_google_cloud_dataproc_v1beta2_YarnApplication_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_YarnApplication_descriptor, @@ -630,7 +689,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Name", "State", "Progress", "TrackingUrl", }); internal_static_google_cloud_dataproc_v1beta2_Job_descriptor = - getDescriptor().getMessageTypes().get(13); + getDescriptor().getMessageTypes().get(14); internal_static_google_cloud_dataproc_v1beta2_Job_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_Job_descriptor, @@ -644,6 +703,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "PigJob", "SparkRJob", "SparkSqlJob", + "PrestoJob", "Status", "StatusHistory", "YarnApplications", @@ -653,6 +713,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Labels", "Scheduling", "JobUuid", + "Done", "TypeJob", }); internal_static_google_cloud_dataproc_v1beta2_Job_LabelsEntry_descriptor = @@ -664,15 +725,23 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Key", "Value", }); internal_static_google_cloud_dataproc_v1beta2_JobScheduling_descriptor = - getDescriptor().getMessageTypes().get(14); + getDescriptor().getMessageTypes().get(15); internal_static_google_cloud_dataproc_v1beta2_JobScheduling_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_JobScheduling_descriptor, new java.lang.String[] { "MaxFailuresPerHour", }); + internal_static_google_cloud_dataproc_v1beta2_JobMetadata_descriptor = + getDescriptor().getMessageTypes().get(16); + internal_static_google_cloud_dataproc_v1beta2_JobMetadata_fieldAccessorTable = + new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_google_cloud_dataproc_v1beta2_JobMetadata_descriptor, + new java.lang.String[] { + "JobId", "Status", "OperationType", "StartTime", + }); internal_static_google_cloud_dataproc_v1beta2_SubmitJobRequest_descriptor = - getDescriptor().getMessageTypes().get(15); + getDescriptor().getMessageTypes().get(17); internal_static_google_cloud_dataproc_v1beta2_SubmitJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_SubmitJobRequest_descriptor, @@ -680,7 +749,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "Job", "RequestId", }); internal_static_google_cloud_dataproc_v1beta2_GetJobRequest_descriptor = - getDescriptor().getMessageTypes().get(16); + getDescriptor().getMessageTypes().get(18); internal_static_google_cloud_dataproc_v1beta2_GetJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_GetJobRequest_descriptor, @@ -688,7 +757,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "JobId", }); internal_static_google_cloud_dataproc_v1beta2_ListJobsRequest_descriptor = - getDescriptor().getMessageTypes().get(17); + getDescriptor().getMessageTypes().get(19); internal_static_google_cloud_dataproc_v1beta2_ListJobsRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_ListJobsRequest_descriptor, @@ -702,7 +771,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Filter", }); internal_static_google_cloud_dataproc_v1beta2_UpdateJobRequest_descriptor = - getDescriptor().getMessageTypes().get(18); + getDescriptor().getMessageTypes().get(20); internal_static_google_cloud_dataproc_v1beta2_UpdateJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_UpdateJobRequest_descriptor, @@ -710,7 +779,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "JobId", "Job", "UpdateMask", }); internal_static_google_cloud_dataproc_v1beta2_ListJobsResponse_descriptor = - getDescriptor().getMessageTypes().get(19); + getDescriptor().getMessageTypes().get(21); internal_static_google_cloud_dataproc_v1beta2_ListJobsResponse_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_ListJobsResponse_descriptor, @@ -718,7 +787,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "Jobs", "NextPageToken", }); internal_static_google_cloud_dataproc_v1beta2_CancelJobRequest_descriptor = - getDescriptor().getMessageTypes().get(20); + getDescriptor().getMessageTypes().get(22); internal_static_google_cloud_dataproc_v1beta2_CancelJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_CancelJobRequest_descriptor, @@ -726,7 +795,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "ProjectId", "Region", "JobId", }); internal_static_google_cloud_dataproc_v1beta2_DeleteJobRequest_descriptor = - getDescriptor().getMessageTypes().get(21); + getDescriptor().getMessageTypes().get(23); internal_static_google_cloud_dataproc_v1beta2_DeleteJobRequest_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dataproc_v1beta2_DeleteJobRequest_descriptor, @@ -740,6 +809,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { registry.add(com.google.api.AnnotationsProto.http); registry.add(com.google.api.ClientProto.methodSignature); registry.add(com.google.api.ClientProto.oauthScopes); + registry.add(com.google.longrunning.OperationsProto.operationInfo); com.google.protobuf.Descriptors.FileDescriptor.internalUpdateFileDescriptor( descriptor, registry); com.google.api.AnnotationsProto.getDescriptor(); diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfig.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfig.java index 9a7d39d0..e73ebe5e 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfig.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfig.java @@ -223,7 +223,8 @@ public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { * * *
-   * Optional. Flag to indicate whether to Kerberize the cluster.
+   * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+   * this field to true to enable Kerberos on a cluster.
    * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1459,7 +1460,8 @@ public Builder mergeFrom( * * *
-     * Optional. Flag to indicate whether to Kerberize the cluster.
+     * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+     * this field to true to enable Kerberos on a cluster.
      * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1473,7 +1475,8 @@ public boolean getEnableKerberos() { * * *
-     * Optional. Flag to indicate whether to Kerberize the cluster.
+     * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+     * this field to true to enable Kerberos on a cluster.
      * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; @@ -1491,7 +1494,8 @@ public Builder setEnableKerberos(boolean value) { * * *
-     * Optional. Flag to indicate whether to Kerberize the cluster.
+     * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+     * this field to true to enable Kerberos on a cluster.
      * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfigOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfigOrBuilder.java index 3e50f8b6..a947cb29 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/KerberosConfigOrBuilder.java @@ -27,7 +27,8 @@ public interface KerberosConfigOrBuilder * * *
-   * Optional. Flag to indicate whether to Kerberize the cluster.
+   * Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set
+   * this field to true to enable Kerberos on a cluster.
    * 
* * bool enable_kerberos = 1 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJob.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJob.java index 711cd541..b01dbd53 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJob.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJob.java @@ -214,6 +214,38 @@ private OrderedJob( prerequisiteStepIds_.add(s); break; } + case 90: + { + com.google.cloud.dataproc.v1beta2.SparkRJob.Builder subBuilder = null; + if (jobTypeCase_ == 11) { + subBuilder = ((com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_).toBuilder(); + } + jobType_ = + input.readMessage( + com.google.cloud.dataproc.v1beta2.SparkRJob.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom((com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_); + jobType_ = subBuilder.buildPartial(); + } + jobTypeCase_ = 11; + break; + } + case 98: + { + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder subBuilder = null; + if (jobTypeCase_ == 12) { + subBuilder = ((com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_).toBuilder(); + } + jobType_ = + input.readMessage( + com.google.cloud.dataproc.v1beta2.PrestoJob.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom((com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_); + jobType_ = subBuilder.buildPartial(); + } + jobTypeCase_ = 12; + break; + } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { @@ -274,7 +306,9 @@ public enum JobTypeCase PYSPARK_JOB(4), HIVE_JOB(5), PIG_JOB(6), + SPARK_R_JOB(11), SPARK_SQL_JOB(7), + PRESTO_JOB(12), JOBTYPE_NOT_SET(0); private final int value; @@ -303,8 +337,12 @@ public static JobTypeCase forNumber(int value) { return HIVE_JOB; case 6: return PIG_JOB; + case 11: + return SPARK_R_JOB; case 7: return SPARK_SQL_JOB; + case 12: + return PRESTO_JOB; case 0: return JOBTYPE_NOT_SET; default: @@ -331,14 +369,14 @@ public JobTypeCase getJobTypeCase() { * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The stepId. */ @@ -361,14 +399,14 @@ public java.lang.String getStepId() { * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for stepId. */ @@ -524,6 +562,54 @@ public com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder() { return com.google.cloud.dataproc.v1beta2.PigJob.getDefaultInstance(); } + public static final int SPARK_R_JOB_FIELD_NUMBER = 11; + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + * + * @return Whether the sparkRJob field is set. + */ + public boolean hasSparkRJob() { + return jobTypeCase_ == 11; + } + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + * + * @return The sparkRJob. + */ + public com.google.cloud.dataproc.v1beta2.SparkRJob getSparkRJob() { + if (jobTypeCase_ == 11) { + return (com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_; + } + return com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); + } + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + public com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilder() { + if (jobTypeCase_ == 11) { + return (com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_; + } + return com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); + } + public static final int SPARK_SQL_JOB_FIELD_NUMBER = 7; /** * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 7; @@ -552,6 +638,54 @@ public com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBu return com.google.cloud.dataproc.v1beta2.SparkSqlJob.getDefaultInstance(); } + public static final int PRESTO_JOB_FIELD_NUMBER = 12; + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + * + * @return Whether the prestoJob field is set. + */ + public boolean hasPrestoJob() { + return jobTypeCase_ == 12; + } + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + * + * @return The prestoJob. + */ + public com.google.cloud.dataproc.v1beta2.PrestoJob getPrestoJob() { + if (jobTypeCase_ == 12) { + return (com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_; + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + public com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder getPrestoJobOrBuilder() { + if (jobTypeCase_ == 12) { + return (com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_; + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + public static final int LABELS_FIELD_NUMBER = 8; private static final class LabelsDefaultEntryHolder { @@ -590,7 +724,7 @@ public int getLabelsCount() { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public boolean containsLabels(java.lang.String key) { if (key == null) { @@ -616,7 +750,7 @@ public java.util.Map getLabels() { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public java.util.Map getLabelsMap() { return internalGetLabels().getMap(); @@ -634,7 +768,7 @@ public java.util.Map getLabelsMap() { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public java.lang.String getLabelsOrDefault(java.lang.String key, java.lang.String defaultValue) { if (key == null) { @@ -656,7 +790,7 @@ public java.lang.String getLabelsOrDefault(java.lang.String key, java.lang.Strin * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public java.lang.String getLabelsOrThrow(java.lang.String key) { if (key == null) { @@ -678,7 +812,9 @@ public java.lang.String getLabelsOrThrow(java.lang.String key) { * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the scheduling field is set. */ @@ -692,7 +828,9 @@ public boolean hasScheduling() { * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The scheduling. */ @@ -708,7 +846,9 @@ public com.google.cloud.dataproc.v1beta2.JobScheduling getScheduling() { * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ public com.google.cloud.dataproc.v1beta2.JobSchedulingOrBuilder getSchedulingOrBuilder() { return getScheduling(); @@ -724,7 +864,8 @@ public com.google.cloud.dataproc.v1beta2.JobSchedulingOrBuilder getSchedulingOrB * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return A list containing the prerequisiteStepIds. */ @@ -739,7 +880,8 @@ public com.google.protobuf.ProtocolStringList getPrerequisiteStepIdsList() { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The count of prerequisiteStepIds. */ @@ -754,7 +896,8 @@ public int getPrerequisiteStepIdsCount() { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param index The index of the element to return. * @return The prerequisiteStepIds at the given index. @@ -770,7 +913,8 @@ public java.lang.String getPrerequisiteStepIds(int index) { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param index The index of the value to return. * @return The bytes of the prerequisiteStepIds at the given index. @@ -823,6 +967,12 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io com.google.protobuf.GeneratedMessageV3.writeString( output, 10, prerequisiteStepIds_.getRaw(i)); } + if (jobTypeCase_ == 11) { + output.writeMessage(11, (com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_); + } + if (jobTypeCase_ == 12) { + output.writeMessage(12, (com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_); + } unknownFields.writeTo(output); } @@ -886,6 +1036,16 @@ public int getSerializedSize() { size += dataSize; size += 1 * getPrerequisiteStepIdsList().size(); } + if (jobTypeCase_ == 11) { + size += + com.google.protobuf.CodedOutputStream.computeMessageSize( + 11, (com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_); + } + if (jobTypeCase_ == 12) { + size += + com.google.protobuf.CodedOutputStream.computeMessageSize( + 12, (com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_); + } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; @@ -926,9 +1086,15 @@ public boolean equals(final java.lang.Object obj) { case 6: if (!getPigJob().equals(other.getPigJob())) return false; break; + case 11: + if (!getSparkRJob().equals(other.getSparkRJob())) return false; + break; case 7: if (!getSparkSqlJob().equals(other.getSparkSqlJob())) return false; break; + case 12: + if (!getPrestoJob().equals(other.getPrestoJob())) return false; + break; case 0: default: } @@ -978,10 +1144,18 @@ public int hashCode() { hash = (37 * hash) + PIG_JOB_FIELD_NUMBER; hash = (53 * hash) + getPigJob().hashCode(); break; + case 11: + hash = (37 * hash) + SPARK_R_JOB_FIELD_NUMBER; + hash = (53 * hash) + getSparkRJob().hashCode(); + break; case 7: hash = (37 * hash) + SPARK_SQL_JOB_FIELD_NUMBER; hash = (53 * hash) + getSparkSqlJob().hashCode(); break; + case 12: + hash = (37 * hash) + PRESTO_JOB_FIELD_NUMBER; + hash = (53 * hash) + getPrestoJob().hashCode(); + break; case 0: default: } @@ -1227,6 +1401,13 @@ public com.google.cloud.dataproc.v1beta2.OrderedJob buildPartial() { result.jobType_ = pigJobBuilder_.build(); } } + if (jobTypeCase_ == 11) { + if (sparkRJobBuilder_ == null) { + result.jobType_ = jobType_; + } else { + result.jobType_ = sparkRJobBuilder_.build(); + } + } if (jobTypeCase_ == 7) { if (sparkSqlJobBuilder_ == null) { result.jobType_ = jobType_; @@ -1234,6 +1415,13 @@ public com.google.cloud.dataproc.v1beta2.OrderedJob buildPartial() { result.jobType_ = sparkSqlJobBuilder_.build(); } } + if (jobTypeCase_ == 12) { + if (prestoJobBuilder_ == null) { + result.jobType_ = jobType_; + } else { + result.jobType_ = prestoJobBuilder_.build(); + } + } result.labels_ = internalGetLabels(); result.labels_.makeImmutable(); if (schedulingBuilder_ == null) { @@ -1340,11 +1528,21 @@ public Builder mergeFrom(com.google.cloud.dataproc.v1beta2.OrderedJob other) { mergePigJob(other.getPigJob()); break; } + case SPARK_R_JOB: + { + mergeSparkRJob(other.getSparkRJob()); + break; + } case SPARK_SQL_JOB: { mergeSparkSqlJob(other.getSparkSqlJob()); break; } + case PRESTO_JOB: + { + mergePrestoJob(other.getPrestoJob()); + break; + } case JOBTYPE_NOT_SET: { break; @@ -1404,14 +1602,14 @@ public Builder clearJobType() { * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The stepId. */ @@ -1434,14 +1632,14 @@ public java.lang.String getStepId() { * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for stepId. */ @@ -1464,14 +1662,14 @@ public com.google.protobuf.ByteString getStepIdBytes() { * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @param value The stepId to set. * @return This builder for chaining. @@ -1493,14 +1691,14 @@ public Builder setStepId(java.lang.String value) { * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return This builder for chaining. */ @@ -1518,14 +1716,14 @@ public Builder clearStepId() { * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @param value The bytes for stepId to set. * @return This builder for chaining. @@ -2228,6 +2426,212 @@ public com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder() { return pigJobBuilder_; } + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.SparkRJob, + com.google.cloud.dataproc.v1beta2.SparkRJob.Builder, + com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder> + sparkRJobBuilder_; + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + * + * @return Whether the sparkRJob field is set. + */ + public boolean hasSparkRJob() { + return jobTypeCase_ == 11; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + * + * @return The sparkRJob. + */ + public com.google.cloud.dataproc.v1beta2.SparkRJob getSparkRJob() { + if (sparkRJobBuilder_ == null) { + if (jobTypeCase_ == 11) { + return (com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_; + } + return com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); + } else { + if (jobTypeCase_ == 11) { + return sparkRJobBuilder_.getMessage(); + } + return com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + public Builder setSparkRJob(com.google.cloud.dataproc.v1beta2.SparkRJob value) { + if (sparkRJobBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + jobType_ = value; + onChanged(); + } else { + sparkRJobBuilder_.setMessage(value); + } + jobTypeCase_ = 11; + return this; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + public Builder setSparkRJob( + com.google.cloud.dataproc.v1beta2.SparkRJob.Builder builderForValue) { + if (sparkRJobBuilder_ == null) { + jobType_ = builderForValue.build(); + onChanged(); + } else { + sparkRJobBuilder_.setMessage(builderForValue.build()); + } + jobTypeCase_ = 11; + return this; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + public Builder mergeSparkRJob(com.google.cloud.dataproc.v1beta2.SparkRJob value) { + if (sparkRJobBuilder_ == null) { + if (jobTypeCase_ == 11 + && jobType_ != com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance()) { + jobType_ = + com.google.cloud.dataproc.v1beta2.SparkRJob.newBuilder( + (com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_) + .mergeFrom(value) + .buildPartial(); + } else { + jobType_ = value; + } + onChanged(); + } else { + if (jobTypeCase_ == 11) { + sparkRJobBuilder_.mergeFrom(value); + } + sparkRJobBuilder_.setMessage(value); + } + jobTypeCase_ = 11; + return this; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + public Builder clearSparkRJob() { + if (sparkRJobBuilder_ == null) { + if (jobTypeCase_ == 11) { + jobTypeCase_ = 0; + jobType_ = null; + onChanged(); + } + } else { + if (jobTypeCase_ == 11) { + jobTypeCase_ = 0; + jobType_ = null; + } + sparkRJobBuilder_.clear(); + } + return this; + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + public com.google.cloud.dataproc.v1beta2.SparkRJob.Builder getSparkRJobBuilder() { + return getSparkRJobFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + public com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilder() { + if ((jobTypeCase_ == 11) && (sparkRJobBuilder_ != null)) { + return sparkRJobBuilder_.getMessageOrBuilder(); + } else { + if (jobTypeCase_ == 11) { + return (com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_; + } + return com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Spark R job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.SparkRJob, + com.google.cloud.dataproc.v1beta2.SparkRJob.Builder, + com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder> + getSparkRJobFieldBuilder() { + if (sparkRJobBuilder_ == null) { + if (!(jobTypeCase_ == 11)) { + jobType_ = com.google.cloud.dataproc.v1beta2.SparkRJob.getDefaultInstance(); + } + sparkRJobBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.SparkRJob, + com.google.cloud.dataproc.v1beta2.SparkRJob.Builder, + com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder>( + (com.google.cloud.dataproc.v1beta2.SparkRJob) jobType_, + getParentForChildren(), + isClean()); + jobType_ = null; + } + jobTypeCase_ = 11; + onChanged(); + ; + return sparkRJobBuilder_; + } + private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.SparkSqlJob, com.google.cloud.dataproc.v1beta2.SparkSqlJob.Builder, @@ -2366,6 +2770,212 @@ public com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBu return sparkSqlJobBuilder_; } + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.PrestoJob, + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder, + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder> + prestoJobBuilder_; + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + * + * @return Whether the prestoJob field is set. + */ + public boolean hasPrestoJob() { + return jobTypeCase_ == 12; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + * + * @return The prestoJob. + */ + public com.google.cloud.dataproc.v1beta2.PrestoJob getPrestoJob() { + if (prestoJobBuilder_ == null) { + if (jobTypeCase_ == 12) { + return (com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_; + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } else { + if (jobTypeCase_ == 12) { + return prestoJobBuilder_.getMessage(); + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + public Builder setPrestoJob(com.google.cloud.dataproc.v1beta2.PrestoJob value) { + if (prestoJobBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + jobType_ = value; + onChanged(); + } else { + prestoJobBuilder_.setMessage(value); + } + jobTypeCase_ = 12; + return this; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + public Builder setPrestoJob( + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder builderForValue) { + if (prestoJobBuilder_ == null) { + jobType_ = builderForValue.build(); + onChanged(); + } else { + prestoJobBuilder_.setMessage(builderForValue.build()); + } + jobTypeCase_ = 12; + return this; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + public Builder mergePrestoJob(com.google.cloud.dataproc.v1beta2.PrestoJob value) { + if (prestoJobBuilder_ == null) { + if (jobTypeCase_ == 12 + && jobType_ != com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance()) { + jobType_ = + com.google.cloud.dataproc.v1beta2.PrestoJob.newBuilder( + (com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_) + .mergeFrom(value) + .buildPartial(); + } else { + jobType_ = value; + } + onChanged(); + } else { + if (jobTypeCase_ == 12) { + prestoJobBuilder_.mergeFrom(value); + } + prestoJobBuilder_.setMessage(value); + } + jobTypeCase_ = 12; + return this; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + public Builder clearPrestoJob() { + if (prestoJobBuilder_ == null) { + if (jobTypeCase_ == 12) { + jobTypeCase_ = 0; + jobType_ = null; + onChanged(); + } + } else { + if (jobTypeCase_ == 12) { + jobTypeCase_ = 0; + jobType_ = null; + } + prestoJobBuilder_.clear(); + } + return this; + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + public com.google.cloud.dataproc.v1beta2.PrestoJob.Builder getPrestoJobBuilder() { + return getPrestoJobFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + public com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder getPrestoJobOrBuilder() { + if ((jobTypeCase_ == 12) && (prestoJobBuilder_ != null)) { + return prestoJobBuilder_.getMessageOrBuilder(); + } else { + if (jobTypeCase_ == 12) { + return (com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_; + } + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + } + /** + * + * + *
+     * Presto job
+     * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.PrestoJob, + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder, + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder> + getPrestoJobFieldBuilder() { + if (prestoJobBuilder_ == null) { + if (!(jobTypeCase_ == 12)) { + jobType_ = com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + prestoJobBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.PrestoJob, + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder, + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder>( + (com.google.cloud.dataproc.v1beta2.PrestoJob) jobType_, + getParentForChildren(), + isClean()); + jobType_ = null; + } + jobTypeCase_ = 12; + onChanged(); + ; + return prestoJobBuilder_; + } + private com.google.protobuf.MapField labels_; private com.google.protobuf.MapField internalGetLabels() { @@ -2404,7 +3014,7 @@ public int getLabelsCount() { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public boolean containsLabels(java.lang.String key) { if (key == null) { @@ -2430,7 +3040,7 @@ public java.util.Map getLabels() { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public java.util.Map getLabelsMap() { return internalGetLabels().getMap(); @@ -2448,7 +3058,7 @@ public java.util.Map getLabelsMap() { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public java.lang.String getLabelsOrDefault( java.lang.String key, java.lang.String defaultValue) { @@ -2471,7 +3081,7 @@ public java.lang.String getLabelsOrDefault( * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public java.lang.String getLabelsOrThrow(java.lang.String key) { if (key == null) { @@ -2501,7 +3111,7 @@ public Builder clearLabels() { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public Builder removeLabels(java.lang.String key) { if (key == null) { @@ -2528,7 +3138,7 @@ public java.util.Map getMutableLabels() { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public Builder putLabels(java.lang.String key, java.lang.String value) { if (key == null) { @@ -2553,7 +3163,7 @@ public Builder putLabels(java.lang.String key, java.lang.String value) { * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ public Builder putAllLabels(java.util.Map values) { internalGetMutableLabels().getMutableMap().putAll(values); @@ -2573,7 +3183,9 @@ public Builder putAllLabels(java.util.Map va * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the scheduling field is set. */ @@ -2587,7 +3199,9 @@ public boolean hasScheduling() { * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The scheduling. */ @@ -2607,7 +3221,9 @@ public com.google.cloud.dataproc.v1beta2.JobScheduling getScheduling() { * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ public Builder setScheduling(com.google.cloud.dataproc.v1beta2.JobScheduling value) { if (schedulingBuilder_ == null) { @@ -2629,7 +3245,9 @@ public Builder setScheduling(com.google.cloud.dataproc.v1beta2.JobScheduling val * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ public Builder setScheduling( com.google.cloud.dataproc.v1beta2.JobScheduling.Builder builderForValue) { @@ -2649,7 +3267,9 @@ public Builder setScheduling( * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ public Builder mergeScheduling(com.google.cloud.dataproc.v1beta2.JobScheduling value) { if (schedulingBuilder_ == null) { @@ -2675,7 +3295,9 @@ public Builder mergeScheduling(com.google.cloud.dataproc.v1beta2.JobScheduling v * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ public Builder clearScheduling() { if (schedulingBuilder_ == null) { @@ -2695,7 +3317,9 @@ public Builder clearScheduling() { * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ public com.google.cloud.dataproc.v1beta2.JobScheduling.Builder getSchedulingBuilder() { @@ -2709,7 +3333,9 @@ public com.google.cloud.dataproc.v1beta2.JobScheduling.Builder getSchedulingBuil * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ public com.google.cloud.dataproc.v1beta2.JobSchedulingOrBuilder getSchedulingOrBuilder() { if (schedulingBuilder_ != null) { @@ -2727,7 +3353,9 @@ public com.google.cloud.dataproc.v1beta2.JobSchedulingOrBuilder getSchedulingOrB * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ private com.google.protobuf.SingleFieldBuilderV3< com.google.cloud.dataproc.v1beta2.JobScheduling, @@ -2763,7 +3391,8 @@ private void ensurePrerequisiteStepIdsIsMutable() { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return A list containing the prerequisiteStepIds. */ @@ -2778,7 +3407,8 @@ public com.google.protobuf.ProtocolStringList getPrerequisiteStepIdsList() { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The count of prerequisiteStepIds. */ @@ -2793,7 +3423,8 @@ public int getPrerequisiteStepIdsCount() { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param index The index of the element to return. * @return The prerequisiteStepIds at the given index. @@ -2809,7 +3440,8 @@ public java.lang.String getPrerequisiteStepIds(int index) { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param index The index of the value to return. * @return The bytes of the prerequisiteStepIds at the given index. @@ -2825,7 +3457,8 @@ public com.google.protobuf.ByteString getPrerequisiteStepIdsBytes(int index) { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param index The index to set the value at. * @param value The prerequisiteStepIds to set. @@ -2848,7 +3481,8 @@ public Builder setPrerequisiteStepIds(int index, java.lang.String value) { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param value The prerequisiteStepIds to add. * @return This builder for chaining. @@ -2870,7 +3504,8 @@ public Builder addPrerequisiteStepIds(java.lang.String value) { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param values The prerequisiteStepIds to add. * @return This builder for chaining. @@ -2889,7 +3524,8 @@ public Builder addAllPrerequisiteStepIds(java.lang.Iterable va * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return This builder for chaining. */ @@ -2907,7 +3543,8 @@ public Builder clearPrerequisiteStepIds() { * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param value The bytes of the prerequisiteStepIds to add. * @return This builder for chaining. diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJobOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJobOrBuilder.java index 6c1267fe..053134b8 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJobOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/OrderedJobOrBuilder.java @@ -31,14 +31,14 @@ public interface OrderedJobOrBuilder * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The stepId. */ @@ -51,14 +51,14 @@ public interface OrderedJobOrBuilder * within the template. * The step id is used as prefix for job id, as job * `goog-dataproc-workflow-step-id` label, and in - * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - * field from other steps. + * [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + * steps. * The id must contain only letters (a-z, A-Z), numbers (0-9), * underscores (_), and hyphens (-). Cannot begin or end with underscore * or hyphen. Must consist of between 3 and 50 characters. * * - * string step_id = 1; + * string step_id = 1 [(.google.api.field_behavior) = REQUIRED]; * * @return The bytes for stepId. */ @@ -139,6 +139,41 @@ public interface OrderedJobOrBuilder /** .google.cloud.dataproc.v1beta2.PigJob pig_job = 6; */ com.google.cloud.dataproc.v1beta2.PigJobOrBuilder getPigJobOrBuilder(); + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + * + * @return Whether the sparkRJob field is set. + */ + boolean hasSparkRJob(); + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + * + * @return The sparkRJob. + */ + com.google.cloud.dataproc.v1beta2.SparkRJob getSparkRJob(); + /** + * + * + *
+   * Spark R job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.SparkRJob spark_r_job = 11; + */ + com.google.cloud.dataproc.v1beta2.SparkRJobOrBuilder getSparkRJobOrBuilder(); + /** * .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 7; * @@ -154,6 +189,41 @@ public interface OrderedJobOrBuilder /** .google.cloud.dataproc.v1beta2.SparkSqlJob spark_sql_job = 7; */ com.google.cloud.dataproc.v1beta2.SparkSqlJobOrBuilder getSparkSqlJobOrBuilder(); + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + * + * @return Whether the prestoJob field is set. + */ + boolean hasPrestoJob(); + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + * + * @return The prestoJob. + */ + com.google.cloud.dataproc.v1beta2.PrestoJob getPrestoJob(); + /** + * + * + *
+   * Presto job
+   * 
+ * + * .google.cloud.dataproc.v1beta2.PrestoJob presto_job = 12; + */ + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder getPrestoJobOrBuilder(); + /** * * @@ -167,7 +237,7 @@ public interface OrderedJobOrBuilder * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ int getLabelsCount(); /** @@ -183,7 +253,7 @@ public interface OrderedJobOrBuilder * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ boolean containsLabels(java.lang.String key); /** Use {@link #getLabelsMap()} instead. */ @@ -202,7 +272,7 @@ public interface OrderedJobOrBuilder * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ java.util.Map getLabelsMap(); /** @@ -218,7 +288,7 @@ public interface OrderedJobOrBuilder * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ java.lang.String getLabelsOrDefault(java.lang.String key, java.lang.String defaultValue); /** @@ -234,7 +304,7 @@ public interface OrderedJobOrBuilder * No more than 32 labels can be associated with a given job. * * - * map<string, string> labels = 8; + * map<string, string> labels = 8 [(.google.api.field_behavior) = OPTIONAL]; */ java.lang.String getLabelsOrThrow(java.lang.String key); @@ -245,7 +315,9 @@ public interface OrderedJobOrBuilder * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return Whether the scheduling field is set. */ @@ -257,7 +329,9 @@ public interface OrderedJobOrBuilder * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The scheduling. */ @@ -269,7 +343,9 @@ public interface OrderedJobOrBuilder * Optional. Job scheduling configuration. * * - * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9; + * + * .google.cloud.dataproc.v1beta2.JobScheduling scheduling = 9 [(.google.api.field_behavior) = OPTIONAL]; + * */ com.google.cloud.dataproc.v1beta2.JobSchedulingOrBuilder getSchedulingOrBuilder(); @@ -281,7 +357,8 @@ public interface OrderedJobOrBuilder * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return A list containing the prerequisiteStepIds. */ @@ -294,7 +371,8 @@ public interface OrderedJobOrBuilder * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @return The count of prerequisiteStepIds. */ @@ -307,7 +385,8 @@ public interface OrderedJobOrBuilder * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param index The index of the element to return. * @return The prerequisiteStepIds at the given index. @@ -321,7 +400,8 @@ public interface OrderedJobOrBuilder * If not specified, the job will start at the beginning of workflow. * * - * repeated string prerequisite_step_ids = 10; + * repeated string prerequisite_step_ids = 10 [(.google.api.field_behavior) = OPTIONAL]; + * * * @param index The index of the value to return. * @return The bytes of the prerequisiteStepIds at the given index. diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PrestoJob.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PrestoJob.java new file mode 100644 index 00000000..af518e68 --- /dev/null +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PrestoJob.java @@ -0,0 +1,2248 @@ +/* + * Copyright 2020 Google LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dataproc/v1beta2/jobs.proto + +package com.google.cloud.dataproc.v1beta2; + +/** + * + * + *
+ * A Dataproc job for running [Presto](https://prestosql.io/) queries.
+ * **IMPORTANT**: The [Dataproc Presto Optional
+ * Component](/dataproc/docs/concepts/components/presto) must be enabled when
+ * the cluster is created to submit a Presto job to the cluster.
+ * 
+ * + * Protobuf type {@code google.cloud.dataproc.v1beta2.PrestoJob} + */ +public final class PrestoJob extends com.google.protobuf.GeneratedMessageV3 + implements + // @@protoc_insertion_point(message_implements:google.cloud.dataproc.v1beta2.PrestoJob) + PrestoJobOrBuilder { + private static final long serialVersionUID = 0L; + // Use PrestoJob.newBuilder() to construct. + private PrestoJob(com.google.protobuf.GeneratedMessageV3.Builder builder) { + super(builder); + } + + private PrestoJob() { + outputFormat_ = ""; + clientTags_ = com.google.protobuf.LazyStringArrayList.EMPTY; + } + + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance(UnusedPrivateParameter unused) { + return new PrestoJob(); + } + + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet getUnknownFields() { + return this.unknownFields; + } + + private PrestoJob( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } + int mutable_bitField0_ = 0; + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + case 10: + { + java.lang.String s = input.readStringRequireUtf8(); + queriesCase_ = 1; + queries_ = s; + break; + } + case 18: + { + com.google.cloud.dataproc.v1beta2.QueryList.Builder subBuilder = null; + if (queriesCase_ == 2) { + subBuilder = ((com.google.cloud.dataproc.v1beta2.QueryList) queries_).toBuilder(); + } + queries_ = + input.readMessage( + com.google.cloud.dataproc.v1beta2.QueryList.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom((com.google.cloud.dataproc.v1beta2.QueryList) queries_); + queries_ = subBuilder.buildPartial(); + } + queriesCase_ = 2; + break; + } + case 24: + { + continueOnFailure_ = input.readBool(); + break; + } + case 34: + { + java.lang.String s = input.readStringRequireUtf8(); + + outputFormat_ = s; + break; + } + case 42: + { + java.lang.String s = input.readStringRequireUtf8(); + if (!((mutable_bitField0_ & 0x00000001) != 0)) { + clientTags_ = new com.google.protobuf.LazyStringArrayList(); + mutable_bitField0_ |= 0x00000001; + } + clientTags_.add(s); + break; + } + case 50: + { + if (!((mutable_bitField0_ & 0x00000002) != 0)) { + properties_ = + com.google.protobuf.MapField.newMapField( + PropertiesDefaultEntryHolder.defaultEntry); + mutable_bitField0_ |= 0x00000002; + } + com.google.protobuf.MapEntry properties__ = + input.readMessage( + PropertiesDefaultEntryHolder.defaultEntry.getParserForType(), + extensionRegistry); + properties_.getMutableMap().put(properties__.getKey(), properties__.getValue()); + break; + } + case 58: + { + com.google.cloud.dataproc.v1beta2.LoggingConfig.Builder subBuilder = null; + if (loggingConfig_ != null) { + subBuilder = loggingConfig_.toBuilder(); + } + loggingConfig_ = + input.readMessage( + com.google.cloud.dataproc.v1beta2.LoggingConfig.parser(), extensionRegistry); + if (subBuilder != null) { + subBuilder.mergeFrom(loggingConfig_); + loggingConfig_ = subBuilder.buildPartial(); + } + + break; + } + default: + { + if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException(e).setUnfinishedMessage(this); + } finally { + if (((mutable_bitField0_ & 0x00000001) != 0)) { + clientTags_ = clientTags_.getUnmodifiableView(); + } + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_PrestoJob_descriptor; + } + + @SuppressWarnings({"rawtypes"}) + @java.lang.Override + protected com.google.protobuf.MapField internalGetMapField(int number) { + switch (number) { + case 6: + return internalGetProperties(); + default: + throw new RuntimeException("Invalid map field number: " + number); + } + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_PrestoJob_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1beta2.PrestoJob.class, + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder.class); + } + + private int queriesCase_ = 0; + private java.lang.Object queries_; + + public enum QueriesCase + implements + com.google.protobuf.Internal.EnumLite, + com.google.protobuf.AbstractMessage.InternalOneOfEnum { + QUERY_FILE_URI(1), + QUERY_LIST(2), + QUERIES_NOT_SET(0); + private final int value; + + private QueriesCase(int value) { + this.value = value; + } + /** + * @param value The number of the enum to look for. + * @return The enum associated with the given number. + * @deprecated Use {@link #forNumber(int)} instead. + */ + @java.lang.Deprecated + public static QueriesCase valueOf(int value) { + return forNumber(value); + } + + public static QueriesCase forNumber(int value) { + switch (value) { + case 1: + return QUERY_FILE_URI; + case 2: + return QUERY_LIST; + case 0: + return QUERIES_NOT_SET; + default: + return null; + } + } + + public int getNumber() { + return this.value; + } + }; + + public QueriesCase getQueriesCase() { + return QueriesCase.forNumber(queriesCase_); + } + + public static final int QUERY_FILE_URI_FIELD_NUMBER = 1; + /** + * + * + *
+   * The HCFS URI of the script that contains SQL queries.
+   * 
+ * + * string query_file_uri = 1; + * + * @return The queryFileUri. + */ + public java.lang.String getQueryFileUri() { + java.lang.Object ref = ""; + if (queriesCase_ == 1) { + ref = queries_; + } + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (queriesCase_ == 1) { + queries_ = s; + } + return s; + } + } + /** + * + * + *
+   * The HCFS URI of the script that contains SQL queries.
+   * 
+ * + * string query_file_uri = 1; + * + * @return The bytes for queryFileUri. + */ + public com.google.protobuf.ByteString getQueryFileUriBytes() { + java.lang.Object ref = ""; + if (queriesCase_ == 1) { + ref = queries_; + } + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + if (queriesCase_ == 1) { + queries_ = b; + } + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + public static final int QUERY_LIST_FIELD_NUMBER = 2; + /** + * + * + *
+   * A list of queries.
+   * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + * + * @return Whether the queryList field is set. + */ + public boolean hasQueryList() { + return queriesCase_ == 2; + } + /** + * + * + *
+   * A list of queries.
+   * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + * + * @return The queryList. + */ + public com.google.cloud.dataproc.v1beta2.QueryList getQueryList() { + if (queriesCase_ == 2) { + return (com.google.cloud.dataproc.v1beta2.QueryList) queries_; + } + return com.google.cloud.dataproc.v1beta2.QueryList.getDefaultInstance(); + } + /** + * + * + *
+   * A list of queries.
+   * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + public com.google.cloud.dataproc.v1beta2.QueryListOrBuilder getQueryListOrBuilder() { + if (queriesCase_ == 2) { + return (com.google.cloud.dataproc.v1beta2.QueryList) queries_; + } + return com.google.cloud.dataproc.v1beta2.QueryList.getDefaultInstance(); + } + + public static final int CONTINUE_ON_FAILURE_FIELD_NUMBER = 3; + private boolean continueOnFailure_; + /** + * + * + *
+   * Optional. Whether to continue executing queries if a query fails.
+   * The default value is `false`. Setting to `true` can be useful when
+   * executing independent parallel queries.
+   * 
+ * + * bool continue_on_failure = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The continueOnFailure. + */ + public boolean getContinueOnFailure() { + return continueOnFailure_; + } + + public static final int OUTPUT_FORMAT_FIELD_NUMBER = 4; + private volatile java.lang.Object outputFormat_; + /** + * + * + *
+   * Optional. The format in which query output will be displayed. See the
+   * Presto documentation for supported output formats
+   * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The outputFormat. + */ + public java.lang.String getOutputFormat() { + java.lang.Object ref = outputFormat_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + outputFormat_ = s; + return s; + } + } + /** + * + * + *
+   * Optional. The format in which query output will be displayed. See the
+   * Presto documentation for supported output formats
+   * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The bytes for outputFormat. + */ + public com.google.protobuf.ByteString getOutputFormatBytes() { + java.lang.Object ref = outputFormat_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + outputFormat_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + public static final int CLIENT_TAGS_FIELD_NUMBER = 5; + private com.google.protobuf.LazyStringList clientTags_; + /** + * + * + *
+   * Optional. Presto client tags to attach to this query
+   * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return A list containing the clientTags. + */ + public com.google.protobuf.ProtocolStringList getClientTagsList() { + return clientTags_; + } + /** + * + * + *
+   * Optional. Presto client tags to attach to this query
+   * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The count of clientTags. + */ + public int getClientTagsCount() { + return clientTags_.size(); + } + /** + * + * + *
+   * Optional. Presto client tags to attach to this query
+   * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param index The index of the element to return. + * @return The clientTags at the given index. + */ + public java.lang.String getClientTags(int index) { + return clientTags_.get(index); + } + /** + * + * + *
+   * Optional. Presto client tags to attach to this query
+   * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param index The index of the value to return. + * @return The bytes of the clientTags at the given index. + */ + public com.google.protobuf.ByteString getClientTagsBytes(int index) { + return clientTags_.getByteString(index); + } + + public static final int PROPERTIES_FIELD_NUMBER = 6; + + private static final class PropertiesDefaultEntryHolder { + static final com.google.protobuf.MapEntry defaultEntry = + com.google.protobuf.MapEntry.newDefaultInstance( + com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_PrestoJob_PropertiesEntry_descriptor, + com.google.protobuf.WireFormat.FieldType.STRING, + "", + com.google.protobuf.WireFormat.FieldType.STRING, + ""); + } + + private com.google.protobuf.MapField properties_; + + private com.google.protobuf.MapField internalGetProperties() { + if (properties_ == null) { + return com.google.protobuf.MapField.emptyMapField(PropertiesDefaultEntryHolder.defaultEntry); + } + return properties_; + } + + public int getPropertiesCount() { + return internalGetProperties().getMap().size(); + } + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public boolean containsProperties(java.lang.String key) { + if (key == null) { + throw new java.lang.NullPointerException(); + } + return internalGetProperties().getMap().containsKey(key); + } + /** Use {@link #getPropertiesMap()} instead. */ + @java.lang.Deprecated + public java.util.Map getProperties() { + return getPropertiesMap(); + } + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public java.util.Map getPropertiesMap() { + return internalGetProperties().getMap(); + } + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public java.lang.String getPropertiesOrDefault( + java.lang.String key, java.lang.String defaultValue) { + if (key == null) { + throw new java.lang.NullPointerException(); + } + java.util.Map map = internalGetProperties().getMap(); + return map.containsKey(key) ? map.get(key) : defaultValue; + } + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public java.lang.String getPropertiesOrThrow(java.lang.String key) { + if (key == null) { + throw new java.lang.NullPointerException(); + } + java.util.Map map = internalGetProperties().getMap(); + if (!map.containsKey(key)) { + throw new java.lang.IllegalArgumentException(); + } + return map.get(key); + } + + public static final int LOGGING_CONFIG_FIELD_NUMBER = 7; + private com.google.cloud.dataproc.v1beta2.LoggingConfig loggingConfig_; + /** + * + * + *
+   * Optional. The runtime log config for job execution.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the loggingConfig field is set. + */ + public boolean hasLoggingConfig() { + return loggingConfig_ != null; + } + /** + * + * + *
+   * Optional. The runtime log config for job execution.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The loggingConfig. + */ + public com.google.cloud.dataproc.v1beta2.LoggingConfig getLoggingConfig() { + return loggingConfig_ == null + ? com.google.cloud.dataproc.v1beta2.LoggingConfig.getDefaultInstance() + : loggingConfig_; + } + /** + * + * + *
+   * Optional. The runtime log config for job execution.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.LoggingConfigOrBuilder getLoggingConfigOrBuilder() { + return getLoggingConfig(); + } + + private byte memoizedIsInitialized = -1; + + @java.lang.Override + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; + + memoizedIsInitialized = 1; + return true; + } + + @java.lang.Override + public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { + if (queriesCase_ == 1) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, queries_); + } + if (queriesCase_ == 2) { + output.writeMessage(2, (com.google.cloud.dataproc.v1beta2.QueryList) queries_); + } + if (continueOnFailure_ != false) { + output.writeBool(3, continueOnFailure_); + } + if (!getOutputFormatBytes().isEmpty()) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 4, outputFormat_); + } + for (int i = 0; i < clientTags_.size(); i++) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 5, clientTags_.getRaw(i)); + } + com.google.protobuf.GeneratedMessageV3.serializeStringMapTo( + output, internalGetProperties(), PropertiesDefaultEntryHolder.defaultEntry, 6); + if (loggingConfig_ != null) { + output.writeMessage(7, getLoggingConfig()); + } + unknownFields.writeTo(output); + } + + @java.lang.Override + public int getSerializedSize() { + int size = memoizedSize; + if (size != -1) return size; + + size = 0; + if (queriesCase_ == 1) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, queries_); + } + if (queriesCase_ == 2) { + size += + com.google.protobuf.CodedOutputStream.computeMessageSize( + 2, (com.google.cloud.dataproc.v1beta2.QueryList) queries_); + } + if (continueOnFailure_ != false) { + size += com.google.protobuf.CodedOutputStream.computeBoolSize(3, continueOnFailure_); + } + if (!getOutputFormatBytes().isEmpty()) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, outputFormat_); + } + { + int dataSize = 0; + for (int i = 0; i < clientTags_.size(); i++) { + dataSize += computeStringSizeNoTag(clientTags_.getRaw(i)); + } + size += dataSize; + size += 1 * getClientTagsList().size(); + } + for (java.util.Map.Entry entry : + internalGetProperties().getMap().entrySet()) { + com.google.protobuf.MapEntry properties__ = + PropertiesDefaultEntryHolder.defaultEntry + .newBuilderForType() + .setKey(entry.getKey()) + .setValue(entry.getValue()) + .build(); + size += com.google.protobuf.CodedOutputStream.computeMessageSize(6, properties__); + } + if (loggingConfig_ != null) { + size += com.google.protobuf.CodedOutputStream.computeMessageSize(7, getLoggingConfig()); + } + size += unknownFields.getSerializedSize(); + memoizedSize = size; + return size; + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof com.google.cloud.dataproc.v1beta2.PrestoJob)) { + return super.equals(obj); + } + com.google.cloud.dataproc.v1beta2.PrestoJob other = + (com.google.cloud.dataproc.v1beta2.PrestoJob) obj; + + if (getContinueOnFailure() != other.getContinueOnFailure()) return false; + if (!getOutputFormat().equals(other.getOutputFormat())) return false; + if (!getClientTagsList().equals(other.getClientTagsList())) return false; + if (!internalGetProperties().equals(other.internalGetProperties())) return false; + if (hasLoggingConfig() != other.hasLoggingConfig()) return false; + if (hasLoggingConfig()) { + if (!getLoggingConfig().equals(other.getLoggingConfig())) return false; + } + if (!getQueriesCase().equals(other.getQueriesCase())) return false; + switch (queriesCase_) { + case 1: + if (!getQueryFileUri().equals(other.getQueryFileUri())) return false; + break; + case 2: + if (!getQueryList().equals(other.getQueryList())) return false; + break; + case 0: + default: + } + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (37 * hash) + CONTINUE_ON_FAILURE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean(getContinueOnFailure()); + hash = (37 * hash) + OUTPUT_FORMAT_FIELD_NUMBER; + hash = (53 * hash) + getOutputFormat().hashCode(); + if (getClientTagsCount() > 0) { + hash = (37 * hash) + CLIENT_TAGS_FIELD_NUMBER; + hash = (53 * hash) + getClientTagsList().hashCode(); + } + if (!internalGetProperties().getMap().isEmpty()) { + hash = (37 * hash) + PROPERTIES_FIELD_NUMBER; + hash = (53 * hash) + internalGetProperties().hashCode(); + } + if (hasLoggingConfig()) { + hash = (37 * hash) + LOGGING_CONFIG_FIELD_NUMBER; + hash = (53 * hash) + getLoggingConfig().hashCode(); + } + switch (queriesCase_) { + case 1: + hash = (37 * hash) + QUERY_FILE_URI_FIELD_NUMBER; + hash = (53 * hash) + getQueryFileUri().hashCode(); + break; + case 2: + hash = (37 * hash) + QUERY_LIST_FIELD_NUMBER; + hash = (53 * hash) + getQueryList().hashCode(); + break; + case 0: + default: + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom(java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom( + java.nio.ByteBuffer data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom( + byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom(java.io.InputStream input) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseDelimitedFrom( + java.io.InputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseDelimitedFrom( + java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseDelimitedWithIOException( + PARSER, input, extensionRegistry); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom( + com.google.protobuf.CodedInputStream input) throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException(PARSER, input); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageV3.parseWithIOException( + PARSER, input, extensionRegistry); + } + + @java.lang.Override + public Builder newBuilderForType() { + return newBuilder(); + } + + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } + + public static Builder newBuilder(com.google.cloud.dataproc.v1beta2.PrestoJob prototype) { + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); + } + + @java.lang.Override + protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * + * + *
+   * A Dataproc job for running [Presto](https://prestosql.io/) queries.
+   * **IMPORTANT**: The [Dataproc Presto Optional
+   * Component](/dataproc/docs/concepts/components/presto) must be enabled when
+   * the cluster is created to submit a Presto job to the cluster.
+   * 
+ * + * Protobuf type {@code google.cloud.dataproc.v1beta2.PrestoJob} + */ + public static final class Builder extends com.google.protobuf.GeneratedMessageV3.Builder + implements + // @@protoc_insertion_point(builder_implements:google.cloud.dataproc.v1beta2.PrestoJob) + com.google.cloud.dataproc.v1beta2.PrestoJobOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_PrestoJob_descriptor; + } + + @SuppressWarnings({"rawtypes"}) + protected com.google.protobuf.MapField internalGetMapField(int number) { + switch (number) { + case 6: + return internalGetProperties(); + default: + throw new RuntimeException("Invalid map field number: " + number); + } + } + + @SuppressWarnings({"rawtypes"}) + protected com.google.protobuf.MapField internalGetMutableMapField(int number) { + switch (number) { + case 6: + return internalGetMutableProperties(); + default: + throw new RuntimeException("Invalid map field number: " + number); + } + } + + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internalGetFieldAccessorTable() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_PrestoJob_fieldAccessorTable + .ensureFieldAccessorsInitialized( + com.google.cloud.dataproc.v1beta2.PrestoJob.class, + com.google.cloud.dataproc.v1beta2.PrestoJob.Builder.class); + } + + // Construct using com.google.cloud.dataproc.v1beta2.PrestoJob.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder(com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders) {} + } + + @java.lang.Override + public Builder clear() { + super.clear(); + continueOnFailure_ = false; + + outputFormat_ = ""; + + clientTags_ = com.google.protobuf.LazyStringArrayList.EMPTY; + bitField0_ = (bitField0_ & ~0x00000001); + internalGetMutableProperties().clear(); + if (loggingConfigBuilder_ == null) { + loggingConfig_ = null; + } else { + loggingConfig_ = null; + loggingConfigBuilder_ = null; + } + queriesCase_ = 0; + queries_ = null; + return this; + } + + @java.lang.Override + public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { + return com.google.cloud.dataproc.v1beta2.JobsProto + .internal_static_google_cloud_dataproc_v1beta2_PrestoJob_descriptor; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.PrestoJob getDefaultInstanceForType() { + return com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance(); + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.PrestoJob build() { + com.google.cloud.dataproc.v1beta2.PrestoJob result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.PrestoJob buildPartial() { + com.google.cloud.dataproc.v1beta2.PrestoJob result = + new com.google.cloud.dataproc.v1beta2.PrestoJob(this); + int from_bitField0_ = bitField0_; + if (queriesCase_ == 1) { + result.queries_ = queries_; + } + if (queriesCase_ == 2) { + if (queryListBuilder_ == null) { + result.queries_ = queries_; + } else { + result.queries_ = queryListBuilder_.build(); + } + } + result.continueOnFailure_ = continueOnFailure_; + result.outputFormat_ = outputFormat_; + if (((bitField0_ & 0x00000001) != 0)) { + clientTags_ = clientTags_.getUnmodifiableView(); + bitField0_ = (bitField0_ & ~0x00000001); + } + result.clientTags_ = clientTags_; + result.properties_ = internalGetProperties(); + result.properties_.makeImmutable(); + if (loggingConfigBuilder_ == null) { + result.loggingConfig_ = loggingConfig_; + } else { + result.loggingConfig_ = loggingConfigBuilder_.build(); + } + result.queriesCase_ = queriesCase_; + onBuilt(); + return result; + } + + @java.lang.Override + public Builder clone() { + return super.clone(); + } + + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.setField(field, value); + } + + @java.lang.Override + public Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + + @java.lang.Override + public Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { + return super.addRepeatedField(field, value); + } + + @java.lang.Override + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof com.google.cloud.dataproc.v1beta2.PrestoJob) { + return mergeFrom((com.google.cloud.dataproc.v1beta2.PrestoJob) other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(com.google.cloud.dataproc.v1beta2.PrestoJob other) { + if (other == com.google.cloud.dataproc.v1beta2.PrestoJob.getDefaultInstance()) return this; + if (other.getContinueOnFailure() != false) { + setContinueOnFailure(other.getContinueOnFailure()); + } + if (!other.getOutputFormat().isEmpty()) { + outputFormat_ = other.outputFormat_; + onChanged(); + } + if (!other.clientTags_.isEmpty()) { + if (clientTags_.isEmpty()) { + clientTags_ = other.clientTags_; + bitField0_ = (bitField0_ & ~0x00000001); + } else { + ensureClientTagsIsMutable(); + clientTags_.addAll(other.clientTags_); + } + onChanged(); + } + internalGetMutableProperties().mergeFrom(other.internalGetProperties()); + if (other.hasLoggingConfig()) { + mergeLoggingConfig(other.getLoggingConfig()); + } + switch (other.getQueriesCase()) { + case QUERY_FILE_URI: + { + queriesCase_ = 1; + queries_ = other.queries_; + onChanged(); + break; + } + case QUERY_LIST: + { + mergeQueryList(other.getQueryList()); + break; + } + case QUERIES_NOT_SET: + { + break; + } + } + this.mergeUnknownFields(other.unknownFields); + onChanged(); + return this; + } + + @java.lang.Override + public final boolean isInitialized() { + return true; + } + + @java.lang.Override + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + com.google.cloud.dataproc.v1beta2.PrestoJob parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (com.google.cloud.dataproc.v1beta2.PrestoJob) e.getUnfinishedMessage(); + throw e.unwrapIOException(); + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + + private int queriesCase_ = 0; + private java.lang.Object queries_; + + public QueriesCase getQueriesCase() { + return QueriesCase.forNumber(queriesCase_); + } + + public Builder clearQueries() { + queriesCase_ = 0; + queries_ = null; + onChanged(); + return this; + } + + private int bitField0_; + + /** + * + * + *
+     * The HCFS URI of the script that contains SQL queries.
+     * 
+ * + * string query_file_uri = 1; + * + * @return The queryFileUri. + */ + public java.lang.String getQueryFileUri() { + java.lang.Object ref = ""; + if (queriesCase_ == 1) { + ref = queries_; + } + if (!(ref instanceof java.lang.String)) { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (queriesCase_ == 1) { + queries_ = s; + } + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * + * + *
+     * The HCFS URI of the script that contains SQL queries.
+     * 
+ * + * string query_file_uri = 1; + * + * @return The bytes for queryFileUri. + */ + public com.google.protobuf.ByteString getQueryFileUriBytes() { + java.lang.Object ref = ""; + if (queriesCase_ == 1) { + ref = queries_; + } + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + if (queriesCase_ == 1) { + queries_ = b; + } + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * + * + *
+     * The HCFS URI of the script that contains SQL queries.
+     * 
+ * + * string query_file_uri = 1; + * + * @param value The queryFileUri to set. + * @return This builder for chaining. + */ + public Builder setQueryFileUri(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + queriesCase_ = 1; + queries_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * The HCFS URI of the script that contains SQL queries.
+     * 
+ * + * string query_file_uri = 1; + * + * @return This builder for chaining. + */ + public Builder clearQueryFileUri() { + if (queriesCase_ == 1) { + queriesCase_ = 0; + queries_ = null; + onChanged(); + } + return this; + } + /** + * + * + *
+     * The HCFS URI of the script that contains SQL queries.
+     * 
+ * + * string query_file_uri = 1; + * + * @param value The bytes for queryFileUri to set. + * @return This builder for chaining. + */ + public Builder setQueryFileUriBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + queriesCase_ = 1; + queries_ = value; + onChanged(); + return this; + } + + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.QueryList, + com.google.cloud.dataproc.v1beta2.QueryList.Builder, + com.google.cloud.dataproc.v1beta2.QueryListOrBuilder> + queryListBuilder_; + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + * + * @return Whether the queryList field is set. + */ + public boolean hasQueryList() { + return queriesCase_ == 2; + } + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + * + * @return The queryList. + */ + public com.google.cloud.dataproc.v1beta2.QueryList getQueryList() { + if (queryListBuilder_ == null) { + if (queriesCase_ == 2) { + return (com.google.cloud.dataproc.v1beta2.QueryList) queries_; + } + return com.google.cloud.dataproc.v1beta2.QueryList.getDefaultInstance(); + } else { + if (queriesCase_ == 2) { + return queryListBuilder_.getMessage(); + } + return com.google.cloud.dataproc.v1beta2.QueryList.getDefaultInstance(); + } + } + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + public Builder setQueryList(com.google.cloud.dataproc.v1beta2.QueryList value) { + if (queryListBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + queries_ = value; + onChanged(); + } else { + queryListBuilder_.setMessage(value); + } + queriesCase_ = 2; + return this; + } + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + public Builder setQueryList( + com.google.cloud.dataproc.v1beta2.QueryList.Builder builderForValue) { + if (queryListBuilder_ == null) { + queries_ = builderForValue.build(); + onChanged(); + } else { + queryListBuilder_.setMessage(builderForValue.build()); + } + queriesCase_ = 2; + return this; + } + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + public Builder mergeQueryList(com.google.cloud.dataproc.v1beta2.QueryList value) { + if (queryListBuilder_ == null) { + if (queriesCase_ == 2 + && queries_ != com.google.cloud.dataproc.v1beta2.QueryList.getDefaultInstance()) { + queries_ = + com.google.cloud.dataproc.v1beta2.QueryList.newBuilder( + (com.google.cloud.dataproc.v1beta2.QueryList) queries_) + .mergeFrom(value) + .buildPartial(); + } else { + queries_ = value; + } + onChanged(); + } else { + if (queriesCase_ == 2) { + queryListBuilder_.mergeFrom(value); + } + queryListBuilder_.setMessage(value); + } + queriesCase_ = 2; + return this; + } + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + public Builder clearQueryList() { + if (queryListBuilder_ == null) { + if (queriesCase_ == 2) { + queriesCase_ = 0; + queries_ = null; + onChanged(); + } + } else { + if (queriesCase_ == 2) { + queriesCase_ = 0; + queries_ = null; + } + queryListBuilder_.clear(); + } + return this; + } + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + public com.google.cloud.dataproc.v1beta2.QueryList.Builder getQueryListBuilder() { + return getQueryListFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + public com.google.cloud.dataproc.v1beta2.QueryListOrBuilder getQueryListOrBuilder() { + if ((queriesCase_ == 2) && (queryListBuilder_ != null)) { + return queryListBuilder_.getMessageOrBuilder(); + } else { + if (queriesCase_ == 2) { + return (com.google.cloud.dataproc.v1beta2.QueryList) queries_; + } + return com.google.cloud.dataproc.v1beta2.QueryList.getDefaultInstance(); + } + } + /** + * + * + *
+     * A list of queries.
+     * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.QueryList, + com.google.cloud.dataproc.v1beta2.QueryList.Builder, + com.google.cloud.dataproc.v1beta2.QueryListOrBuilder> + getQueryListFieldBuilder() { + if (queryListBuilder_ == null) { + if (!(queriesCase_ == 2)) { + queries_ = com.google.cloud.dataproc.v1beta2.QueryList.getDefaultInstance(); + } + queryListBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.QueryList, + com.google.cloud.dataproc.v1beta2.QueryList.Builder, + com.google.cloud.dataproc.v1beta2.QueryListOrBuilder>( + (com.google.cloud.dataproc.v1beta2.QueryList) queries_, + getParentForChildren(), + isClean()); + queries_ = null; + } + queriesCase_ = 2; + onChanged(); + ; + return queryListBuilder_; + } + + private boolean continueOnFailure_; + /** + * + * + *
+     * Optional. Whether to continue executing queries if a query fails.
+     * The default value is `false`. Setting to `true` can be useful when
+     * executing independent parallel queries.
+     * 
+ * + * bool continue_on_failure = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The continueOnFailure. + */ + public boolean getContinueOnFailure() { + return continueOnFailure_; + } + /** + * + * + *
+     * Optional. Whether to continue executing queries if a query fails.
+     * The default value is `false`. Setting to `true` can be useful when
+     * executing independent parallel queries.
+     * 
+ * + * bool continue_on_failure = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param value The continueOnFailure to set. + * @return This builder for chaining. + */ + public Builder setContinueOnFailure(boolean value) { + + continueOnFailure_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Whether to continue executing queries if a query fails.
+     * The default value is `false`. Setting to `true` can be useful when
+     * executing independent parallel queries.
+     * 
+ * + * bool continue_on_failure = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return This builder for chaining. + */ + public Builder clearContinueOnFailure() { + + continueOnFailure_ = false; + onChanged(); + return this; + } + + private java.lang.Object outputFormat_ = ""; + /** + * + * + *
+     * Optional. The format in which query output will be displayed. See the
+     * Presto documentation for supported output formats
+     * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The outputFormat. + */ + public java.lang.String getOutputFormat() { + java.lang.Object ref = outputFormat_; + if (!(ref instanceof java.lang.String)) { + com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + outputFormat_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * + * + *
+     * Optional. The format in which query output will be displayed. See the
+     * Presto documentation for supported output formats
+     * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The bytes for outputFormat. + */ + public com.google.protobuf.ByteString getOutputFormatBytes() { + java.lang.Object ref = outputFormat_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8((java.lang.String) ref); + outputFormat_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * + * + *
+     * Optional. The format in which query output will be displayed. See the
+     * Presto documentation for supported output formats
+     * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param value The outputFormat to set. + * @return This builder for chaining. + */ + public Builder setOutputFormat(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + + outputFormat_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. The format in which query output will be displayed. See the
+     * Presto documentation for supported output formats
+     * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return This builder for chaining. + */ + public Builder clearOutputFormat() { + + outputFormat_ = getDefaultInstance().getOutputFormat(); + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. The format in which query output will be displayed. See the
+     * Presto documentation for supported output formats
+     * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param value The bytes for outputFormat to set. + * @return This builder for chaining. + */ + public Builder setOutputFormatBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + + outputFormat_ = value; + onChanged(); + return this; + } + + private com.google.protobuf.LazyStringList clientTags_ = + com.google.protobuf.LazyStringArrayList.EMPTY; + + private void ensureClientTagsIsMutable() { + if (!((bitField0_ & 0x00000001) != 0)) { + clientTags_ = new com.google.protobuf.LazyStringArrayList(clientTags_); + bitField0_ |= 0x00000001; + } + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return A list containing the clientTags. + */ + public com.google.protobuf.ProtocolStringList getClientTagsList() { + return clientTags_.getUnmodifiableView(); + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The count of clientTags. + */ + public int getClientTagsCount() { + return clientTags_.size(); + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param index The index of the element to return. + * @return The clientTags at the given index. + */ + public java.lang.String getClientTags(int index) { + return clientTags_.get(index); + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param index The index of the value to return. + * @return The bytes of the clientTags at the given index. + */ + public com.google.protobuf.ByteString getClientTagsBytes(int index) { + return clientTags_.getByteString(index); + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param index The index to set the value at. + * @param value The clientTags to set. + * @return This builder for chaining. + */ + public Builder setClientTags(int index, java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + ensureClientTagsIsMutable(); + clientTags_.set(index, value); + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param value The clientTags to add. + * @return This builder for chaining. + */ + public Builder addClientTags(java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + ensureClientTagsIsMutable(); + clientTags_.add(value); + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param values The clientTags to add. + * @return This builder for chaining. + */ + public Builder addAllClientTags(java.lang.Iterable values) { + ensureClientTagsIsMutable(); + com.google.protobuf.AbstractMessageLite.Builder.addAll(values, clientTags_); + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return This builder for chaining. + */ + public Builder clearClientTags() { + clientTags_ = com.google.protobuf.LazyStringArrayList.EMPTY; + bitField0_ = (bitField0_ & ~0x00000001); + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Presto client tags to attach to this query
+     * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param value The bytes of the clientTags to add. + * @return This builder for chaining. + */ + public Builder addClientTagsBytes(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + checkByteStringIsUtf8(value); + ensureClientTagsIsMutable(); + clientTags_.add(value); + onChanged(); + return this; + } + + private com.google.protobuf.MapField properties_; + + private com.google.protobuf.MapField + internalGetProperties() { + if (properties_ == null) { + return com.google.protobuf.MapField.emptyMapField( + PropertiesDefaultEntryHolder.defaultEntry); + } + return properties_; + } + + private com.google.protobuf.MapField + internalGetMutableProperties() { + onChanged(); + ; + if (properties_ == null) { + properties_ = + com.google.protobuf.MapField.newMapField(PropertiesDefaultEntryHolder.defaultEntry); + } + if (!properties_.isMutable()) { + properties_ = properties_.copy(); + } + return properties_; + } + + public int getPropertiesCount() { + return internalGetProperties().getMap().size(); + } + /** + * + * + *
+     * Optional. A mapping of property names to values. Used to set Presto
+     * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+     * Equivalent to using the --session flag in the Presto CLI
+     * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public boolean containsProperties(java.lang.String key) { + if (key == null) { + throw new java.lang.NullPointerException(); + } + return internalGetProperties().getMap().containsKey(key); + } + /** Use {@link #getPropertiesMap()} instead. */ + @java.lang.Deprecated + public java.util.Map getProperties() { + return getPropertiesMap(); + } + /** + * + * + *
+     * Optional. A mapping of property names to values. Used to set Presto
+     * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+     * Equivalent to using the --session flag in the Presto CLI
+     * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public java.util.Map getPropertiesMap() { + return internalGetProperties().getMap(); + } + /** + * + * + *
+     * Optional. A mapping of property names to values. Used to set Presto
+     * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+     * Equivalent to using the --session flag in the Presto CLI
+     * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public java.lang.String getPropertiesOrDefault( + java.lang.String key, java.lang.String defaultValue) { + if (key == null) { + throw new java.lang.NullPointerException(); + } + java.util.Map map = internalGetProperties().getMap(); + return map.containsKey(key) ? map.get(key) : defaultValue; + } + /** + * + * + *
+     * Optional. A mapping of property names to values. Used to set Presto
+     * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+     * Equivalent to using the --session flag in the Presto CLI
+     * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public java.lang.String getPropertiesOrThrow(java.lang.String key) { + if (key == null) { + throw new java.lang.NullPointerException(); + } + java.util.Map map = internalGetProperties().getMap(); + if (!map.containsKey(key)) { + throw new java.lang.IllegalArgumentException(); + } + return map.get(key); + } + + public Builder clearProperties() { + internalGetMutableProperties().getMutableMap().clear(); + return this; + } + /** + * + * + *
+     * Optional. A mapping of property names to values. Used to set Presto
+     * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+     * Equivalent to using the --session flag in the Presto CLI
+     * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder removeProperties(java.lang.String key) { + if (key == null) { + throw new java.lang.NullPointerException(); + } + internalGetMutableProperties().getMutableMap().remove(key); + return this; + } + /** Use alternate mutation accessors instead. */ + @java.lang.Deprecated + public java.util.Map getMutableProperties() { + return internalGetMutableProperties().getMutableMap(); + } + /** + * + * + *
+     * Optional. A mapping of property names to values. Used to set Presto
+     * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+     * Equivalent to using the --session flag in the Presto CLI
+     * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder putProperties(java.lang.String key, java.lang.String value) { + if (key == null) { + throw new java.lang.NullPointerException(); + } + if (value == null) { + throw new java.lang.NullPointerException(); + } + internalGetMutableProperties().getMutableMap().put(key, value); + return this; + } + /** + * + * + *
+     * Optional. A mapping of property names to values. Used to set Presto
+     * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+     * Equivalent to using the --session flag in the Presto CLI
+     * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder putAllProperties(java.util.Map values) { + internalGetMutableProperties().getMutableMap().putAll(values); + return this; + } + + private com.google.cloud.dataproc.v1beta2.LoggingConfig loggingConfig_; + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.LoggingConfig, + com.google.cloud.dataproc.v1beta2.LoggingConfig.Builder, + com.google.cloud.dataproc.v1beta2.LoggingConfigOrBuilder> + loggingConfigBuilder_; + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the loggingConfig field is set. + */ + public boolean hasLoggingConfig() { + return loggingConfigBuilder_ != null || loggingConfig_ != null; + } + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The loggingConfig. + */ + public com.google.cloud.dataproc.v1beta2.LoggingConfig getLoggingConfig() { + if (loggingConfigBuilder_ == null) { + return loggingConfig_ == null + ? com.google.cloud.dataproc.v1beta2.LoggingConfig.getDefaultInstance() + : loggingConfig_; + } else { + return loggingConfigBuilder_.getMessage(); + } + } + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setLoggingConfig(com.google.cloud.dataproc.v1beta2.LoggingConfig value) { + if (loggingConfigBuilder_ == null) { + if (value == null) { + throw new NullPointerException(); + } + loggingConfig_ = value; + onChanged(); + } else { + loggingConfigBuilder_.setMessage(value); + } + + return this; + } + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder setLoggingConfig( + com.google.cloud.dataproc.v1beta2.LoggingConfig.Builder builderForValue) { + if (loggingConfigBuilder_ == null) { + loggingConfig_ = builderForValue.build(); + onChanged(); + } else { + loggingConfigBuilder_.setMessage(builderForValue.build()); + } + + return this; + } + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder mergeLoggingConfig(com.google.cloud.dataproc.v1beta2.LoggingConfig value) { + if (loggingConfigBuilder_ == null) { + if (loggingConfig_ != null) { + loggingConfig_ = + com.google.cloud.dataproc.v1beta2.LoggingConfig.newBuilder(loggingConfig_) + .mergeFrom(value) + .buildPartial(); + } else { + loggingConfig_ = value; + } + onChanged(); + } else { + loggingConfigBuilder_.mergeFrom(value); + } + + return this; + } + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public Builder clearLoggingConfig() { + if (loggingConfigBuilder_ == null) { + loggingConfig_ = null; + onChanged(); + } else { + loggingConfig_ = null; + loggingConfigBuilder_ = null; + } + + return this; + } + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.LoggingConfig.Builder getLoggingConfigBuilder() { + + onChanged(); + return getLoggingConfigFieldBuilder().getBuilder(); + } + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + public com.google.cloud.dataproc.v1beta2.LoggingConfigOrBuilder getLoggingConfigOrBuilder() { + if (loggingConfigBuilder_ != null) { + return loggingConfigBuilder_.getMessageOrBuilder(); + } else { + return loggingConfig_ == null + ? com.google.cloud.dataproc.v1beta2.LoggingConfig.getDefaultInstance() + : loggingConfig_; + } + } + /** + * + * + *
+     * Optional. The runtime log config for job execution.
+     * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + private com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.LoggingConfig, + com.google.cloud.dataproc.v1beta2.LoggingConfig.Builder, + com.google.cloud.dataproc.v1beta2.LoggingConfigOrBuilder> + getLoggingConfigFieldBuilder() { + if (loggingConfigBuilder_ == null) { + loggingConfigBuilder_ = + new com.google.protobuf.SingleFieldBuilderV3< + com.google.cloud.dataproc.v1beta2.LoggingConfig, + com.google.cloud.dataproc.v1beta2.LoggingConfig.Builder, + com.google.cloud.dataproc.v1beta2.LoggingConfigOrBuilder>( + getLoggingConfig(), getParentForChildren(), isClean()); + loggingConfig_ = null; + } + return loggingConfigBuilder_; + } + + @java.lang.Override + public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + + // @@protoc_insertion_point(builder_scope:google.cloud.dataproc.v1beta2.PrestoJob) + } + + // @@protoc_insertion_point(class_scope:google.cloud.dataproc.v1beta2.PrestoJob) + private static final com.google.cloud.dataproc.v1beta2.PrestoJob DEFAULT_INSTANCE; + + static { + DEFAULT_INSTANCE = new com.google.cloud.dataproc.v1beta2.PrestoJob(); + } + + public static com.google.cloud.dataproc.v1beta2.PrestoJob getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + private static final com.google.protobuf.Parser PARSER = + new com.google.protobuf.AbstractParser() { + @java.lang.Override + public PrestoJob parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new PrestoJob(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public com.google.cloud.dataproc.v1beta2.PrestoJob getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } +} diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PrestoJobOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PrestoJobOrBuilder.java new file mode 100644 index 00000000..371c7b70 --- /dev/null +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/PrestoJobOrBuilder.java @@ -0,0 +1,290 @@ +/* + * Copyright 2020 Google LLC + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dataproc/v1beta2/jobs.proto + +package com.google.cloud.dataproc.v1beta2; + +public interface PrestoJobOrBuilder + extends + // @@protoc_insertion_point(interface_extends:google.cloud.dataproc.v1beta2.PrestoJob) + com.google.protobuf.MessageOrBuilder { + + /** + * + * + *
+   * The HCFS URI of the script that contains SQL queries.
+   * 
+ * + * string query_file_uri = 1; + * + * @return The queryFileUri. + */ + java.lang.String getQueryFileUri(); + /** + * + * + *
+   * The HCFS URI of the script that contains SQL queries.
+   * 
+ * + * string query_file_uri = 1; + * + * @return The bytes for queryFileUri. + */ + com.google.protobuf.ByteString getQueryFileUriBytes(); + + /** + * + * + *
+   * A list of queries.
+   * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + * + * @return Whether the queryList field is set. + */ + boolean hasQueryList(); + /** + * + * + *
+   * A list of queries.
+   * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + * + * @return The queryList. + */ + com.google.cloud.dataproc.v1beta2.QueryList getQueryList(); + /** + * + * + *
+   * A list of queries.
+   * 
+ * + * .google.cloud.dataproc.v1beta2.QueryList query_list = 2; + */ + com.google.cloud.dataproc.v1beta2.QueryListOrBuilder getQueryListOrBuilder(); + + /** + * + * + *
+   * Optional. Whether to continue executing queries if a query fails.
+   * The default value is `false`. Setting to `true` can be useful when
+   * executing independent parallel queries.
+   * 
+ * + * bool continue_on_failure = 3 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The continueOnFailure. + */ + boolean getContinueOnFailure(); + + /** + * + * + *
+   * Optional. The format in which query output will be displayed. See the
+   * Presto documentation for supported output formats
+   * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The outputFormat. + */ + java.lang.String getOutputFormat(); + /** + * + * + *
+   * Optional. The format in which query output will be displayed. See the
+   * Presto documentation for supported output formats
+   * 
+ * + * string output_format = 4 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The bytes for outputFormat. + */ + com.google.protobuf.ByteString getOutputFormatBytes(); + + /** + * + * + *
+   * Optional. Presto client tags to attach to this query
+   * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return A list containing the clientTags. + */ + java.util.List getClientTagsList(); + /** + * + * + *
+   * Optional. Presto client tags to attach to this query
+   * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @return The count of clientTags. + */ + int getClientTagsCount(); + /** + * + * + *
+   * Optional. Presto client tags to attach to this query
+   * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param index The index of the element to return. + * @return The clientTags at the given index. + */ + java.lang.String getClientTags(int index); + /** + * + * + *
+   * Optional. Presto client tags to attach to this query
+   * 
+ * + * repeated string client_tags = 5 [(.google.api.field_behavior) = OPTIONAL]; + * + * @param index The index of the value to return. + * @return The bytes of the clientTags at the given index. + */ + com.google.protobuf.ByteString getClientTagsBytes(int index); + + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + int getPropertiesCount(); + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + boolean containsProperties(java.lang.String key); + /** Use {@link #getPropertiesMap()} instead. */ + @java.lang.Deprecated + java.util.Map getProperties(); + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + java.util.Map getPropertiesMap(); + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + java.lang.String getPropertiesOrDefault(java.lang.String key, java.lang.String defaultValue); + /** + * + * + *
+   * Optional. A mapping of property names to values. Used to set Presto
+   * [session properties](https://prestodb.io/docs/current/sql/set-session.html)
+   * Equivalent to using the --session flag in the Presto CLI
+   * 
+ * + * map<string, string> properties = 6 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + java.lang.String getPropertiesOrThrow(java.lang.String key); + + /** + * + * + *
+   * Optional. The runtime log config for job execution.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return Whether the loggingConfig field is set. + */ + boolean hasLoggingConfig(); + /** + * + * + *
+   * Optional. The runtime log config for job execution.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + * + * @return The loggingConfig. + */ + com.google.cloud.dataproc.v1beta2.LoggingConfig getLoggingConfig(); + /** + * + * + *
+   * Optional. The runtime log config for job execution.
+   * 
+ * + * + * .google.cloud.dataproc.v1beta2.LoggingConfig logging_config = 7 [(.google.api.field_behavior) = OPTIONAL]; + * + */ + com.google.cloud.dataproc.v1beta2.LoggingConfigOrBuilder getLoggingConfigOrBuilder(); + + public com.google.cloud.dataproc.v1beta2.PrestoJob.QueriesCase getQueriesCase(); +} diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SharedProto.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SharedProto.java index 06bcf8af..58e8dbe5 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SharedProto.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SharedProto.java @@ -37,14 +37,15 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { java.lang.String[] descriptorData = { "\n*google/cloud/dataproc/v1beta2/shared.p" + "roto\022\035google.cloud.dataproc.v1beta2\032\034goo" - + "gle/api/annotations.proto*\225\001\n\tComponent\022" + + "gle/api/annotations.proto*\302\001\n\tComponent\022" + "\031\n\025COMPONENT_UNSPECIFIED\020\000\022\014\n\010ANACONDA\020\005" - + "\022\t\n\005DRUID\020\t\022\020\n\014HIVE_WEBHCAT\020\003\022\013\n\007JUPYTER" - + "\020\001\022\014\n\010KERBEROS\020\007\022\n\n\006PRESTO\020\006\022\014\n\010ZEPPELIN" - + "\020\004\022\r\n\tZOOKEEPER\020\010By\n!com.google.cloud.da" - + "taproc.v1beta2B\013SharedProtoP\001ZEgoogle.go" - + "lang.org/genproto/googleapis/cloud/datap" - + "roc/v1beta2;dataprocb\006proto3" + + "\022\n\n\006DOCKER\020\r\022\t\n\005DRUID\020\t\022\t\n\005FLINK\020\016\022\020\n\014HI" + + "VE_WEBHCAT\020\003\022\013\n\007JUPYTER\020\001\022\014\n\010KERBEROS\020\007\022" + + "\n\n\006PRESTO\020\006\022\n\n\006RANGER\020\014\022\010\n\004SOLR\020\n\022\014\n\010ZEP" + + "PELIN\020\004\022\r\n\tZOOKEEPER\020\010By\n!com.google.clo" + + "ud.dataproc.v1beta2B\013SharedProtoP\001ZEgoog" + + "le.golang.org/genproto/googleapis/cloud/" + + "dataproc/v1beta2;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfig.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfig.java index 602335da..659ef005 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfig.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfig.java @@ -175,10 +175,10 @@ protected com.google.protobuf.MapField internalGetMapField(int number) { *
    * Optional. The version of software inside the cluster. It must be one of the
    * supported [Dataproc
-   * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+   * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
-   * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+   * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
    * If unspecified, it defaults to the latest Debian version.
    * 
* @@ -203,10 +203,10 @@ public java.lang.String getImageVersion() { *
    * Optional. The version of software inside the cluster. It must be one of the
    * supported [Dataproc
-   * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+   * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
-   * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+   * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
    * If unspecified, it defaults to the latest Debian version.
    * 
* @@ -268,8 +268,8 @@ public int getPropertiesCount() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -303,8 +303,8 @@ public java.util.Map getProperties() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -330,8 +330,8 @@ public java.util.Map getPropertiesMap() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -362,8 +362,8 @@ public java.lang.String getPropertiesOrDefault( * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -884,10 +884,10 @@ public Builder mergeFrom( *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -912,10 +912,10 @@ public java.lang.String getImageVersion() { *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -940,10 +940,10 @@ public com.google.protobuf.ByteString getImageVersionBytes() { *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -967,10 +967,10 @@ public Builder setImageVersion(java.lang.String value) { *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -990,10 +990,10 @@ public Builder clearImageVersion() { *
      * Optional. The version of software inside the cluster. It must be one of the
      * supported [Dataproc
-     * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+     * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
      * such as "1.2" (including a subminor version, such as "1.2.29"), or the
      * ["preview"
-     * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+     * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
      * If unspecified, it defaults to the latest Debian version.
      * 
* @@ -1058,8 +1058,8 @@ public int getPropertiesCount() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1093,8 +1093,8 @@ public java.util.Map getProperties() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1120,8 +1120,8 @@ public java.util.Map getPropertiesMap() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1152,8 +1152,8 @@ public java.lang.String getPropertiesOrDefault( * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1191,8 +1191,8 @@ public Builder clearProperties() { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1227,8 +1227,8 @@ public java.util.Map getMutableProperties() * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -1261,8 +1261,8 @@ public Builder putProperties(java.lang.String key, java.lang.String value) { * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfigOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfigOrBuilder.java index c58c696f..6b2f2fef 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfigOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/SoftwareConfigOrBuilder.java @@ -29,10 +29,10 @@ public interface SoftwareConfigOrBuilder *
    * Optional. The version of software inside the cluster. It must be one of the
    * supported [Dataproc
-   * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+   * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
-   * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+   * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
    * If unspecified, it defaults to the latest Debian version.
    * 
* @@ -47,10 +47,10 @@ public interface SoftwareConfigOrBuilder *
    * Optional. The version of software inside the cluster. It must be one of the
    * supported [Dataproc
-   * Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
+   * Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
    * such as "1.2" (including a subminor version, such as "1.2.29"), or the
    * ["preview"
-   * version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
+   * version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions).
    * If unspecified, it defaults to the latest Debian version.
    * 
* @@ -77,8 +77,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -102,8 +102,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -130,8 +130,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -155,8 +155,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; @@ -180,8 +180,8 @@ public interface SoftwareConfigOrBuilder * * pig: `pig.properties` * * spark: `spark-defaults.conf` * * yarn: `yarn-site.xml` - * For more information, see - * [Cluster properties](/dataproc/docs/concepts/cluster-properties). + * For more information, see [Cluster + * properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). * * * map<string, string> properties = 2 [(.google.api.field_behavior) = OPTIONAL]; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameter.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameter.java index 92202abc..5f89a52c 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameter.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameter.java @@ -227,10 +227,10 @@ public com.google.protobuf.ByteString getNameBytes() { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -276,10 +276,10 @@ public com.google.protobuf.ProtocolStringList getFieldsList() { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -325,10 +325,10 @@ public int getFieldsCount() { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -375,10 +375,10 @@ public java.lang.String getFields(int index) { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1071,10 +1071,10 @@ private void ensureFieldsIsMutable() { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1120,10 +1120,10 @@ public com.google.protobuf.ProtocolStringList getFieldsList() { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1169,10 +1169,10 @@ public int getFieldsCount() { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1219,10 +1219,10 @@ public java.lang.String getFields(int index) { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1269,10 +1269,10 @@ public com.google.protobuf.ByteString getFieldsBytes(int index) { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1326,10 +1326,10 @@ public Builder setFields(int index, java.lang.String value) { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1382,10 +1382,10 @@ public Builder addFields(java.lang.String value) { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1435,10 +1435,10 @@ public Builder addAllFields(java.lang.Iterable values) { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -1487,10 +1487,10 @@ public Builder clearFields() { * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameterOrBuilder.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameterOrBuilder.java index 580307f3..93c43366 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameterOrBuilder.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/TemplateParameterOrBuilder.java @@ -67,10 +67,10 @@ public interface TemplateParameterOrBuilder * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -114,10 +114,10 @@ public interface TemplateParameterOrBuilder * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -161,10 +161,10 @@ public interface TemplateParameterOrBuilder * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] @@ -209,10 +209,10 @@ public interface TemplateParameterOrBuilder * Required. Paths to all fields that the parameter replaces. * A field is allowed to appear in at most one parameter's list of field * paths. - * A field path is similar in syntax to a - * [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - * field path that references the zone field of a workflow template's cluster - * selector would be specified as `placement.clusterSelector.zone`. + * A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + * For example, a field path that references the zone field of a workflow + * template's cluster selector would be specified as + * `placement.clusterSelector.zone`. * Also, field paths can reference fields using the following syntax: * * Values in maps can be referenced by key: * * labels['key'] diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java index a993eec1..c9429838 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java +++ b/proto-google-cloud-dataproc-v1beta2/src/main/java/com/google/cloud/dataproc/v1beta2/WorkflowTemplatesProto.java @@ -186,156 +186,160 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "\004zone\030\001 \001(\t\022Y\n\016cluster_labels\030\002 \003(\0132A.go" + "ogle.cloud.dataproc.v1beta2.ClusterSelec" + "tor.ClusterLabelsEntry\0324\n\022ClusterLabelsE" - + "ntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\373\004\n" - + "\nOrderedJob\022\017\n\007step_id\030\001 \001(\t\022>\n\nhadoop_j" - + "ob\030\002 \001(\0132(.google.cloud.dataproc.v1beta2" - + ".HadoopJobH\000\022<\n\tspark_job\030\003 \001(\0132\'.google" - + ".cloud.dataproc.v1beta2.SparkJobH\000\022@\n\013py" - + "spark_job\030\004 \001(\0132).google.cloud.dataproc." - + "v1beta2.PySparkJobH\000\022:\n\010hive_job\030\005 \001(\0132&" - + ".google.cloud.dataproc.v1beta2.HiveJobH\000" - + "\0228\n\007pig_job\030\006 \001(\0132%.google.cloud.datapro" - + "c.v1beta2.PigJobH\000\022C\n\rspark_sql_job\030\007 \001(" - + "\0132*.google.cloud.dataproc.v1beta2.SparkS" - + "qlJobH\000\022E\n\006labels\030\010 \003(\01325.google.cloud.d" - + "ataproc.v1beta2.OrderedJob.LabelsEntry\022@" - + "\n\nscheduling\030\t \001(\0132,.google.cloud.datapr" - + "oc.v1beta2.JobScheduling\022\035\n\025prerequisite" - + "_step_ids\030\n \003(\t\032-\n\013LabelsEntry\022\013\n\003key\030\001 " - + "\001(\t\022\r\n\005value\030\002 \001(\t:\0028\001B\n\n\010job_type\"\216\001\n\021T" - + "emplateParameter\022\014\n\004name\030\001 \001(\t\022\016\n\006fields" - + "\030\002 \003(\t\022\023\n\013description\030\003 \001(\t\022F\n\nvalidatio" - + "n\030\004 \001(\01322.google.cloud.dataproc.v1beta2." - + "ParameterValidation\"\253\001\n\023ParameterValidat" - + "ion\022?\n\005regex\030\001 \001(\0132..google.cloud.datapr" - + "oc.v1beta2.RegexValidationH\000\022@\n\006values\030\002" - + " \001(\0132..google.cloud.dataproc.v1beta2.Val" - + "ueValidationH\000B\021\n\017validation_type\"\"\n\017Reg" - + "exValidation\022\017\n\007regexes\030\001 \003(\t\"!\n\017ValueVa" - + "lidation\022\016\n\006values\030\001 \003(\t\"\310\005\n\020WorkflowMet" - + "adata\022\025\n\010template\030\001 \001(\tB\003\340A\003\022\024\n\007version\030" - + "\002 \001(\005B\003\340A\003\022L\n\016create_cluster\030\003 \001(\0132/.goo" - + "gle.cloud.dataproc.v1beta2.ClusterOperat" - + "ionB\003\340A\003\022@\n\005graph\030\004 \001(\0132,.google.cloud.d" - + "ataproc.v1beta2.WorkflowGraphB\003\340A\003\022L\n\016de" - + "lete_cluster\030\005 \001(\0132/.google.cloud.datapr" - + "oc.v1beta2.ClusterOperationB\003\340A\003\022I\n\005stat" - + "e\030\006 \001(\01625.google.cloud.dataproc.v1beta2." - + "WorkflowMetadata.StateB\003\340A\003\022\031\n\014cluster_n" - + "ame\030\007 \001(\tB\003\340A\003\022S\n\nparameters\030\010 \003(\0132?.goo" - + "gle.cloud.dataproc.v1beta2.WorkflowMetad" - + "ata.ParametersEntry\0223\n\nstart_time\030\t \001(\0132" - + "\032.google.protobuf.TimestampB\003\340A\003\0221\n\010end_" - + "time\030\n \001(\0132\032.google.protobuf.TimestampB\003" - + "\340A\003\022\031\n\014cluster_uuid\030\013 \001(\tB\003\340A\003\0321\n\017Parame" - + "tersEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028" - + "\001\"8\n\005State\022\013\n\007UNKNOWN\020\000\022\013\n\007PENDING\020\001\022\013\n\007" - + "RUNNING\020\002\022\010\n\004DONE\020\003\"T\n\020ClusterOperation\022" - + "\031\n\014operation_id\030\001 \001(\tB\003\340A\003\022\022\n\005error\030\002 \001(" - + "\tB\003\340A\003\022\021\n\004done\030\003 \001(\010B\003\340A\003\"P\n\rWorkflowGra" - + "ph\022?\n\005nodes\030\001 \003(\0132+.google.cloud.datapro" - + "c.v1beta2.WorkflowNodeB\003\340A\003\"\251\002\n\014Workflow" - + "Node\022\024\n\007step_id\030\001 \001(\tB\003\340A\003\022\"\n\025prerequisi" - + "te_step_ids\030\002 \003(\tB\003\340A\003\022\023\n\006job_id\030\003 \001(\tB\003" - + "\340A\003\022I\n\005state\030\005 \001(\01625.google.cloud.datapr" - + "oc.v1beta2.WorkflowNode.NodeStateB\003\340A\003\022\022" - + "\n\005error\030\006 \001(\tB\003\340A\003\"k\n\tNodeState\022\033\n\027NODE_" - + "STATUS_UNSPECIFIED\020\000\022\013\n\007BLOCKED\020\001\022\014\n\010RUN" - + "NABLE\020\002\022\013\n\007RUNNING\020\003\022\r\n\tCOMPLETED\020\004\022\n\n\006F" - + "AILED\020\005\"\251\001\n\035CreateWorkflowTemplateReques" - + "t\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(dataproc.goog" - + "leapis.com/WorkflowTemplate\022F\n\010template\030" - + "\002 \001(\0132/.google.cloud.dataproc.v1beta2.Wo" - + "rkflowTemplateB\003\340A\002\"m\n\032GetWorkflowTempla" - + "teRequest\022>\n\004name\030\001 \001(\tB0\340A\002\372A*\n(datapro" - + "c.googleapis.com/WorkflowTemplate\022\017\n\007ver" - + "sion\030\002 \001(\005\"\274\002\n\"InstantiateWorkflowTempla" - + "teRequest\022>\n\004name\030\001 \001(\tB0\340A\002\372A*\n(datapro" - + "c.googleapis.com/WorkflowTemplate\022\017\n\007ver" - + "sion\030\002 \001(\005\022\027\n\013instance_id\030\003 \001(\tB\002\030\001\022\022\n\nr" - + "equest_id\030\005 \001(\t\022e\n\nparameters\030\004 \003(\0132Q.go" - + "ogle.cloud.dataproc.v1beta2.InstantiateW" - + "orkflowTemplateRequest.ParametersEntry\0321" - + "\n\017ParametersEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030" - + "\002 \001(\t:\0028\001\"\335\001\n(InstantiateInlineWorkflowT" - + "emplateRequest\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(" - + "dataproc.googleapis.com/WorkflowTemplate" - + "\022F\n\010template\030\002 \001(\0132/.google.cloud.datapr" - + "oc.v1beta2.WorkflowTemplateB\003\340A\002\022\023\n\013inst" - + "ance_id\030\003 \001(\t\022\022\n\nrequest_id\030\004 \001(\t\"g\n\035Upd" - + "ateWorkflowTemplateRequest\022F\n\010template\030\001" - + " \001(\0132/.google.cloud.dataproc.v1beta2.Wor" - + "kflowTemplateB\003\340A\002\"\207\001\n\034ListWorkflowTempl" - + "atesRequest\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(dat" - + "aproc.googleapis.com/WorkflowTemplate\022\021\n" - + "\tpage_size\030\002 \001(\005\022\022\n\npage_token\030\003 \001(\t\"\206\001\n" - + "\035ListWorkflowTemplatesResponse\022G\n\ttempla" - + "tes\030\001 \003(\0132/.google.cloud.dataproc.v1beta" - + "2.WorkflowTemplateB\003\340A\003\022\034\n\017next_page_tok" - + "en\030\002 \001(\tB\003\340A\003\"p\n\035DeleteWorkflowTemplateR" - + "equest\022>\n\004name\030\001 \001(\tB0\340A\002\372A*\n(dataproc.g" - + "oogleapis.com/WorkflowTemplate\022\017\n\007versio" - + "n\030\002 \001(\0052\351\021\n\027WorkflowTemplateService\022\260\002\n\026" - + "CreateWorkflowTemplate\022<.google.cloud.da" - + "taproc.v1beta2.CreateWorkflowTemplateReq" - + "uest\032/.google.cloud.dataproc.v1beta2.Wor" - + "kflowTemplate\"\246\001\202\323\344\223\002\214\001\"8/v1beta2/{paren" - + "t=projects/*/regions/*}/workflowTemplate" - + "s:\010templateZF\":/v1beta2/{parent=projects" - + "/*/locations/*}/workflowTemplates:\010templ" - + "ate\332A\020parent, template\022\211\002\n\023GetWorkflowTe" - + "mplate\0229.google.cloud.dataproc.v1beta2.G" - + "etWorkflowTemplateRequest\032/.google.cloud" - + ".dataproc.v1beta2.WorkflowTemplate\"\205\001\202\323\344" - + "\223\002x\0228/v1beta2/{name=projects/*/regions/*" - + "/workflowTemplates/*}Z<\022:/v1beta2/{name=" - + "projects/*/locations/*/workflowTemplates" - + "/*}\332A\004name\022\345\002\n\033InstantiateWorkflowTempla" - + "te\022A.google.cloud.dataproc.v1beta2.Insta" - + "ntiateWorkflowTemplateRequest\032\035.google.l" - + "ongrunning.Operation\"\343\001\202\323\344\223\002\226\001\"D/v1beta2" - + "/{name=projects/*/regions/*/workflowTemp" - + "lates/*}:instantiate:\001*ZK\"F/v1beta2/{nam" - + "e=projects/*/locations/*/workflowTemplat" - + "es/*}:instantiate:\001*\332A\004name\332A\020name, para" - + "meters\312A)\n\025google.protobuf.Empty\022\020Workfl" - + "owMetadata\022\204\003\n!InstantiateInlineWorkflow" - + "Template\022G.google.cloud.dataproc.v1beta2" - + ".InstantiateInlineWorkflowTemplateReques" - + "t\032\035.google.longrunning.Operation\"\366\001\202\323\344\223\002" - + "\260\001\"L/v1beta2/{parent=projects/*/location" - + "s/*}/workflowTemplates:instantiateInline" - + ":\010templateZV\"J/v1beta2/{parent=projects/" - + "*/regions/*}/workflowTemplates:instantia" - + "teInline:\010template\332A\020parent, template\312A)" - + "\n\025google.protobuf.Empty\022\020WorkflowMetadat" - + "a\022\272\002\n\026UpdateWorkflowTemplate\022<.google.cl" - + "oud.dataproc.v1beta2.UpdateWorkflowTempl" - + "ateRequest\032/.google.cloud.dataproc.v1bet" - + "a2.WorkflowTemplate\"\260\001\202\323\344\223\002\236\001\032A/v1beta2/" - + "{template.name=projects/*/regions/*/work" - + "flowTemplates/*}:\010templateZO\032C/v1beta2/{" - + "template.name=projects/*/locations/*/wor" - + "kflowTemplates/*}:\010template\332A\010template\022\234" - + "\002\n\025ListWorkflowTemplates\022;.google.cloud." - + "dataproc.v1beta2.ListWorkflowTemplatesRe" - + "quest\032<.google.cloud.dataproc.v1beta2.Li" - + "stWorkflowTemplatesResponse\"\207\001\202\323\344\223\002x\0228/v" - + "1beta2/{parent=projects/*/regions/*}/wor" - + "kflowTemplatesZ<\022:/v1beta2/{parent=proje" - + "cts/*/locations/*}/workflowTemplates\332A\006p" - + "arent\022\366\001\n\026DeleteWorkflowTemplate\022<.googl" - + "e.cloud.dataproc.v1beta2.DeleteWorkflowT" - + "emplateRequest\032\026.google.protobuf.Empty\"\205" - + "\001\202\323\344\223\002x*8/v1beta2/{name=projects/*/regio" - + "ns/*/workflowTemplates/*}Z<*:/v1beta2/{n" - + "ame=projects/*/locations/*/workflowTempl" - + "ates/*}\332A\004name\032K\312A\027dataproc.googleapis.c" - + "om\322A.https://www.googleapis.com/auth/clo" - + "ud-platformB\204\001\n!com.google.cloud.datapro" - + "c.v1beta2B\026WorkflowTemplatesProtoP\001ZEgoo" - + "gle.golang.org/genproto/googleapis/cloud" - + "/dataproc/v1beta2;dataprocb\006proto3" + + "ntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\220\006\n" + + "\nOrderedJob\022\024\n\007step_id\030\001 \001(\tB\003\340A\002\022>\n\nhad" + + "oop_job\030\002 \001(\0132(.google.cloud.dataproc.v1" + + "beta2.HadoopJobH\000\022<\n\tspark_job\030\003 \001(\0132\'.g" + + "oogle.cloud.dataproc.v1beta2.SparkJobH\000\022" + + "@\n\013pyspark_job\030\004 \001(\0132).google.cloud.data" + + "proc.v1beta2.PySparkJobH\000\022:\n\010hive_job\030\005 " + + "\001(\0132&.google.cloud.dataproc.v1beta2.Hive" + + "JobH\000\0228\n\007pig_job\030\006 \001(\0132%.google.cloud.da" + + "taproc.v1beta2.PigJobH\000\022?\n\013spark_r_job\030\013" + + " \001(\0132(.google.cloud.dataproc.v1beta2.Spa" + + "rkRJobH\000\022C\n\rspark_sql_job\030\007 \001(\0132*.google" + + ".cloud.dataproc.v1beta2.SparkSqlJobH\000\022>\n" + + "\npresto_job\030\014 \001(\0132(.google.cloud.datapro" + + "c.v1beta2.PrestoJobH\000\022J\n\006labels\030\010 \003(\01325." + + "google.cloud.dataproc.v1beta2.OrderedJob" + + ".LabelsEntryB\003\340A\001\022E\n\nscheduling\030\t \001(\0132,." + + "google.cloud.dataproc.v1beta2.JobSchedul" + + "ingB\003\340A\001\022\"\n\025prerequisite_step_ids\030\n \003(\tB" + + "\003\340A\001\032-\n\013LabelsEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005valu" + + "e\030\002 \001(\t:\0028\001B\n\n\010job_type\"\216\001\n\021TemplatePara" + + "meter\022\014\n\004name\030\001 \001(\t\022\016\n\006fields\030\002 \003(\t\022\023\n\013d" + + "escription\030\003 \001(\t\022F\n\nvalidation\030\004 \001(\01322.g" + + "oogle.cloud.dataproc.v1beta2.ParameterVa" + + "lidation\"\253\001\n\023ParameterValidation\022?\n\005rege" + + "x\030\001 \001(\0132..google.cloud.dataproc.v1beta2." + + "RegexValidationH\000\022@\n\006values\030\002 \001(\0132..goog" + + "le.cloud.dataproc.v1beta2.ValueValidatio" + + "nH\000B\021\n\017validation_type\"\"\n\017RegexValidatio" + + "n\022\017\n\007regexes\030\001 \003(\t\"!\n\017ValueValidation\022\016\n" + + "\006values\030\001 \003(\t\"\310\005\n\020WorkflowMetadata\022\025\n\010te" + + "mplate\030\001 \001(\tB\003\340A\003\022\024\n\007version\030\002 \001(\005B\003\340A\003\022" + + "L\n\016create_cluster\030\003 \001(\0132/.google.cloud.d" + + "ataproc.v1beta2.ClusterOperationB\003\340A\003\022@\n" + + "\005graph\030\004 \001(\0132,.google.cloud.dataproc.v1b" + + "eta2.WorkflowGraphB\003\340A\003\022L\n\016delete_cluste" + + "r\030\005 \001(\0132/.google.cloud.dataproc.v1beta2." + + "ClusterOperationB\003\340A\003\022I\n\005state\030\006 \001(\01625.g" + + "oogle.cloud.dataproc.v1beta2.WorkflowMet" + + "adata.StateB\003\340A\003\022\031\n\014cluster_name\030\007 \001(\tB\003" + + "\340A\003\022S\n\nparameters\030\010 \003(\0132?.google.cloud.d" + + "ataproc.v1beta2.WorkflowMetadata.Paramet" + + "ersEntry\0223\n\nstart_time\030\t \001(\0132\032.google.pr" + + "otobuf.TimestampB\003\340A\003\0221\n\010end_time\030\n \001(\0132" + + "\032.google.protobuf.TimestampB\003\340A\003\022\031\n\014clus" + + "ter_uuid\030\013 \001(\tB\003\340A\003\0321\n\017ParametersEntry\022\013" + + "\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"8\n\005State\022" + + "\013\n\007UNKNOWN\020\000\022\013\n\007PENDING\020\001\022\013\n\007RUNNING\020\002\022\010" + + "\n\004DONE\020\003\"T\n\020ClusterOperation\022\031\n\014operatio" + + "n_id\030\001 \001(\tB\003\340A\003\022\022\n\005error\030\002 \001(\tB\003\340A\003\022\021\n\004d" + + "one\030\003 \001(\010B\003\340A\003\"P\n\rWorkflowGraph\022?\n\005nodes" + + "\030\001 \003(\0132+.google.cloud.dataproc.v1beta2.W" + + "orkflowNodeB\003\340A\003\"\251\002\n\014WorkflowNode\022\024\n\007ste" + + "p_id\030\001 \001(\tB\003\340A\003\022\"\n\025prerequisite_step_ids" + + "\030\002 \003(\tB\003\340A\003\022\023\n\006job_id\030\003 \001(\tB\003\340A\003\022I\n\005stat" + + "e\030\005 \001(\01625.google.cloud.dataproc.v1beta2." + + "WorkflowNode.NodeStateB\003\340A\003\022\022\n\005error\030\006 \001" + + "(\tB\003\340A\003\"k\n\tNodeState\022\033\n\027NODE_STATUS_UNSP" + + "ECIFIED\020\000\022\013\n\007BLOCKED\020\001\022\014\n\010RUNNABLE\020\002\022\013\n\007" + + "RUNNING\020\003\022\r\n\tCOMPLETED\020\004\022\n\n\006FAILED\020\005\"\251\001\n" + + "\035CreateWorkflowTemplateRequest\022@\n\006parent" + + "\030\001 \001(\tB0\340A\002\372A*\022(dataproc.googleapis.com/" + + "WorkflowTemplate\022F\n\010template\030\002 \001(\0132/.goo" + + "gle.cloud.dataproc.v1beta2.WorkflowTempl" + + "ateB\003\340A\002\"m\n\032GetWorkflowTemplateRequest\022>" + + "\n\004name\030\001 \001(\tB0\340A\002\372A*\n(dataproc.googleapi" + + "s.com/WorkflowTemplate\022\017\n\007version\030\002 \001(\005\"" + + "\274\002\n\"InstantiateWorkflowTemplateRequest\022>" + + "\n\004name\030\001 \001(\tB0\340A\002\372A*\n(dataproc.googleapi" + + "s.com/WorkflowTemplate\022\017\n\007version\030\002 \001(\005\022" + + "\027\n\013instance_id\030\003 \001(\tB\002\030\001\022\022\n\nrequest_id\030\005" + + " \001(\t\022e\n\nparameters\030\004 \003(\0132Q.google.cloud." + + "dataproc.v1beta2.InstantiateWorkflowTemp" + + "lateRequest.ParametersEntry\0321\n\017Parameter" + + "sEntry\022\013\n\003key\030\001 \001(\t\022\r\n\005value\030\002 \001(\t:\0028\001\"\335" + + "\001\n(InstantiateInlineWorkflowTemplateRequ" + + "est\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(dataproc.go" + + "ogleapis.com/WorkflowTemplate\022F\n\010templat" + + "e\030\002 \001(\0132/.google.cloud.dataproc.v1beta2." + + "WorkflowTemplateB\003\340A\002\022\023\n\013instance_id\030\003 \001" + + "(\t\022\022\n\nrequest_id\030\004 \001(\t\"g\n\035UpdateWorkflow" + + "TemplateRequest\022F\n\010template\030\001 \001(\0132/.goog" + + "le.cloud.dataproc.v1beta2.WorkflowTempla" + + "teB\003\340A\002\"\207\001\n\034ListWorkflowTemplatesRequest" + + "\022@\n\006parent\030\001 \001(\tB0\340A\002\372A*\022(dataproc.googl" + + "eapis.com/WorkflowTemplate\022\021\n\tpage_size\030" + + "\002 \001(\005\022\022\n\npage_token\030\003 \001(\t\"\206\001\n\035ListWorkfl" + + "owTemplatesResponse\022G\n\ttemplates\030\001 \003(\0132/" + + ".google.cloud.dataproc.v1beta2.WorkflowT" + + "emplateB\003\340A\003\022\034\n\017next_page_token\030\002 \001(\tB\003\340" + + "A\003\"p\n\035DeleteWorkflowTemplateRequest\022>\n\004n" + + "ame\030\001 \001(\tB0\340A\002\372A*\n(dataproc.googleapis.c" + + "om/WorkflowTemplate\022\017\n\007version\030\002 \001(\0052\351\021\n" + + "\027WorkflowTemplateService\022\260\002\n\026CreateWorkf" + + "lowTemplate\022<.google.cloud.dataproc.v1be" + + "ta2.CreateWorkflowTemplateRequest\032/.goog" + + "le.cloud.dataproc.v1beta2.WorkflowTempla" + + "te\"\246\001\202\323\344\223\002\214\001\"8/v1beta2/{parent=projects/" + + "*/regions/*}/workflowTemplates:\010template" + + "ZF\":/v1beta2/{parent=projects/*/location" + + "s/*}/workflowTemplates:\010template\332A\020paren" + + "t, template\022\211\002\n\023GetWorkflowTemplate\0229.go" + + "ogle.cloud.dataproc.v1beta2.GetWorkflowT" + + "emplateRequest\032/.google.cloud.dataproc.v" + + "1beta2.WorkflowTemplate\"\205\001\202\323\344\223\002x\0228/v1bet" + + "a2/{name=projects/*/regions/*/workflowTe" + + "mplates/*}Z<\022:/v1beta2/{name=projects/*/" + + "locations/*/workflowTemplates/*}\332A\004name\022" + + "\345\002\n\033InstantiateWorkflowTemplate\022A.google" + + ".cloud.dataproc.v1beta2.InstantiateWorkf" + + "lowTemplateRequest\032\035.google.longrunning." + + "Operation\"\343\001\202\323\344\223\002\226\001\"D/v1beta2/{name=proj" + + "ects/*/regions/*/workflowTemplates/*}:in" + + "stantiate:\001*ZK\"F/v1beta2/{name=projects/" + + "*/locations/*/workflowTemplates/*}:insta" + + "ntiate:\001*\332A\004name\332A\020name, parameters\312A)\n\025" + + "google.protobuf.Empty\022\020WorkflowMetadata\022" + + "\204\003\n!InstantiateInlineWorkflowTemplate\022G." + + "google.cloud.dataproc.v1beta2.Instantiat" + + "eInlineWorkflowTemplateRequest\032\035.google." + + "longrunning.Operation\"\366\001\202\323\344\223\002\260\001\"L/v1beta" + + "2/{parent=projects/*/locations/*}/workfl" + + "owTemplates:instantiateInline:\010templateZ" + + "V\"J/v1beta2/{parent=projects/*/regions/*" + + "}/workflowTemplates:instantiateInline:\010t" + + "emplate\332A\020parent, template\312A)\n\025google.pr" + + "otobuf.Empty\022\020WorkflowMetadata\022\272\002\n\026Updat" + + "eWorkflowTemplate\022<.google.cloud.datapro" + + "c.v1beta2.UpdateWorkflowTemplateRequest\032" + + "/.google.cloud.dataproc.v1beta2.Workflow" + + "Template\"\260\001\202\323\344\223\002\236\001\032A/v1beta2/{template.n" + + "ame=projects/*/regions/*/workflowTemplat" + + "es/*}:\010templateZO\032C/v1beta2/{template.na" + + "me=projects/*/locations/*/workflowTempla" + + "tes/*}:\010template\332A\010template\022\234\002\n\025ListWork" + + "flowTemplates\022;.google.cloud.dataproc.v1" + + "beta2.ListWorkflowTemplatesRequest\032<.goo" + + "gle.cloud.dataproc.v1beta2.ListWorkflowT" + + "emplatesResponse\"\207\001\202\323\344\223\002x\0228/v1beta2/{par" + + "ent=projects/*/regions/*}/workflowTempla" + + "tesZ<\022:/v1beta2/{parent=projects/*/locat" + + "ions/*}/workflowTemplates\332A\006parent\022\366\001\n\026D" + + "eleteWorkflowTemplate\022<.google.cloud.dat" + + "aproc.v1beta2.DeleteWorkflowTemplateRequ" + + "est\032\026.google.protobuf.Empty\"\205\001\202\323\344\223\002x*8/v" + + "1beta2/{name=projects/*/regions/*/workfl" + + "owTemplates/*}Z<*:/v1beta2/{name=project" + + "s/*/locations/*/workflowTemplates/*}\332A\004n" + + "ame\032K\312A\027dataproc.googleapis.com\322A.https:" + + "//www.googleapis.com/auth/cloud-platform" + + "B\204\001\n!com.google.cloud.dataproc.v1beta2B\026" + + "WorkflowTemplatesProtoP\001ZEgoogle.golang." + + "org/genproto/googleapis/cloud/dataproc/v" + + "1beta2;dataprocb\006proto3" }; descriptor = com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom( @@ -433,7 +437,9 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { "PysparkJob", "HiveJob", "PigJob", + "SparkRJob", "SparkSqlJob", + "PrestoJob", "Labels", "Scheduling", "PrerequisiteStepIds", diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/autoscaling_policies.proto b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/autoscaling_policies.proto index a5a7d56e..a7d6376b 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/autoscaling_policies.proto +++ b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/autoscaling_policies.proto @@ -36,12 +36,10 @@ option (google.api.resource_definition) = { // Cloud Dataproc API. service AutoscalingPolicyService { option (google.api.default_host) = "dataproc.googleapis.com"; - option (google.api.oauth_scopes) = - "https://www.googleapis.com/auth/cloud-platform"; + option (google.api.oauth_scopes) = "https://www.googleapis.com/auth/cloud-platform"; // Creates new autoscaling policy. - rpc CreateAutoscalingPolicy(CreateAutoscalingPolicyRequest) - returns (AutoscalingPolicy) { + rpc CreateAutoscalingPolicy(CreateAutoscalingPolicyRequest) returns (AutoscalingPolicy) { option (google.api.http) = { post: "/v1beta2/{parent=projects/*/locations/*}/autoscalingPolicies" body: "policy" @@ -57,8 +55,7 @@ service AutoscalingPolicyService { // // Disabled check for update_mask, because all updates will be full // replacements. - rpc UpdateAutoscalingPolicy(UpdateAutoscalingPolicyRequest) - returns (AutoscalingPolicy) { + rpc UpdateAutoscalingPolicy(UpdateAutoscalingPolicyRequest) returns (AutoscalingPolicy) { option (google.api.http) = { put: "/v1beta2/{policy.name=projects/*/locations/*/autoscalingPolicies/*}" body: "policy" @@ -71,8 +68,7 @@ service AutoscalingPolicyService { } // Retrieves autoscaling policy. - rpc GetAutoscalingPolicy(GetAutoscalingPolicyRequest) - returns (AutoscalingPolicy) { + rpc GetAutoscalingPolicy(GetAutoscalingPolicyRequest) returns (AutoscalingPolicy) { option (google.api.http) = { get: "/v1beta2/{name=projects/*/locations/*/autoscalingPolicies/*}" additional_bindings { @@ -83,8 +79,7 @@ service AutoscalingPolicyService { } // Lists autoscaling policies in the project. - rpc ListAutoscalingPolicies(ListAutoscalingPoliciesRequest) - returns (ListAutoscalingPoliciesResponse) { + rpc ListAutoscalingPolicies(ListAutoscalingPoliciesRequest) returns (ListAutoscalingPoliciesResponse) { option (google.api.http) = { get: "/v1beta2/{parent=projects/*/locations/*}/autoscalingPolicies" additional_bindings { @@ -96,8 +91,7 @@ service AutoscalingPolicyService { // Deletes an autoscaling policy. It is an error to delete an autoscaling // policy that is in use by one or more clusters. - rpc DeleteAutoscalingPolicy(DeleteAutoscalingPolicyRequest) - returns (google.protobuf.Empty) { + rpc DeleteAutoscalingPolicy(DeleteAutoscalingPolicyRequest) returns (google.protobuf.Empty) { option (google.api.http) = { delete: "/v1beta2/{name=projects/*/locations/*/autoscalingPolicies/*}" additional_bindings { @@ -142,26 +136,22 @@ message AutoscalingPolicy { } // Required. Describes how the autoscaler will operate for primary workers. - InstanceGroupAutoscalingPolicyConfig worker_config = 4 - [(google.api.field_behavior) = REQUIRED]; + InstanceGroupAutoscalingPolicyConfig worker_config = 4 [(google.api.field_behavior) = REQUIRED]; // Optional. Describes how the autoscaler will operate for secondary workers. - InstanceGroupAutoscalingPolicyConfig secondary_worker_config = 5 - [(google.api.field_behavior) = OPTIONAL]; + InstanceGroupAutoscalingPolicyConfig secondary_worker_config = 5 [(google.api.field_behavior) = OPTIONAL]; } // Basic algorithm for autoscaling. message BasicAutoscalingAlgorithm { // Required. YARN autoscaling configuration. - BasicYarnAutoscalingConfig yarn_config = 1 - [(google.api.field_behavior) = REQUIRED]; + BasicYarnAutoscalingConfig yarn_config = 1 [(google.api.field_behavior) = REQUIRED]; // Optional. Duration between scaling events. A scaling period starts after // the update operation from the previous event has completed. // // Bounds: [2m, 1d]. Default: 2m. - google.protobuf.Duration cooldown_period = 2 - [(google.api.field_behavior) = OPTIONAL]; + google.protobuf.Duration cooldown_period = 2 [(google.api.field_behavior) = OPTIONAL]; } // Basic autoscaling configurations for YARN. @@ -172,8 +162,7 @@ message BasicYarnAutoscalingConfig { // downscaling operations. // // Bounds: [0s, 1d]. - google.protobuf.Duration graceful_decommission_timeout = 5 - [(google.api.field_behavior) = REQUIRED]; + google.protobuf.Duration graceful_decommission_timeout = 5 [(google.api.field_behavior) = REQUIRED]; // Required. Fraction of average pending memory in the last cooldown period // for which to add workers. A scale-up factor of 1.0 will result in scaling @@ -200,8 +189,7 @@ message BasicYarnAutoscalingConfig { // on any recommended change. // // Bounds: [0.0, 1.0]. Default: 0.0. - double scale_up_min_worker_fraction = 3 - [(google.api.field_behavior) = OPTIONAL]; + double scale_up_min_worker_fraction = 3 [(google.api.field_behavior) = OPTIONAL]; // Optional. Minimum scale-down threshold as a fraction of total cluster size // before scaling occurs. For example, in a 20-worker cluster, a threshold of @@ -210,8 +198,7 @@ message BasicYarnAutoscalingConfig { // on any recommended change. // // Bounds: [0.0, 1.0]. Default: 0.0. - double scale_down_min_worker_fraction = 4 - [(google.api.field_behavior) = OPTIONAL]; + double scale_down_min_worker_fraction = 4 [(google.api.field_behavior) = OPTIONAL]; } // Configuration for the size bounds of an instance group, including its @@ -354,8 +341,7 @@ message ListAutoscalingPoliciesRequest { // A response to a request to list autoscaling policies in a project. message ListAutoscalingPoliciesResponse { // Output only. Autoscaling policies list. - repeated AutoscalingPolicy policies = 1 - [(google.api.field_behavior) = OUTPUT_ONLY]; + repeated AutoscalingPolicy policies = 1 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. This token is included in the response if there are more // results to fetch. diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto index 2e9e648c..93e6fe79 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto +++ b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/clusters.proto @@ -89,7 +89,7 @@ service ClusterController { option (google.api.method_signature) = "project_id, region, cluster_name"; } - // Lists all regions/{region}/clusters in a project. + // Lists all regions/{region}/clusters in a project alphabetically. rpc ListClusters(ListClustersRequest) returns (ListClustersResponse) { option (google.api.http) = { get: "/v1beta2/projects/{project_id}/regions/{region}/clusters" @@ -168,7 +168,7 @@ message ClusterConfig { // Compute Engine zone where your cluster is deployed, and then create // and manage this project-level, per-location bucket (see // [Dataproc staging - // bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)). + // bucket](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/staging-bucket)). string config_bucket = 1 [(google.api.field_behavior) = OPTIONAL]; // Optional. The shared Compute Engine config settings for @@ -220,6 +220,33 @@ message ClusterConfig { // Optional. Security related configuration. SecurityConfig security_config = 18 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. The Kubernetes Engine config for Dataproc clusters deployed to Kubernetes. + // Setting this is considered mutually exclusive with Compute Engine-based + // options such as `gce_cluster_config`, `master_config`, `worker_config`, + // `secondary_worker_config`, and `autoscaling_config`. + GkeClusterConfig gke_cluster_config = 19 [(google.api.field_behavior) = OPTIONAL]; +} + +// The GKE config for this cluster. +message GkeClusterConfig { + // A full, namespace-isolated deployment target for an existing GKE cluster. + message NamespacedGkeDeploymentTarget { + // Optional. The target GKE cluster to deploy to. + // Format: 'projects/{project}/locations/{location}/clusters/{cluster_id}' + string target_gke_cluster = 1 [ + (google.api.field_behavior) = OPTIONAL, + (google.api.resource_reference) = { + type: "container.googleapis.com/Cluster" + } + ]; + + // Optional. A namespace within the GKE cluster to deploy into. + string cluster_namespace = 2 [(google.api.field_behavior) = OPTIONAL]; + } + + // Optional. A target for the deployment. + NamespacedGkeDeploymentTarget namespaced_gke_deployment_target = 1 [(google.api.field_behavior) = OPTIONAL]; } // Endpoint config for this cluster @@ -274,7 +301,8 @@ message GceClusterConfig { // communications. Cannot be specified with subnetwork_uri. If neither // `network_uri` nor `subnetwork_uri` is specified, the "default" network of // the project is used, if it exists. Cannot be a "Custom Subnet Network" (see - // [Using Subnetworks](/compute/docs/subnetworks) for more information). + // [Using Subnetworks](https://cloud.google.com/compute/docs/subnetworks) for + // more information). // // A full URL, partial URI, or short name are valid. Examples: // @@ -302,15 +330,15 @@ message GceClusterConfig { bool internal_ip_only = 7 [(google.api.field_behavior) = OPTIONAL]; // Optional. The [Dataproc service - // account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc) + // account](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc) // (also see [VM Data Plane - // identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity)) + // identity](https://cloud.google.com/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity)) // used by Dataproc cluster VM instances to access Google Cloud Platform // services. // // If not specified, the // [Compute Engine default service - // account](/compute/docs/access/service-accounts#default_service_account) + // account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) // is used. string service_account = 8 [(google.api.field_behavior) = OPTIONAL]; @@ -330,8 +358,8 @@ message GceClusterConfig { // * https://www.googleapis.com/auth/devstorage.full_control repeated string service_account_scopes = 3 [(google.api.field_behavior) = OPTIONAL]; - // The Compute Engine tags to add to all instances (see - // [Tagging instances](/compute/docs/label-or-tag-resources#tags)). + // The Compute Engine tags to add to all instances (see [Tagging + // instances](https://cloud.google.com/compute/docs/label-or-tag-resources#tags)). repeated string tags = 4; // The Compute Engine metadata entries to add to all instances (see @@ -354,9 +382,24 @@ message InstanceGroupConfig { // from `cluster_name`, `num_instances`, and the instance group. repeated string instance_names = 2 [(google.api.field_behavior) = OUTPUT_ONLY]; - // Optional. The Compute Engine image resource used for cluster - // instances. It can be specified or may be inferred from - // `SoftwareConfig.image_version`. + // Optional. The Compute Engine image resource used for cluster instances. + // + // The URI can represent an image or image family. + // + // Image examples: + // + // * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id]` + // * `projects/[project_id]/global/images/[image-id]` + // * `image-id` + // + // Image family examples. Dataproc will use the most recent + // image from the family: + // + // * `https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name]` + // * `projects/[project_id]/global/images/family/[custom-image-family-name]` + // + // If the URI is unspecified, it will be inferred from + // `SoftwareConfig.image_version` or the system default. string image_uri = 3 [(google.api.field_behavior) = OPTIONAL]; // Optional. The Compute Engine machine type used for cluster instances. @@ -369,7 +412,7 @@ message InstanceGroupConfig { // // **Auto Zone Exception**: If you are using the Dataproc // [Auto Zone - // Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + // Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) // feature, you must use the short name of the machine type // resource, for example, `n1-standard-2`. string machine_type_uri = 4 [(google.api.field_behavior) = OPTIONAL]; @@ -377,9 +420,9 @@ message InstanceGroupConfig { // Optional. Disk option config settings. DiskConfig disk_config = 5 [(google.api.field_behavior) = OPTIONAL]; - // Optional. Specifies that this instance group contains preemptible + // Output only. Specifies that this instance group contains preemptible // instances. - bool is_preemptible = 6 [(google.api.field_behavior) = OPTIONAL]; + bool is_preemptible = 6 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The config for Compute Engine Instance Group // Manager that manages this group. @@ -391,8 +434,8 @@ message InstanceGroupConfig { repeated AcceleratorConfig accelerators = 8 [(google.api.field_behavior) = OPTIONAL]; // Specifies the minimum cpu platform for the Instance Group. - // See [Dataproc→Minimum CPU Platform] - // (/dataproc/docs/concepts/compute/dataproc-min-cpu). + // See [Dataproc -> Minimum CPU + // Platform](https://cloud.google.com/dataproc/docs/concepts/compute/dataproc-min-cpu). string min_cpu_platform = 9; } @@ -407,12 +450,13 @@ message ManagedGroupConfig { } // Specifies the type and number of accelerator cards attached to the instances -// of an instance group (see [GPUs on Compute Engine](/compute/docs/gpus/)). +// of an instance group (see [GPUs on Compute +// Engine](https://cloud.google.com/compute/docs/gpus/)). message AcceleratorConfig { // Full URL, partial URI, or short name of the accelerator type resource to // expose to this instance. See // [Compute Engine - // AcceleratorTypes](/compute/docs/reference/beta/acceleratorTypes) + // AcceleratorTypes](https://cloud.google.com/compute/docs/reference/beta/acceleratorTypes) // // Examples // * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` @@ -421,7 +465,7 @@ message AcceleratorConfig { // // **Auto Zone Exception**: If you are using the Dataproc // [Auto Zone - // Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + // Placement](https://cloud.google.com/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) // feature, you must use the short name of the accelerator type // resource, for example, `nvidia-tesla-k80`. string accelerator_type_uri = 1; @@ -487,7 +531,8 @@ message SecurityConfig { // Specifies Kerberos related configuration. message KerberosConfig { - // Optional. Flag to indicate whether to Kerberize the cluster. + // Optional. Flag to indicate whether to Kerberize the cluster (default: false). Set + // this field to true to enable Kerberos on a cluster. bool enable_kerberos = 1 [(google.api.field_behavior) = OPTIONAL]; // Required. The Cloud Storage URI of a KMS encrypted file containing the root @@ -591,6 +636,15 @@ message ClusterStatus { // The cluster is being updated. It continues to accept and process jobs. UPDATING = 5; + + // The cluster is being stopped. It cannot be used. + STOPPING = 6; + + // The cluster is currently stopped. It is not ready for use. + STOPPED = 7; + + // The cluster is being started. It is not ready for use. + STARTING = 8; } // The cluster substate. @@ -631,10 +685,10 @@ message ClusterStatus { message SoftwareConfig { // Optional. The version of software inside the cluster. It must be one of the // supported [Dataproc - // Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions), + // Versions](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions), // such as "1.2" (including a subminor version, such as "1.2.29"), or the // ["preview" - // version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). + // version](https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). // If unspecified, it defaults to the latest Debian version. string image_version = 1 [(google.api.field_behavior) = OPTIONAL]; @@ -654,8 +708,8 @@ message SoftwareConfig { // * spark: `spark-defaults.conf` // * yarn: `yarn-site.xml` // - // For more information, see - // [Cluster properties](/dataproc/docs/concepts/cluster-properties). + // For more information, see [Cluster + // properties](https://cloud.google.com/dataproc/docs/concepts/cluster-properties). map properties = 2 [(google.api.field_behavior) = OPTIONAL]; // The set of optional components to activate on the cluster. diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto index 3208822f..325dc9e5 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto +++ b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/jobs.proto @@ -43,6 +43,19 @@ service JobController { option (google.api.method_signature) = "project_id, region, job"; } + // Submits job to a cluster. + rpc SubmitJobAsOperation(SubmitJobRequest) returns (google.longrunning.Operation) { + option (google.api.http) = { + post: "/v1beta2/projects/{project_id}/regions/{region}/jobs:submitAsOperation" + body: "*" + }; + option (google.api.method_signature) = "project_id, region, job"; + option (google.longrunning.operation_info) = { + response_type: "Job" + metadata_type: "JobMetadata" + }; + } + // Gets the resource representation for a job in a project. rpc GetJob(GetJobRequest) returns (Job) { option (google.api.http) = { @@ -420,6 +433,42 @@ message SparkRJob { LoggingConfig logging_config = 6 [(google.api.field_behavior) = OPTIONAL]; } +// A Dataproc job for running [Presto](https://prestosql.io/) queries. +// **IMPORTANT**: The [Dataproc Presto Optional +// Component](/dataproc/docs/concepts/components/presto) must be enabled when +// the cluster is created to submit a Presto job to the cluster. +message PrestoJob { + // Required. The sequence of Presto queries to execute, specified as + // either an HCFS file URI or as a list of queries. + oneof queries { + // The HCFS URI of the script that contains SQL queries. + string query_file_uri = 1; + + // A list of queries. + QueryList query_list = 2; + } + + // Optional. Whether to continue executing queries if a query fails. + // The default value is `false`. Setting to `true` can be useful when + // executing independent parallel queries. + bool continue_on_failure = 3 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. The format in which query output will be displayed. See the + // Presto documentation for supported output formats + string output_format = 4 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. Presto client tags to attach to this query + repeated string client_tags = 5 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. A mapping of property names to values. Used to set Presto + // [session properties](https://prestodb.io/docs/current/sql/set-session.html) + // Equivalent to using the --session flag in the Presto CLI + map properties = 6 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. The runtime log config for job execution. + LoggingConfig logging_config = 7 [(google.api.field_behavior) = OPTIONAL]; +} + // Dataproc job config. message JobPlacement { // Required. The name of the cluster where the job will be submitted. @@ -591,19 +640,29 @@ message Job { // Required. The application/framework-specific portion of the job. oneof type_job { - HadoopJob hadoop_job = 3; + // Optional. Job is a Hadoop job. + HadoopJob hadoop_job = 3 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. Job is a Spark job. + SparkJob spark_job = 4 [(google.api.field_behavior) = OPTIONAL]; - SparkJob spark_job = 4; + // Optional. Job is a PySpark job. + PySparkJob pyspark_job = 5 [(google.api.field_behavior) = OPTIONAL]; - PySparkJob pyspark_job = 5; + // Optional. Job is a Hive job. + HiveJob hive_job = 6 [(google.api.field_behavior) = OPTIONAL]; - HiveJob hive_job = 6; + // Optional. Job is a Pig job. + PigJob pig_job = 7 [(google.api.field_behavior) = OPTIONAL]; - PigJob pig_job = 7; + // Optional. Job is a SparkR job. + SparkRJob spark_r_job = 21 [(google.api.field_behavior) = OPTIONAL]; - SparkRJob spark_r_job = 21; + // Optional. Job is a SparkSql job. + SparkSqlJob spark_sql_job = 12 [(google.api.field_behavior) = OPTIONAL]; - SparkSqlJob spark_sql_job = 12; + // Optional. Job is a Presto job. + PrestoJob presto_job = 23 [(google.api.field_behavior) = OPTIONAL]; } // Output only. The job status. Additional application-specific @@ -649,6 +708,12 @@ message Job { // over time. This is in contrast to a user-settable reference.job_id that // may be reused over time. string job_uuid = 22 [(google.api.field_behavior) = OUTPUT_ONLY]; + + // Output only. Indicates whether the job is completed. If the value is `false`, + // the job is still in progress. If `true`, the job is completed, and + // `status.state` field will indicate if it was successful, failed, + // or cancelled. + bool done = 24 [(google.api.field_behavior) = OUTPUT_ONLY]; } // Job scheduling options. @@ -664,6 +729,21 @@ message JobScheduling { int32 max_failures_per_hour = 1 [(google.api.field_behavior) = OPTIONAL]; } +// Job Operation metadata. +message JobMetadata { + // Output only. The job id. + string job_id = 1 [(google.api.field_behavior) = OUTPUT_ONLY]; + + // Output only. Most recent job status. + JobStatus status = 2 [(google.api.field_behavior) = OUTPUT_ONLY]; + + // Output only. Operation type. + string operation_type = 3 [(google.api.field_behavior) = OUTPUT_ONLY]; + + // Output only. Job submission time. + google.protobuf.Timestamp start_time = 4 [(google.api.field_behavior) = OUTPUT_ONLY]; +} + // A request to submit a job. message SubmitJobRequest { // Required. The ID of the Google Cloud Platform project that the job diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/shared.proto b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/shared.proto index eba80918..130ae554 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/shared.proto +++ b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/shared.proto @@ -31,9 +31,15 @@ enum Component { // The Anaconda python distribution. ANACONDA = 5; + // Docker + DOCKER = 13; + // The Druid query engine. DRUID = 9; + // Flink + FLINK = 14; + // The Hive Web HCatalog (the REST service for accessing HCatalog). HIVE_WEBHCAT = 3; @@ -46,6 +52,12 @@ enum Component { // The Presto query engine. PRESTO = 6; + // The Ranger service. + RANGER = 12; + + // The Solr service. + SOLR = 10; + // The Zeppelin notebook. ZEPPELIN = 4; diff --git a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto index a46282d9..e5ef680b 100644 --- a/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto +++ b/proto-google-cloud-dataproc-v1beta2/src/main/proto/google/cloud/dataproc/v1beta2/workflow_templates.proto @@ -35,12 +35,10 @@ option java_package = "com.google.cloud.dataproc.v1beta2"; // Dataproc API. service WorkflowTemplateService { option (google.api.default_host) = "dataproc.googleapis.com"; - option (google.api.oauth_scopes) = - "https://www.googleapis.com/auth/cloud-platform"; + option (google.api.oauth_scopes) = "https://www.googleapis.com/auth/cloud-platform"; // Creates new workflow template. - rpc CreateWorkflowTemplate(CreateWorkflowTemplateRequest) - returns (WorkflowTemplate) { + rpc CreateWorkflowTemplate(CreateWorkflowTemplateRequest) returns (WorkflowTemplate) { option (google.api.http) = { post: "/v1beta2/{parent=projects/*/regions/*}/workflowTemplates" body: "template" @@ -56,8 +54,7 @@ service WorkflowTemplateService { // // Can retrieve previously instantiated template by specifying optional // version parameter. - rpc GetWorkflowTemplate(GetWorkflowTemplateRequest) - returns (WorkflowTemplate) { + rpc GetWorkflowTemplate(GetWorkflowTemplateRequest) returns (WorkflowTemplate) { option (google.api.http) = { get: "/v1beta2/{name=projects/*/regions/*/workflowTemplates/*}" additional_bindings { @@ -87,8 +84,7 @@ service WorkflowTemplateService { // On successful completion, // [Operation.response][google.longrunning.Operation.response] will be // [Empty][google.protobuf.Empty]. - rpc InstantiateWorkflowTemplate(InstantiateWorkflowTemplateRequest) - returns (google.longrunning.Operation) { + rpc InstantiateWorkflowTemplate(InstantiateWorkflowTemplateRequest) returns (google.longrunning.Operation) { option (google.api.http) = { post: "/v1beta2/{name=projects/*/regions/*/workflowTemplates/*}:instantiate" body: "*" @@ -108,8 +104,7 @@ service WorkflowTemplateService { // Instantiates a template and begins execution. // // This method is equivalent to executing the sequence - // [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate], - // [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate], + // [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate], // [DeleteWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.DeleteWorkflowTemplate]. // // The returned Operation can be used to track execution of @@ -130,9 +125,7 @@ service WorkflowTemplateService { // On successful completion, // [Operation.response][google.longrunning.Operation.response] will be // [Empty][google.protobuf.Empty]. - rpc InstantiateInlineWorkflowTemplate( - InstantiateInlineWorkflowTemplateRequest) - returns (google.longrunning.Operation) { + rpc InstantiateInlineWorkflowTemplate(InstantiateInlineWorkflowTemplateRequest) returns (google.longrunning.Operation) { option (google.api.http) = { post: "/v1beta2/{parent=projects/*/locations/*}/workflowTemplates:instantiateInline" body: "template" @@ -150,8 +143,7 @@ service WorkflowTemplateService { // Updates (replaces) workflow template. The updated template // must contain version that matches the current server version. - rpc UpdateWorkflowTemplate(UpdateWorkflowTemplateRequest) - returns (WorkflowTemplate) { + rpc UpdateWorkflowTemplate(UpdateWorkflowTemplateRequest) returns (WorkflowTemplate) { option (google.api.http) = { put: "/v1beta2/{template.name=projects/*/regions/*/workflowTemplates/*}" body: "template" @@ -164,8 +156,7 @@ service WorkflowTemplateService { } // Lists workflows that match the specified filter in the request. - rpc ListWorkflowTemplates(ListWorkflowTemplatesRequest) - returns (ListWorkflowTemplatesResponse) { + rpc ListWorkflowTemplates(ListWorkflowTemplatesRequest) returns (ListWorkflowTemplatesResponse) { option (google.api.http) = { get: "/v1beta2/{parent=projects/*/regions/*}/workflowTemplates" additional_bindings { @@ -176,8 +167,7 @@ service WorkflowTemplateService { } // Deletes a workflow template. It does not cancel in-progress workflows. - rpc DeleteWorkflowTemplate(DeleteWorkflowTemplateRequest) - returns (google.protobuf.Empty) { + rpc DeleteWorkflowTemplate(DeleteWorkflowTemplateRequest) returns (google.protobuf.Empty) { option (google.api.http) = { delete: "/v1beta2/{name=projects/*/regions/*/workflowTemplates/*}" additional_bindings { @@ -230,12 +220,10 @@ message WorkflowTemplate { int32 version = 3 [(google.api.field_behavior) = OPTIONAL]; // Output only. The time template was created. - google.protobuf.Timestamp create_time = 4 - [(google.api.field_behavior) = OUTPUT_ONLY]; + google.protobuf.Timestamp create_time = 4 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The time template was last updated. - google.protobuf.Timestamp update_time = 5 - [(google.api.field_behavior) = OUTPUT_ONLY]; + google.protobuf.Timestamp update_time = 5 [(google.api.field_behavior) = OUTPUT_ONLY]; // Optional. The labels to associate with this template. These labels // will be propagated to all jobs and clusters created by the workflow @@ -260,8 +248,7 @@ message WorkflowTemplate { // Optional. Template parameters whose values are substituted into the // template. Values for parameters must be provided when the template is // instantiated. - repeated TemplateParameter parameters = 9 - [(google.api.field_behavior) = OPTIONAL]; + repeated TemplateParameter parameters = 9 [(google.api.field_behavior) = OPTIONAL]; } // Specifies workflow execution target. @@ -329,13 +316,13 @@ message OrderedJob { // // The step id is used as prefix for job id, as job // `goog-dataproc-workflow-step-id` label, and in - // [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] - // field from other steps. + // [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + // steps. // // The id must contain only letters (a-z, A-Z), numbers (0-9), // underscores (_), and hyphens (-). Cannot begin or end with underscore // or hyphen. Must consist of between 3 and 50 characters. - string step_id = 1; + string step_id = 1 [(google.api.field_behavior) = REQUIRED]; // Required. The job definition. oneof job_type { @@ -349,7 +336,13 @@ message OrderedJob { PigJob pig_job = 6; + // Spark R job + SparkRJob spark_r_job = 11; + SparkSqlJob spark_sql_job = 7; + + // Presto job + PrestoJob presto_job = 12; } // Optional. The labels to associate with this job. @@ -362,14 +355,14 @@ message OrderedJob { // the following regular expression: [\p{Ll}\p{Lo}\p{N}_-]{0,63} // // No more than 32 labels can be associated with a given job. - map labels = 8; + map labels = 8 [(google.api.field_behavior) = OPTIONAL]; // Optional. Job scheduling configuration. - JobScheduling scheduling = 9; + JobScheduling scheduling = 9 [(google.api.field_behavior) = OPTIONAL]; // Optional. The optional list of prerequisite job step_ids. // If not specified, the job will start at the beginning of workflow. - repeated string prerequisite_step_ids = 10; + repeated string prerequisite_step_ids = 10 [(google.api.field_behavior) = OPTIONAL]; } // A configurable parameter that replaces one or more fields in the template. @@ -395,10 +388,10 @@ message TemplateParameter { // A field is allowed to appear in at most one parameter's list of field // paths. // - // A field path is similar in syntax to a - // [google.protobuf.FieldMask][google.protobuf.FieldMask]. For example, a - // field path that references the zone field of a workflow template's cluster - // selector would be specified as `placement.clusterSelector.zone`. + // A field path is similar in syntax to a [google.protobuf.FieldMask][google.protobuf.FieldMask]. + // For example, a field path that references the zone field of a workflow + // template's cluster selector would be specified as + // `placement.clusterSelector.zone`. // // Also, field paths can reference fields using the following syntax: // @@ -505,15 +498,13 @@ message WorkflowMetadata { int32 version = 2 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The create cluster operation metadata. - ClusterOperation create_cluster = 3 - [(google.api.field_behavior) = OUTPUT_ONLY]; + ClusterOperation create_cluster = 3 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The workflow graph. WorkflowGraph graph = 4 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The delete cluster operation metadata. - ClusterOperation delete_cluster = 5 - [(google.api.field_behavior) = OUTPUT_ONLY]; + ClusterOperation delete_cluster = 5 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The workflow state. State state = 6 [(google.api.field_behavior) = OUTPUT_ONLY]; @@ -525,12 +516,10 @@ message WorkflowMetadata { map parameters = 8; // Output only. Workflow start time. - google.protobuf.Timestamp start_time = 9 - [(google.api.field_behavior) = OUTPUT_ONLY]; + google.protobuf.Timestamp start_time = 9 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. Workflow end time. - google.protobuf.Timestamp end_time = 10 - [(google.api.field_behavior) = OUTPUT_ONLY]; + google.protobuf.Timestamp end_time = 10 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The UUID of target cluster. string cluster_uuid = 11 [(google.api.field_behavior) = OUTPUT_ONLY]; @@ -582,8 +571,7 @@ message WorkflowNode { string step_id = 1 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. Node's prerequisite nodes. - repeated string prerequisite_step_ids = 2 - [(google.api.field_behavior) = OUTPUT_ONLY]; + repeated string prerequisite_step_ids = 2 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. The job id; populated after the node enters RUNNING state. string job_id = 3 [(google.api.field_behavior) = OUTPUT_ONLY]; @@ -765,8 +753,7 @@ message ListWorkflowTemplatesRequest { // A response to a request to list workflow templates in a project. message ListWorkflowTemplatesResponse { // Output only. WorkflowTemplates list. - repeated WorkflowTemplate templates = 1 - [(google.api.field_behavior) = OUTPUT_ONLY]; + repeated WorkflowTemplate templates = 1 [(google.api.field_behavior) = OUTPUT_ONLY]; // Output only. This token is included in the response if there are more // results to fetch. To fetch additional results, provide this value as the diff --git a/synth.metadata b/synth.metadata index 4e03e6ed..6f38e857 100644 --- a/synth.metadata +++ b/synth.metadata @@ -12,16 +12,17 @@ "git": { "name": "googleapis", "remote": "https://github.com/googleapis/googleapis.git", - "sha": "aba342359b6743353195ca53f944fe71e6fb6cd4", - "internalRef": "304207684" + "sha": "744591190e828440f72745aef217f883afd1fd71", + "internalRef": "305323909" } }, { "git": { "name": "googleapis", "remote": "https://github.com/googleapis/googleapis.git", - "sha": "aba342359b6743353195ca53f944fe71e6fb6cd4", - "internalRef": "304207684" + "sha": "744591190e828440f72745aef217f883afd1fd71", + "internalRef": "305323909", + "log": "744591190e828440f72745aef217f883afd1fd71\ndocs: change relative URLs to absolute URLs to fix broken links.\n\nPiperOrigin-RevId: 305323909\n\n1247c135ceaedfe04261d27a64aaecf78ffbae74\nchore: enable gapicv2 for videointelligence/v1beta2 API\n\nCommitter: @miraleung\nPiperOrigin-RevId: 305321976\n\n633c8b13227b9e3810749964d580e5be504db488\nchore: enable gapicv2 for videointelligence/v1p1beta1 API\n\nCommitter: @miraleung\nPiperOrigin-RevId: 305320877\n\n29aac60f121dc43382b37ff92f2dbb692d94143a\ndocs: fix broken link to policy reference documentation.\n\nPiperOrigin-RevId: 305319540\n\n54ddbbf14c489b8a2f0731aa39408c016f5a8387\nbazel: update gapic-generator-go to v0.13.0\n\nChanges include:\n* add clientHook feature\n\nPiperOrigin-RevId: 305289945\n\n823facb4ca6a4b36b817ce955a790dcb40cf808f\nchore: enable gapicv2 for videointelligence/v1p3beta1\n\nCommitter: @miraleung\nPiperOrigin-RevId: 305155214\n\n6b9c969d42bcb0f8206675bd868ed7d1ddcdaef9\nAdd API for bigqueryreservation v1.\n\nPiperOrigin-RevId: 305151484\n\n514f7d27811832a9f58b83d6f6305d894b097cf6\nchore: enable gapicv2 for phishingprotection/v1beta1 API\n\nCommitter: @miraleung\nPiperOrigin-RevId: 305126983\n\nff74d47d47280e6bbcbad1a7c82b1e0959c472ec\nfix: PHP-related fixes in BUILD.bazel and service.yamls\n\nThis PR also adds the rules for all 7 langauges in OsLogin and Kms BUILD.bazel files. Those build files were missing rules for 5 langagues, including PHP.\n\nThis PR is the prerequisite for migrating PHP synth.py scripts from artman to bazel.\n\nThe fixes in service.yaml fix regression made during proto annotation migration. This became visible only during PHP generation, because only PHP depends on the affected sections of the service.yaml config.\n\nPiperOrigin-RevId: 305108224\n\nfdbc7b1f63969307c71143a0c24fdfd02e739df6\nEnable gapicv2 for osconfig/agentendpoint/v1\n\nCommitter: @miraleung\nPiperOrigin-RevId: 305086443\n\n1490d30e1ae339570dd7826ba625a603ede91a08\nEnable gapicv2 for osconfig/v1beta\n\nCommitter: @miraleung\nPiperOrigin-RevId: 305069755\n\n7bf824e82e5c3549642b150dc4a9579602000f34\nEnable gapicv2 for iam/credentials/v1\n\nCommitter: @miraleung\nPiperOrigin-RevId: 305065283\n\n9ff6fd3b22f99167827e89aae7778408b5e82425\nUpdates Dataproc v1 API:\n- Adds Dataproc Jobs.SubmitJobAsOperation RPC\n- Adds SparkR and Presto job types to WorkflowTemplates\n- Adds new Optional Components\n- Clarifies usage of some APIs\n\nPiperOrigin-RevId: 305053617\n\ncad0f5137a70d0d14a8d9acbfcee98e4cd3e9662\nUpdates to Dataproc v1beta2 API:\n- Adds SparkR and Presto job types to WorkflowTemplates\n- Adds new Optional Components\n- Clarifies usage of some APIs\n\nPiperOrigin-RevId: 305053062\n\na005f045a301535eeb4c4b3fa7bb94eec9d22a8b\nAdd support for Cloud EKM to the Cloud KMS service and resource protos.\n\nPiperOrigin-RevId: 305026790\n\n5077b1e4674afdbbf11dac3f5f43d36285ba53ff\nchore: set Ruby namespace in proto options\n\nPiperOrigin-RevId: 304836531\n\nd6cb4997910eda04c0c66c0f2fd043eeaa0f660d\nchore: enable gapic v2 and proto annotation for documentai API.\n\ncommitter @summer-ji-eng\n\nPiperOrigin-RevId: 304724866\n\n490bc556608bfa5b1548c9374b06152fa33d657e\nEnable gapicv2 for devtools/remoteworkers/v1test2\n\nCommitter: @miraleung\nPiperOrigin-RevId: 304718691\n\n9f78ce31a5bd7f4a63e3cf0ddf28221557adb7ed\nEnable gapicv2 for managedidentities/v1beta1\n\nCommitter: @miraleung\nPiperOrigin-RevId: 304718676\n\n6e17d259b8e320bc51aa240cefef05ec753e2b83\ndocs: treat a dummy example URL as a string literal instead of a link\n\nPiperOrigin-RevId: 304716376\n\na8d76f99d3073aaccabdcc122c798a63e812c4fe\ndocs: change relative URLs to absolute URLs to fix broken links.\n\nPiperOrigin-RevId: 304702368\n\n65c749bc6a1d240416a0e6979381b67f97aff907\ndocs: fix formatting of some regexes and string literals.\n\nPiperOrigin-RevId: 304701150\n\n9119eefcd2b5ce845a680fa4ec4093ed733498f0\nchore: set Ruby namespace in proto options\n\nPiperOrigin-RevId: 304698702\n\n62a2a7cc33d3535638d220df238823eefcca930d\nchore: set Ruby namespace in proto options\n\nPiperOrigin-RevId: 304696461\n\n23848c8f64a5e81a239d6133378468185f1756dc\nchore: set Ruby namespace in proto options\n\nPiperOrigin-RevId: 304696192\n\n9514fa9e390a4c0715972c5b510cf4c10ad049a1\ndocs: change relative URLs to absolute URLs to fix broken links.\n\nPiperOrigin-RevId: 304695334\n\n0f7b1509a9a452808c3d07fe90fedfcea763d7d5\nfix: change config_schema_version to 2.0.0 for containeranalysis v1 gapic config.\n\ncommitter: @hzyi-google\nPiperOrigin-RevId: 304672648\n\n3d52f3c126fbfc31f067a7f54737b7f0dfbce163\nDialogflow weekly v2 library update:\n- Change `parent` field's resource_reference to specify child_type instead of type per client library generation requirement;\n- Change Session with its child resource pattern to support both projects/{project}/agent/sessions/{session} and projects/{project}/agent/environments/{environment}/users/{user}/sessions/{session};\n- Fix `method_signature`\n- Regular documentation update\n\nImportant updates are also posted at:\nhttps://cloud.google.com/dialogflow/docs/release-notes\n\nPiperOrigin-RevId: 304635286\n\n4a6a01ce0ead505c245d11a2ce156de34800c58f\ndocs: change a relative URL to an absolute URL to fix broken links.\n\nPiperOrigin-RevId: 304633630\n\n1b969c28a6579265e89cd35e6c2ecacc89970e2d\nchore: set Ruby namespace in proto options\n\nPiperOrigin-RevId: 304620317\n\n5378173a889f9c7d83e36e52d38a6267190de692\nAdd v1beta2 SubmitJobAsOperation RPC to Dataproc.\n\nPiperOrigin-RevId: 304594381\n\n3d5d228a58bdf875e6147b228db3159010c735ee\nEnable visibility on the GetSnapshot and GetTopicSnapshots methods\n\nPiperOrigin-RevId: 304500143\n\n1346f5c11a0376bc82c5c03259b304de52671518\nRefresh public client library for Cloud Monitoring.\nIncrease default ListTimeSeries deadline to 90s.\n\nPiperOrigin-RevId: 304467075\n\n4a6cfccb4a32cb28db7366295d90078c6af3112f\ndocs: Fix link in monitoring summary description.\n\nPiperOrigin-RevId: 304438717\n\n6e2ea9a4b63a264090f1bd4b5b25766a3f634f3a\nfix: add interface name overrides back to logging gapic config.\n\ncommitter: @hzyi-google\nPiperOrigin-RevId: 304438432\n\n01f3ccbaa66cf3ae4e3b9fd140b1ecfbe54a3ed0\nFix: Add package/namespace options for C#, Ruby and PHP in remaining protos.\n\nPiperOrigin-RevId: 304416658\n\nb89abbf12cd3c086abddcd79adb0a653349f960a\nchore(google/maps): Rename nox.py to noxfile.py\n\nupstream wip fix here: https://github.com/googleapis/gapic-generator/pull/3156\n\nPiperOrigin-RevId: 304411146\n\nbbf26d7f11fae7789b13959bf100983c496c9807\nSynchronize new proto/yaml changes.\n\nPiperOrigin-RevId: 304409906\n\n75047719f704d61f405cac6d7439637ab36c1232\nfix Dataproc: add missing `REQUIRED` annotation.\n\nPiperOrigin-RevId: 304305096\n\n7e1b880fd5d477176cf9b0bb9b137b9bea56e787\nAdd Ruby/PHP namespace options\n\nPiperOrigin-RevId: 304288402\n\n" } }, {