diff --git a/google-cloud-bigquerystorage/clirr-ignored-differences.xml b/google-cloud-bigquerystorage/clirr-ignored-differences.xml
deleted file mode 100644
index 309241d8a9..0000000000
--- a/google-cloud-bigquerystorage/clirr-ignored-differences.xml
+++ /dev/null
@@ -1,46 +0,0 @@
-
-
-
This class provides the ability to make remote calls to the backing service through method * calls that map to API methods. Sample code to get started: * - *
- *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ProjectName parent = ProjectName.of("[PROJECT]");
- * ReadSession readSession = ReadSession.newBuilder().build();
- * int maxStreamCount = 0;
- * ReadSession response = baseBigQueryReadClient.createReadSession(parent, readSession, maxStreamCount);
- * }
- *
- *
- *
- * Note: close() needs to be called on the baseBigQueryReadClient object to clean up resources + *
Note: close() needs to be called on the BaseBigQueryReadClient object to clean up resources * such as threads. In the example above, try-with-resources is used, which automatically calls * close(). * @@ -74,30 +64,28 @@ * *
To customize credentials: * - *
- *
+ * {@code
* BaseBigQueryReadSettings baseBigQueryReadSettings =
* BaseBigQueryReadSettings.newBuilder()
* .setCredentialsProvider(FixedCredentialsProvider.create(myCredentials))
* .build();
* BaseBigQueryReadClient baseBigQueryReadClient =
* BaseBigQueryReadClient.create(baseBigQueryReadSettings);
- *
- *
+ * }
*
- * To customize the endpoint:
+ * To customize the endpoint: * - *
- *
+ * {@code
* BaseBigQueryReadSettings baseBigQueryReadSettings =
* BaseBigQueryReadSettings.newBuilder().setEndpoint(myEndpoint).build();
* BaseBigQueryReadClient baseBigQueryReadClient =
* BaseBigQueryReadClient.create(baseBigQueryReadSettings);
- *
- *
+ * }
+ *
+ * Please refer to the GitHub repository's samples for more quickstart code snippets. */ -@Generated("by gapic-generator") @BetaApi +@Generated("by gapic-generator") public class BaseBigQueryReadClient implements BackgroundResource { private final BaseBigQueryReadSettings settings; private final BigQueryReadStub stub; @@ -118,7 +106,7 @@ public static final BaseBigQueryReadClient create(BaseBigQueryReadSettings setti /** * Constructs an instance of BaseBigQueryReadClient, using the given stub for making calls. This - * is for advanced usage - prefer to use BaseBigQueryReadSettings}. + * is for advanced usage - prefer using create(BaseBigQueryReadSettings). */ @BetaApi("A restructuring of stub classes is planned, so this may break in the future") public static final BaseBigQueryReadClient create(BigQueryReadStub stub) { @@ -150,7 +138,7 @@ public BigQueryReadStub getStub() { return stub; } - // AUTO-GENERATED DOCUMENTATION AND METHOD + // AUTO-GENERATED DOCUMENTATION AND METHOD. /** * Creates a new read session. A read session divides the contents of a BigQuery table into one or * more streams, which can then be used to read data from the table. The read session also @@ -169,17 +157,6 @@ public BigQueryReadStub getStub() { *
Read sessions automatically expire 24 hours after they are created and do not require manual * clean-up by the caller. * - *
Sample code: - * - *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ProjectName parent = ProjectName.of("[PROJECT]");
- * ReadSession readSession = ReadSession.newBuilder().build();
- * int maxStreamCount = 0;
- * ReadSession response = baseBigQueryReadClient.createReadSession(parent, readSession, maxStreamCount);
- * }
- *
- *
* @param parent Required. The request project that owns the session, in the form of
* `projects/{project_id}`.
* @param readSession Required. Session to be created.
@@ -202,7 +179,7 @@ public final ReadSession createReadSession(
return createReadSession(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Creates a new read session. A read session divides the contents of a BigQuery table into one or
* more streams, which can then be used to read data from the table. The read session also
@@ -221,17 +198,6 @@ public final ReadSession createReadSession(
* Read sessions automatically expire 24 hours after they are created and do not require manual * clean-up by the caller. * - *
Sample code: - * - *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ProjectName parent = ProjectName.of("[PROJECT]");
- * ReadSession readSession = ReadSession.newBuilder().build();
- * int maxStreamCount = 0;
- * ReadSession response = baseBigQueryReadClient.createReadSession(parent.toString(), readSession, maxStreamCount);
- * }
- *
- *
* @param parent Required. The request project that owns the session, in the form of
* `projects/{project_id}`.
* @param readSession Required. Session to be created.
@@ -254,7 +220,7 @@ public final ReadSession createReadSession(
return createReadSession(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Creates a new read session. A read session divides the contents of a BigQuery table into one or
* more streams, which can then be used to read data from the table. The read session also
@@ -273,20 +239,6 @@ public final ReadSession createReadSession(
* Read sessions automatically expire 24 hours after they are created and do not require manual * clean-up by the caller. * - *
Sample code: - * - *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ProjectName parent = ProjectName.of("[PROJECT]");
- * ReadSession readSession = ReadSession.newBuilder().build();
- * CreateReadSessionRequest request = CreateReadSessionRequest.newBuilder()
- * .setParent(parent.toString())
- * .setReadSession(readSession)
- * .build();
- * ReadSession response = baseBigQueryReadClient.createReadSession(request);
- * }
- *
- *
* @param request The request object containing all of the parameters for the API call.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
@@ -294,7 +246,7 @@ public final ReadSession createReadSession(CreateReadSessionRequest request) {
return createReadSessionCallable().call(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Creates a new read session. A read session divides the contents of a BigQuery table into one or
* more streams, which can then be used to read data from the table. The read session also
@@ -314,26 +266,12 @@ public final ReadSession createReadSession(CreateReadSessionRequest request) {
* clean-up by the caller.
*
* Sample code: - * - *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ProjectName parent = ProjectName.of("[PROJECT]");
- * ReadSession readSession = ReadSession.newBuilder().build();
- * CreateReadSessionRequest request = CreateReadSessionRequest.newBuilder()
- * .setParent(parent.toString())
- * .setReadSession(readSession)
- * .build();
- * ApiFuture<ReadSession> future = baseBigQueryReadClient.createReadSessionCallable().futureCall(request);
- * // Do something
- * ReadSession response = future.get();
- * }
- *
*/
public final UnaryCallableSample code: - * - *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ReadStreamName readStream = ReadStreamName.of("[PROJECT]", "[LOCATION]", "[SESSION]", "[STREAM]");
- * ReadRowsRequest request = ReadRowsRequest.newBuilder()
- * .setReadStream(readStream.toString())
- * .build();
- *
- * ServerStream<ReadRowsResponse> stream = baseBigQueryReadClient.readRowsCallable().call(request);
- * for (ReadRowsResponse response : stream) {
- * // Do something when receive a response
- * }
- * }
- *
*/
public final ServerStreamingCallableSample code: - * - *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ReadStreamName name = ReadStreamName.of("[PROJECT]", "[LOCATION]", "[SESSION]", "[STREAM]");
- * SplitReadStreamRequest request = SplitReadStreamRequest.newBuilder()
- * .setName(name.toString())
- * .build();
- * SplitReadStreamResponse response = baseBigQueryReadClient.splitReadStream(request);
- * }
- *
- *
* @param request The request object containing all of the parameters for the API call.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
@@ -394,7 +306,7 @@ public final SplitReadStreamResponse splitReadStream(SplitReadStreamRequest requ
return splitReadStreamCallable().call(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Splits a given `ReadStream` into two `ReadStream` objects. These `ReadStream` objects are
* referred to as the primary and the residual streams of the split. The original `ReadStream` can
@@ -408,18 +320,6 @@ public final SplitReadStreamResponse splitReadStream(SplitReadStreamRequest requ
* to completion.
*
* Sample code: - * - *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ReadStreamName name = ReadStreamName.of("[PROJECT]", "[LOCATION]", "[SESSION]", "[STREAM]");
- * SplitReadStreamRequest request = SplitReadStreamRequest.newBuilder()
- * .setName(name.toString())
- * .build();
- * ApiFuture<SplitReadStreamResponse> future = baseBigQueryReadClient.splitReadStreamCallable().futureCall(request);
- * // Do something
- * SplitReadStreamResponse response = future.get();
- * }
- *
*/
public final UnaryCallableFor example, to set the total timeout of createReadSession to 30 seconds: * - *
- *
+ * {@code
* BaseBigQueryReadSettings.Builder baseBigQueryReadSettingsBuilder =
* BaseBigQueryReadSettings.newBuilder();
* baseBigQueryReadSettingsBuilder
* .createReadSessionSettings()
* .setRetrySettings(
- * baseBigQueryReadSettingsBuilder.createReadSessionSettings().getRetrySettings().toBuilder()
+ * baseBigQueryReadSettingsBuilder
+ * .createReadSessionSettings()
+ * .getRetrySettings()
+ * .toBuilder()
* .setTotalTimeout(Duration.ofSeconds(30))
* .build());
* BaseBigQueryReadSettings baseBigQueryReadSettings = baseBigQueryReadSettingsBuilder.build();
- *
- *
+ * }
*/
-@Generated("by gapic-generator")
-@BetaApi
+@Generated("by gapic-generator-java")
public class BaseBigQueryReadSettings extends ClientSettingsThe interfaces provided are listed below, along with usage samples. - * - *
====================== BaseBigQueryReadClient ====================== + *
======================= BigQueryReadClient ======================= * *
Service Description: BigQuery Read API. * *
The Read API can be used to read data from BigQuery. * - *
Sample for BaseBigQueryReadClient: - * - *
- *
- * try (BaseBigQueryReadClient baseBigQueryReadClient = BaseBigQueryReadClient.create()) {
- * ProjectName parent = ProjectName.of("[PROJECT]");
- * ReadSession readSession = ReadSession.newBuilder().build();
- * int maxStreamCount = 0;
- * ReadSession response = baseBigQueryReadClient.createReadSession(parent, readSession, maxStreamCount);
- * }
- *
- *
+ * Sample for BigQueryReadClient: */ -@Generated("by gapic-generator") +@Generated("by gapic-generator-java") package com.google.cloud.bigquery.storage.v1; import javax.annotation.Generated; diff --git a/google-cloud-bigquerystorage/src/main/java/com/google/cloud/bigquery/storage/v1/stub/BigQueryReadStub.java b/google-cloud-bigquerystorage/src/main/java/com/google/cloud/bigquery/storage/v1/stub/BigQueryReadStub.java index 01bff92268..85cb247aaf 100644 --- a/google-cloud-bigquerystorage/src/main/java/com/google/cloud/bigquery/storage/v1/stub/BigQueryReadStub.java +++ b/google-cloud-bigquerystorage/src/main/java/com/google/cloud/bigquery/storage/v1/stub/BigQueryReadStub.java @@ -5,7 +5,7 @@ * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * - * https://www.apache.org/licenses/LICENSE-2.0 + * https://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, @@ -13,9 +13,9 @@ * See the License for the specific language governing permissions and * limitations under the License. */ + package com.google.cloud.bigquery.storage.v1.stub; -import com.google.api.core.BetaApi; import com.google.api.gax.core.BackgroundResource; import com.google.api.gax.rpc.ServerStreamingCallable; import com.google.api.gax.rpc.UnaryCallable; @@ -27,14 +27,13 @@ import com.google.cloud.bigquery.storage.v1.SplitReadStreamResponse; import javax.annotation.Generated; -// AUTO-GENERATED DOCUMENTATION AND CLASS +// AUTO-GENERATED DOCUMENTATION AND CLASS. /** - * Base stub class for BigQuery Storage API. + * Base stub class for the BigQueryRead service API. * *
This class is for advanced usage and reflects the underlying API directly.
*/
@Generated("by gapic-generator")
-@BetaApi("A restructuring of stub classes is planned, so this may break in the future")
public abstract class BigQueryReadStub implements BackgroundResource {
public UnaryCallable For example, to set the total timeout of createReadSession to 30 seconds:
*
- * This class is for advanced usage.
*/
@Generated("by gapic-generator")
-@BetaApi("The surface for use by generated code is not stable yet and may change in the future.")
public class GrpcBigQueryReadCallableFactory implements GrpcStubCallableFactory {
+
@Override
public This class is for advanced usage and reflects the underlying API directly.
*/
-@Generated("by gapic-generator")
-@BetaApi("A restructuring of stub classes is planned, so this may break in the future")
+@Generated("by gapic-generator-java")
public class GrpcBigQueryReadStub extends BigQueryReadStub {
-
private static final MethodDescriptor This class provides the ability to make remote calls to the backing service through method
* calls that map to API methods. Sample code to get started:
*
- * Note: close() needs to be called on the bigQueryWriteClient object to clean up resources such
+ * Note: close() needs to be called on the BigQueryWriteClient object to clean up resources such
* as threads. In the example above, try-with-resources is used, which automatically calls close().
*
* The surface of this class includes several types of Java methods for each of the API's
@@ -83,30 +63,26 @@
*
* To customize credentials:
*
- * To customize the endpoint:
*
- * Please refer to the GitHub repository's samples for more quickstart code snippets.
*/
-@Generated("by gapic-generator")
@BetaApi
+@Generated("by gapic-generator")
public class BigQueryWriteClient implements BackgroundResource {
private final BigQueryWriteSettings settings;
private final BigQueryWriteStub stub;
@@ -127,7 +103,7 @@ public static final BigQueryWriteClient create(BigQueryWriteSettings settings)
/**
* Constructs an instance of BigQueryWriteClient, using the given stub for making calls. This is
- * for advanced usage - prefer to use BigQueryWriteSettings}.
+ * for advanced usage - prefer using create(BigQueryWriteSettings).
*/
@BetaApi("A restructuring of stub classes is planned, so this may break in the future")
public static final BigQueryWriteClient create(BigQueryWriteStub stub) {
@@ -159,109 +135,66 @@ public BigQueryWriteStub getStub() {
return stub;
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Creates a write stream to the given table.
*
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * Sample code:
- *
- * For example, to set the total timeout of createWriteStream to 30 seconds:
*
- * The interfaces provided are listed below, along with usage samples.
- *
- * =================== BigQueryWriteClient ===================
+ * ======================= BigQueryWriteClient =======================
*
* Service Description: BigQuery Write API.
*
* The Write API can be used to write data to BigQuery.
*
* Sample for BigQueryWriteClient:
- *
- * This class is for advanced usage and reflects the underlying API directly.
*/
@Generated("by gapic-generator")
-@BetaApi("A restructuring of stub classes is planned, so this may break in the future")
public abstract class BigQueryWriteStub implements BackgroundResource {
- public UnaryCallable For example, to set the total timeout of createWriteStream to 30 seconds:
*
- * This class is for advanced usage.
*/
@Generated("by gapic-generator")
-@BetaApi("The surface for use by generated code is not stable yet and may change in the future.")
public class GrpcBigQueryWriteCallableFactory implements GrpcStubCallableFactory {
+
@Override
public This class is for advanced usage and reflects the underlying API directly.
*/
-@Generated("by gapic-generator")
-@BetaApi("A restructuring of stub classes is planned, so this may break in the future")
+@Generated("by gapic-generator-java")
public class GrpcBigQueryWriteStub extends BigQueryWriteStub {
-
- private static final MethodDescriptor
- *
+ * BigQueryReadStubSettings bigQueryReadSettings = bigQueryReadSettingsBuilder.build();
+ * }
*/
-@Generated("by gapic-generator")
@BetaApi
+@Generated("by gapic-generator-java")
public class BigQueryReadStubSettings extends StubSettings
- * BigQueryReadStubSettings.Builder baseBigQueryReadSettingsBuilder =
+ *
- * {@code
+ * BigQueryReadStubSettings.Builder bigQueryReadSettingsBuilder =
* BigQueryReadStubSettings.newBuilder();
- * baseBigQueryReadSettingsBuilder
+ * bigQueryReadSettingsBuilder
* .createReadSessionSettings()
* .setRetrySettings(
- * baseBigQueryReadSettingsBuilder.createReadSessionSettings().getRetrySettings().toBuilder()
+ * bigQueryReadSettingsBuilder
+ * .createReadSessionSettings()
+ * .getRetrySettings()
+ * .toBuilder()
* .setTotalTimeout(Duration.ofSeconds(30))
* .build());
- * BigQueryReadStubSettings baseBigQueryReadSettings = baseBigQueryReadSettingsBuilder.build();
- *
- *
- *
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * WriteStream writeStream = WriteStream.newBuilder().build();
- * WriteStream response = bigQueryWriteClient.createWriteStream(parent, writeStream);
- * }
- *
- *
- *
+ * BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create(bigQueryWriteSettings);
+ * }
*
- * To customize the endpoint:
+ *
+ *
- * {@code
* BigQueryWriteSettings bigQueryWriteSettings =
* BigQueryWriteSettings.newBuilder()
* .setCredentialsProvider(FixedCredentialsProvider.create(myCredentials))
* .build();
- * BigQueryWriteClient bigQueryWriteClient =
- * BigQueryWriteClient.create(bigQueryWriteSettings);
- *
- *
+ * BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create(bigQueryWriteSettings);
+ * }
+ *
+ *
+ *
- * {@code
* BigQueryWriteSettings bigQueryWriteSettings =
* BigQueryWriteSettings.newBuilder().setEndpoint(myEndpoint).build();
- * BigQueryWriteClient bigQueryWriteClient =
- * BigQueryWriteClient.create(bigQueryWriteSettings);
- *
- *
* @param parent Required. Reference to the table to which the stream belongs, in the format of
* `projects/{project}/datasets/{dataset}/tables/{table}`.
* @param writeStream Required. Stream to be created.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final WriteStream createWriteStream(TableName parent, WriteStream writeStream) {
- CreateWriteStreamRequest request =
- CreateWriteStreamRequest.newBuilder()
+ public final Stream.WriteStream createWriteStream(
+ TableName parent, Stream.WriteStream writeStream) {
+ Storage.CreateWriteStreamRequest request =
+ Storage.CreateWriteStreamRequest.newBuilder()
.setParent(parent == null ? null : parent.toString())
.setWriteStream(writeStream)
.build();
return createWriteStream(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Creates a write stream to the given table.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * WriteStream writeStream = WriteStream.newBuilder().build();
- * WriteStream response = bigQueryWriteClient.createWriteStream(parent, writeStream);
- * }
- *
- *
* @param parent Required. Reference to the table to which the stream belongs, in the format of
* `projects/{project}/datasets/{dataset}/tables/{table}`.
* @param writeStream Required. Stream to be created.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final WriteStream createWriteStream(String parent, WriteStream writeStream) {
- CreateWriteStreamRequest request =
- CreateWriteStreamRequest.newBuilder().setParent(parent).setWriteStream(writeStream).build();
+ public final Stream.WriteStream createWriteStream(String parent, Stream.WriteStream writeStream) {
+ Storage.CreateWriteStreamRequest request =
+ Storage.CreateWriteStreamRequest.newBuilder()
+ .setParent(parent)
+ .setWriteStream(writeStream)
+ .build();
return createWriteStream(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Creates a write stream to the given table.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * WriteStream writeStream = WriteStream.newBuilder().build();
- * WriteStream response = bigQueryWriteClient.createWriteStream(parent.toString(), writeStream);
- * }
- *
- *
* @param request The request object containing all of the parameters for the API call.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final WriteStream createWriteStream(CreateWriteStreamRequest request) {
+ public final Stream.WriteStream createWriteStream(Storage.CreateWriteStreamRequest request) {
return createWriteStreamCallable().call(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Creates a write stream to the given table.
*
*
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * WriteStream writeStream = WriteStream.newBuilder().build();
- * CreateWriteStreamRequest request = CreateWriteStreamRequest.newBuilder()
- * .setParent(parent.toString())
- * .setWriteStream(writeStream)
- * .build();
- * WriteStream response = bigQueryWriteClient.createWriteStream(request);
- * }
- *
*/
- public final UnaryCallable
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * WriteStream writeStream = WriteStream.newBuilder().build();
- * CreateWriteStreamRequest request = CreateWriteStreamRequest.newBuilder()
- * .setParent(parent.toString())
- * .setWriteStream(writeStream)
- * .build();
- * ApiFuture<WriteStream> future = bigQueryWriteClient.createWriteStreamCallable().futureCall(request);
- * // Do something
- * WriteStream response = future.get();
- * }
- *
*/
- public final BidiStreamingCallable
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * BidiStream<AppendRowsRequest, AppendRowsResponse> bidiStream =
- * bigQueryWriteClient.appendRowsCallable().call();
- *
- * WriteStreamName writeStream = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * AppendRowsRequest request = AppendRowsRequest.newBuilder()
- * .setWriteStream(writeStream.toString())
- * .build();
- * bidiStream.send(request);
- * for (AppendRowsResponse response : bidiStream) {
- * // Do something when receive a response
- * }
- * }
- *
- *
* @param name Required. Name of the stream to get, in the form of
* `projects/{project}/datasets/{dataset}/tables/{table}/streams/{stream}`.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final WriteStream getWriteStream(WriteStreamName name) {
- GetWriteStreamRequest request =
- GetWriteStreamRequest.newBuilder().setName(name == null ? null : name.toString()).build();
+ public final Stream.WriteStream getWriteStream(WriteStreamName name) {
+ Storage.GetWriteStreamRequest request =
+ Storage.GetWriteStreamRequest.newBuilder()
+ .setName(name == null ? null : name.toString())
+ .build();
return getWriteStream(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Gets a write stream.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName name = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * WriteStream response = bigQueryWriteClient.getWriteStream(name);
- * }
- *
- *
* @param name Required. Name of the stream to get, in the form of
* `projects/{project}/datasets/{dataset}/tables/{table}/streams/{stream}`.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final WriteStream getWriteStream(String name) {
- GetWriteStreamRequest request = GetWriteStreamRequest.newBuilder().setName(name).build();
+ public final Stream.WriteStream getWriteStream(String name) {
+ Storage.GetWriteStreamRequest request =
+ Storage.GetWriteStreamRequest.newBuilder().setName(name).build();
return getWriteStream(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Gets a write stream.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName name = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * WriteStream response = bigQueryWriteClient.getWriteStream(name.toString());
- * }
- *
- *
* @param request The request object containing all of the parameters for the API call.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final WriteStream getWriteStream(GetWriteStreamRequest request) {
+ public final Stream.WriteStream getWriteStream(Storage.GetWriteStreamRequest request) {
return getWriteStreamCallable().call(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Gets a write stream.
*
*
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName name = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * GetWriteStreamRequest request = GetWriteStreamRequest.newBuilder()
- * .setName(name.toString())
- * .build();
- * WriteStream response = bigQueryWriteClient.getWriteStream(request);
- * }
- *
*/
- public final UnaryCallable
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName name = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * GetWriteStreamRequest request = GetWriteStreamRequest.newBuilder()
- * .setName(name.toString())
- * .build();
- * ApiFuture<WriteStream> future = bigQueryWriteClient.getWriteStreamCallable().futureCall(request);
- * // Do something
- * WriteStream response = future.get();
- * }
- *
- *
* @param name Required. Name of the stream to finalize, in the form of
* `projects/{project}/datasets/{dataset}/tables/{table}/streams/{stream}`.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final FinalizeWriteStreamResponse finalizeWriteStream(WriteStreamName name) {
- FinalizeWriteStreamRequest request =
- FinalizeWriteStreamRequest.newBuilder()
+ public final Storage.FinalizeWriteStreamResponse finalizeWriteStream(WriteStreamName name) {
+ Storage.FinalizeWriteStreamRequest request =
+ Storage.FinalizeWriteStreamRequest.newBuilder()
.setName(name == null ? null : name.toString())
.build();
return finalizeWriteStream(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Finalize a write stream so that no new data can be appended to the stream.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName name = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * FinalizeWriteStreamResponse response = bigQueryWriteClient.finalizeWriteStream(name);
- * }
- *
- *
* @param name Required. Name of the stream to finalize, in the form of
* `projects/{project}/datasets/{dataset}/tables/{table}/streams/{stream}`.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final FinalizeWriteStreamResponse finalizeWriteStream(String name) {
- FinalizeWriteStreamRequest request =
- FinalizeWriteStreamRequest.newBuilder().setName(name).build();
+ public final Storage.FinalizeWriteStreamResponse finalizeWriteStream(String name) {
+ Storage.FinalizeWriteStreamRequest request =
+ Storage.FinalizeWriteStreamRequest.newBuilder().setName(name).build();
return finalizeWriteStream(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Finalize a write stream so that no new data can be appended to the stream.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName name = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * FinalizeWriteStreamResponse response = bigQueryWriteClient.finalizeWriteStream(name.toString());
- * }
- *
- *
* @param request The request object containing all of the parameters for the API call.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final FinalizeWriteStreamResponse finalizeWriteStream(FinalizeWriteStreamRequest request) {
+ public final Storage.FinalizeWriteStreamResponse finalizeWriteStream(
+ Storage.FinalizeWriteStreamRequest request) {
return finalizeWriteStreamCallable().call(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Finalize a write stream so that no new data can be appended to the stream.
*
*
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName name = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * FinalizeWriteStreamRequest request = FinalizeWriteStreamRequest.newBuilder()
- * .setName(name.toString())
- * .build();
- * FinalizeWriteStreamResponse response = bigQueryWriteClient.finalizeWriteStream(request);
- * }
- *
*/
- public final UnaryCallable
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName name = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * FinalizeWriteStreamRequest request = FinalizeWriteStreamRequest.newBuilder()
- * .setName(name.toString())
- * .build();
- * ApiFuture<FinalizeWriteStreamResponse> future = bigQueryWriteClient.finalizeWriteStreamCallable().futureCall(request);
- * // Do something
- * FinalizeWriteStreamResponse response = future.get();
- * }
- *
- *
* @param parent Required. Parent table that all the streams should belong to, in the form of
* `projects/{project}/datasets/{dataset}/tables/{table}`.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final BatchCommitWriteStreamsResponse batchCommitWriteStreams(TableName parent) {
- BatchCommitWriteStreamsRequest request =
- BatchCommitWriteStreamsRequest.newBuilder()
+ public final Storage.BatchCommitWriteStreamsResponse batchCommitWriteStreams(TableName parent) {
+ Storage.BatchCommitWriteStreamsRequest request =
+ Storage.BatchCommitWriteStreamsRequest.newBuilder()
.setParent(parent == null ? null : parent.toString())
.build();
return batchCommitWriteStreams(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Atomically commits a group of `PENDING` streams that belong to the same `parent` table. Streams
* must be finalized before commit and cannot be committed multiple times. Once a stream is
* committed, data in the stream becomes available for read operations.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * BatchCommitWriteStreamsResponse response = bigQueryWriteClient.batchCommitWriteStreams(parent);
- * }
- *
- *
* @param parent Required. Parent table that all the streams should belong to, in the form of
* `projects/{project}/datasets/{dataset}/tables/{table}`.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final BatchCommitWriteStreamsResponse batchCommitWriteStreams(String parent) {
- BatchCommitWriteStreamsRequest request =
- BatchCommitWriteStreamsRequest.newBuilder().setParent(parent).build();
+ public final Storage.BatchCommitWriteStreamsResponse batchCommitWriteStreams(String parent) {
+ Storage.BatchCommitWriteStreamsRequest request =
+ Storage.BatchCommitWriteStreamsRequest.newBuilder().setParent(parent).build();
return batchCommitWriteStreams(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Atomically commits a group of `PENDING` streams that belong to the same `parent` table. Streams
* must be finalized before commit and cannot be committed multiple times. Once a stream is
* committed, data in the stream becomes available for read operations.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * BatchCommitWriteStreamsResponse response = bigQueryWriteClient.batchCommitWriteStreams(parent.toString());
- * }
- *
- *
* @param request The request object containing all of the parameters for the API call.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final BatchCommitWriteStreamsResponse batchCommitWriteStreams(
- BatchCommitWriteStreamsRequest request) {
+ public final Storage.BatchCommitWriteStreamsResponse batchCommitWriteStreams(
+ Storage.BatchCommitWriteStreamsRequest request) {
return batchCommitWriteStreamsCallable().call(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Atomically commits a group of `PENDING` streams that belong to the same `parent` table. Streams
* must be finalized before commit and cannot be committed multiple times. Once a stream is
* committed, data in the stream becomes available for read operations.
*
*
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * List<String> writeStreams = new ArrayList<>();
- * BatchCommitWriteStreamsRequest request = BatchCommitWriteStreamsRequest.newBuilder()
- * .setParent(parent.toString())
- * .addAllWriteStreams(writeStreams)
- * .build();
- * BatchCommitWriteStreamsResponse response = bigQueryWriteClient.batchCommitWriteStreams(request);
- * }
- *
*/
- public final UnaryCallable
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * List<String> writeStreams = new ArrayList<>();
- * BatchCommitWriteStreamsRequest request = BatchCommitWriteStreamsRequest.newBuilder()
- * .setParent(parent.toString())
- * .addAllWriteStreams(writeStreams)
- * .build();
- * ApiFuture<BatchCommitWriteStreamsResponse> future = bigQueryWriteClient.batchCommitWriteStreamsCallable().futureCall(request);
- * // Do something
- * BatchCommitWriteStreamsResponse response = future.get();
- * }
- *
- *
* @param writeStream Required. The stream that is the target of the flush operation.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final FlushRowsResponse flushRows(WriteStreamName writeStream) {
- FlushRowsRequest request =
- FlushRowsRequest.newBuilder()
+ public final Storage.FlushRowsResponse flushRows(WriteStreamName writeStream) {
+ Storage.FlushRowsRequest request =
+ Storage.FlushRowsRequest.newBuilder()
.setWriteStream(writeStream == null ? null : writeStream.toString())
.build();
return flushRows(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Flushes rows to a BUFFERED stream. If users are appending rows to BUFFERED stream, flush
* operation is required in order for the rows to become available for reading. A Flush operation
* flushes up to any previously flushed offset in a BUFFERED stream, to the offset specified in
* the request.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName writeStream = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * FlushRowsResponse response = bigQueryWriteClient.flushRows(writeStream);
- * }
- *
- *
* @param writeStream Required. The stream that is the target of the flush operation.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final FlushRowsResponse flushRows(String writeStream) {
- FlushRowsRequest request = FlushRowsRequest.newBuilder().setWriteStream(writeStream).build();
+ public final Storage.FlushRowsResponse flushRows(String writeStream) {
+ Storage.FlushRowsRequest request =
+ Storage.FlushRowsRequest.newBuilder().setWriteStream(writeStream).build();
return flushRows(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Flushes rows to a BUFFERED stream. If users are appending rows to BUFFERED stream, flush
* operation is required in order for the rows to become available for reading. A Flush operation
* flushes up to any previously flushed offset in a BUFFERED stream, to the offset specified in
* the request.
*
- *
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName writeStream = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * FlushRowsResponse response = bigQueryWriteClient.flushRows(writeStream.toString());
- * }
- *
- *
* @param request The request object containing all of the parameters for the API call.
* @throws com.google.api.gax.rpc.ApiException if the remote call fails
*/
- public final FlushRowsResponse flushRows(FlushRowsRequest request) {
+ public final Storage.FlushRowsResponse flushRows(Storage.FlushRowsRequest request) {
return flushRowsCallable().call(request);
}
- // AUTO-GENERATED DOCUMENTATION AND METHOD
+ // AUTO-GENERATED DOCUMENTATION AND METHOD.
/**
* Flushes rows to a BUFFERED stream. If users are appending rows to BUFFERED stream, flush
* operation is required in order for the rows to become available for reading. A Flush operation
@@ -678,20 +444,9 @@ public final FlushRowsResponse flushRows(FlushRowsRequest request) {
* the request.
*
*
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName writeStream = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * FlushRowsRequest request = FlushRowsRequest.newBuilder()
- * .setWriteStream(writeStream.toString())
- * .build();
- * FlushRowsResponse response = bigQueryWriteClient.flushRows(request);
- * }
- *
*/
- public final UnaryCallable
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * WriteStreamName writeStream = WriteStreamName.of("[PROJECT]", "[DATASET]", "[TABLE]", "[STREAM]");
- * FlushRowsRequest request = FlushRowsRequest.newBuilder()
- * .setWriteStream(writeStream.toString())
- * .build();
- * ApiFuture<FlushRowsResponse> future = bigQueryWriteClient.flushRowsCallable().futureCall(request);
- * // Do something
- * FlushRowsResponse response = future.get();
- * }
- *
- *
+ * }
*/
-@Generated("by gapic-generator")
-@BetaApi
+@Generated("by gapic-generator-java")
public class BigQueryWriteSettings extends ClientSettings
- * BigQueryWriteSettings.Builder bigQueryWriteSettingsBuilder =
- * BigQueryWriteSettings.newBuilder();
+ *
- * {@code
+ * BigQueryWriteSettings.Builder bigQueryWriteSettingsBuilder = BigQueryWriteSettings.newBuilder();
* bigQueryWriteSettingsBuilder
* .createWriteStreamSettings()
* .setRetrySettings(
- * bigQueryWriteSettingsBuilder.createWriteStreamSettings().getRetrySettings().toBuilder()
+ * bigQueryWriteSettingsBuilder
+ * .createWriteStreamSettings()
+ * .getRetrySettings()
+ * .toBuilder()
* .setTotalTimeout(Duration.ofSeconds(30))
* .build());
* BigQueryWriteSettings bigQueryWriteSettings = bigQueryWriteSettingsBuilder.build();
- *
- *
*/
-@Generated("by gapic-generator")
+@Generated("by gapic-generator-java")
package com.google.cloud.bigquery.storage.v1alpha2;
import javax.annotation.Generated;
diff --git a/google-cloud-bigquerystorage/src/main/java/com/google/cloud/bigquery/storage/v1alpha2/stub/BigQueryWriteStub.java b/google-cloud-bigquerystorage/src/main/java/com/google/cloud/bigquery/storage/v1alpha2/stub/BigQueryWriteStub.java
index c86dcd8a28..27ef0b03d7 100644
--- a/google-cloud-bigquerystorage/src/main/java/com/google/cloud/bigquery/storage/v1alpha2/stub/BigQueryWriteStub.java
+++ b/google-cloud-bigquerystorage/src/main/java/com/google/cloud/bigquery/storage/v1alpha2/stub/BigQueryWriteStub.java
@@ -5,7 +5,7 @@
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
- * https://www.apache.org/licenses/LICENSE-2.0
+ * https://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
@@ -13,58 +13,51 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
+
package com.google.cloud.bigquery.storage.v1alpha2.stub;
-import com.google.api.core.BetaApi;
import com.google.api.gax.core.BackgroundResource;
import com.google.api.gax.rpc.BidiStreamingCallable;
import com.google.api.gax.rpc.UnaryCallable;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.AppendRowsRequest;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.AppendRowsResponse;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.BatchCommitWriteStreamsRequest;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.BatchCommitWriteStreamsResponse;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.CreateWriteStreamRequest;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.FinalizeWriteStreamRequest;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.FinalizeWriteStreamResponse;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.FlushRowsRequest;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.FlushRowsResponse;
-import com.google.cloud.bigquery.storage.v1alpha2.Storage.GetWriteStreamRequest;
-import com.google.cloud.bigquery.storage.v1alpha2.Stream.WriteStream;
+import com.google.cloud.bigquery.storage.v1alpha2.Storage;
+import com.google.cloud.bigquery.storage.v1alpha2.Stream;
import javax.annotation.Generated;
-// AUTO-GENERATED DOCUMENTATION AND CLASS
+// AUTO-GENERATED DOCUMENTATION AND CLASS.
/**
- * Base stub class for BigQuery Storage API.
+ * Base stub class for the BigQueryWrite service API.
*
*
- * try (BigQueryWriteClient bigQueryWriteClient = BigQueryWriteClient.create()) {
- * TableName parent = TableName.of("[PROJECT]", "[DATASET]", "[TABLE]");
- * WriteStream writeStream = WriteStream.newBuilder().build();
- * WriteStream response = bigQueryWriteClient.createWriteStream(parent, writeStream);
- * }
- *
- *
- *
+ * }
*/
-@Generated("by gapic-generator")
@BetaApi
+@Generated("by gapic-generator-java")
public class BigQueryWriteStubSettings extends StubSettings
+ *
- * {@code
* BigQueryWriteStubSettings.Builder bigQueryWriteSettingsBuilder =
* BigQueryWriteStubSettings.newBuilder();
* bigQueryWriteSettingsBuilder
* .createWriteStreamSettings()
* .setRetrySettings(
- * bigQueryWriteSettingsBuilder.createWriteStreamSettings().getRetrySettings().toBuilder()
+ * bigQueryWriteSettingsBuilder
+ * .createWriteStreamSettings()
+ * .getRetrySettings()
+ * .toBuilder()
* .setTotalTimeout(Duration.ofSeconds(30))
* .build());
* BigQueryWriteStubSettings bigQueryWriteSettings = bigQueryWriteSettingsBuilder.build();
- *