From a2d137d233e7a401976fbe1fd8ff81145dda515d Mon Sep 17 00:00:00 2001 From: Cody Oss <6331106+codyoss@users.noreply.github.com> Date: Thu, 13 May 2021 15:17:52 -0600 Subject: [PATCH] feat(bigquery/storage): remove alpha client (#4100) This api version should no longer be used. Updates: #4098 --- .../apiv1alpha2/big_query_write_client.go | 358 ------------------ .../big_query_write_client_example_test.go | 171 --------- bigquery/storage/apiv1alpha2/doc.go | 116 ------ .../storage/apiv1alpha2/gapic_metadata.json | 48 --- internal/gapicgen/generator/config.go | 9 - .../BigQueryWriteClient/AppendRows/main.go | 63 --- .../BatchCommitWriteStreams/main.go | 46 --- .../CreateWriteStream/main.go | 46 --- .../FinalizeWriteStream/main.go | 46 --- .../BigQueryWriteClient/FlushRows/main.go | 46 --- .../GetWriteStream/main.go | 46 --- 11 files changed, 995 deletions(-) delete mode 100644 bigquery/storage/apiv1alpha2/big_query_write_client.go delete mode 100644 bigquery/storage/apiv1alpha2/big_query_write_client_example_test.go delete mode 100644 bigquery/storage/apiv1alpha2/doc.go delete mode 100644 bigquery/storage/apiv1alpha2/gapic_metadata.json delete mode 100644 internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/AppendRows/main.go delete mode 100644 internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/BatchCommitWriteStreams/main.go delete mode 100644 internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/CreateWriteStream/main.go delete mode 100644 internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/FinalizeWriteStream/main.go delete mode 100644 internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/FlushRows/main.go delete mode 100644 internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/GetWriteStream/main.go diff --git a/bigquery/storage/apiv1alpha2/big_query_write_client.go b/bigquery/storage/apiv1alpha2/big_query_write_client.go deleted file mode 100644 index ef512bed3a8..00000000000 --- a/bigquery/storage/apiv1alpha2/big_query_write_client.go +++ /dev/null @@ -1,358 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// https://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Code generated by protoc-gen-go_gapic. DO NOT EDIT. - -package storage - -import ( - "context" - "fmt" - "math" - "net/url" - "time" - - gax "github.com/googleapis/gax-go/v2" - "google.golang.org/api/option" - "google.golang.org/api/option/internaloption" - gtransport "google.golang.org/api/transport/grpc" - storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - "google.golang.org/grpc" - "google.golang.org/grpc/codes" - "google.golang.org/grpc/metadata" -) - -var newBigQueryWriteClientHook clientHook - -// BigQueryWriteCallOptions contains the retry settings for each method of BigQueryWriteClient. -type BigQueryWriteCallOptions struct { - CreateWriteStream []gax.CallOption - AppendRows []gax.CallOption - GetWriteStream []gax.CallOption - FinalizeWriteStream []gax.CallOption - BatchCommitWriteStreams []gax.CallOption - FlushRows []gax.CallOption -} - -func defaultBigQueryWriteClientOptions() []option.ClientOption { - return []option.ClientOption{ - internaloption.WithDefaultEndpoint("bigquerystorage.googleapis.com:443"), - internaloption.WithDefaultMTLSEndpoint("bigquerystorage.mtls.googleapis.com:443"), - internaloption.WithDefaultAudience("https://bigquerystorage.googleapis.com/"), - internaloption.WithDefaultScopes(DefaultAuthScopes()...), - option.WithGRPCDialOption(grpc.WithDisableServiceConfig()), - option.WithGRPCDialOption(grpc.WithDefaultCallOptions( - grpc.MaxCallRecvMsgSize(math.MaxInt32))), - } -} - -func defaultBigQueryWriteCallOptions() *BigQueryWriteCallOptions { - return &BigQueryWriteCallOptions{ - CreateWriteStream: []gax.CallOption{ - gax.WithRetry(func() gax.Retryer { - return gax.OnCodes([]codes.Code{ - codes.DeadlineExceeded, - codes.Unavailable, - codes.ResourceExhausted, - }, gax.Backoff{ - Initial: 100 * time.Millisecond, - Max: 60000 * time.Millisecond, - Multiplier: 1.30, - }) - }), - }, - AppendRows: []gax.CallOption{ - gax.WithRetry(func() gax.Retryer { - return gax.OnCodes([]codes.Code{ - codes.Unavailable, - codes.ResourceExhausted, - }, gax.Backoff{ - Initial: 100 * time.Millisecond, - Max: 60000 * time.Millisecond, - Multiplier: 1.30, - }) - }), - }, - GetWriteStream: []gax.CallOption{ - gax.WithRetry(func() gax.Retryer { - return gax.OnCodes([]codes.Code{ - codes.DeadlineExceeded, - codes.Unavailable, - }, gax.Backoff{ - Initial: 100 * time.Millisecond, - Max: 60000 * time.Millisecond, - Multiplier: 1.30, - }) - }), - }, - FinalizeWriteStream: []gax.CallOption{ - gax.WithRetry(func() gax.Retryer { - return gax.OnCodes([]codes.Code{ - codes.DeadlineExceeded, - codes.Unavailable, - }, gax.Backoff{ - Initial: 100 * time.Millisecond, - Max: 60000 * time.Millisecond, - Multiplier: 1.30, - }) - }), - }, - BatchCommitWriteStreams: []gax.CallOption{ - gax.WithRetry(func() gax.Retryer { - return gax.OnCodes([]codes.Code{ - codes.DeadlineExceeded, - codes.Unavailable, - }, gax.Backoff{ - Initial: 100 * time.Millisecond, - Max: 60000 * time.Millisecond, - Multiplier: 1.30, - }) - }), - }, - FlushRows: []gax.CallOption{}, - } -} - -// BigQueryWriteClient is a client for interacting with BigQuery Storage API. -// -// Methods, except Close, may be called concurrently. However, fields must not be modified concurrently with method calls. -type BigQueryWriteClient struct { - // Connection pool of gRPC connections to the service. - connPool gtransport.ConnPool - - // flag to opt out of default deadlines via GOOGLE_API_GO_EXPERIMENTAL_DISABLE_DEFAULT_DEADLINE - disableDeadlines bool - - // The gRPC API client. - bigQueryWriteClient storagepb.BigQueryWriteClient - - // The call options for this service. - CallOptions *BigQueryWriteCallOptions - - // The x-goog-* metadata to be sent with each request. - xGoogMetadata metadata.MD -} - -// NewBigQueryWriteClient creates a new big query write client. -// -// BigQuery Write API. -// -// The Write API can be used to write data to BigQuery. -func NewBigQueryWriteClient(ctx context.Context, opts ...option.ClientOption) (*BigQueryWriteClient, error) { - clientOpts := defaultBigQueryWriteClientOptions() - - if newBigQueryWriteClientHook != nil { - hookOpts, err := newBigQueryWriteClientHook(ctx, clientHookParams{}) - if err != nil { - return nil, err - } - clientOpts = append(clientOpts, hookOpts...) - } - - disableDeadlines, err := checkDisableDeadlines() - if err != nil { - return nil, err - } - - connPool, err := gtransport.DialPool(ctx, append(clientOpts, opts...)...) - if err != nil { - return nil, err - } - c := &BigQueryWriteClient{ - connPool: connPool, - disableDeadlines: disableDeadlines, - CallOptions: defaultBigQueryWriteCallOptions(), - - bigQueryWriteClient: storagepb.NewBigQueryWriteClient(connPool), - } - c.setGoogleClientInfo() - - return c, nil -} - -// Connection returns a connection to the API service. -// -// Deprecated. -func (c *BigQueryWriteClient) Connection() *grpc.ClientConn { - return c.connPool.Conn() -} - -// Close closes the connection to the API service. The user should invoke this when -// the client is no longer required. -func (c *BigQueryWriteClient) Close() error { - return c.connPool.Close() -} - -// setGoogleClientInfo sets the name and version of the application in -// the `x-goog-api-client` header passed on each request. Intended for -// use by Google-written clients. -func (c *BigQueryWriteClient) setGoogleClientInfo(keyval ...string) { - kv := append([]string{"gl-go", versionGo()}, keyval...) - kv = append(kv, "gapic", versionClient, "gax", gax.Version, "grpc", grpc.Version) - c.xGoogMetadata = metadata.Pairs("x-goog-api-client", gax.XGoogHeader(kv...)) -} - -// CreateWriteStream creates a write stream to the given table. -// Additionally, every table has a special COMMITTED stream named ‘_default’ -// to which data can be written. This stream doesn’t need to be created using -// CreateWriteStream. It is a stream that can be used simultaneously by any -// number of clients. Data written to this stream is considered committed as -// soon as an acknowledgement is received. -func (c *BigQueryWriteClient) CreateWriteStream(ctx context.Context, req *storagepb.CreateWriteStreamRequest, opts ...gax.CallOption) (*storagepb.WriteStream, error) { - if _, ok := ctx.Deadline(); !ok && !c.disableDeadlines { - cctx, cancel := context.WithTimeout(ctx, 600000*time.Millisecond) - defer cancel() - ctx = cctx - } - md := metadata.Pairs("x-goog-request-params", fmt.Sprintf("%s=%v", "parent", url.QueryEscape(req.GetParent()))) - ctx = insertMetadata(ctx, c.xGoogMetadata, md) - opts = append(c.CallOptions.CreateWriteStream[0:len(c.CallOptions.CreateWriteStream):len(c.CallOptions.CreateWriteStream)], opts...) - var resp *storagepb.WriteStream - err := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error { - var err error - resp, err = c.bigQueryWriteClient.CreateWriteStream(ctx, req, settings.GRPC...) - return err - }, opts...) - if err != nil { - return nil, err - } - return resp, nil -} - -// AppendRows appends data to the given stream. -// -// If offset is specified, the offset is checked against the end of -// stream. The server returns OUT_OF_RANGE in AppendRowsResponse if an -// attempt is made to append to an offset beyond the current end of the stream -// or ALREADY_EXISTS if user provids an offset that has already been -// written to. User can retry with adjusted offset within the same RPC -// stream. If offset is not specified, append happens at the end of the -// stream. -// -// The response contains the offset at which the append happened. Responses -// are received in the same order in which requests are sent. There will be -// one response for each successful request. If the offset is not set in -// response, it means append didn’t happen due to some errors. If one request -// fails, all the subsequent requests will also fail until a success request -// is made again. -// -// If the stream is of PENDING type, data will only be available for read -// operations after the stream is committed. -func (c *BigQueryWriteClient) AppendRows(ctx context.Context, opts ...gax.CallOption) (storagepb.BigQueryWrite_AppendRowsClient, error) { - ctx = insertMetadata(ctx, c.xGoogMetadata) - opts = append(c.CallOptions.AppendRows[0:len(c.CallOptions.AppendRows):len(c.CallOptions.AppendRows)], opts...) - var resp storagepb.BigQueryWrite_AppendRowsClient - err := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error { - var err error - resp, err = c.bigQueryWriteClient.AppendRows(ctx, settings.GRPC...) - return err - }, opts...) - if err != nil { - return nil, err - } - return resp, nil -} - -// GetWriteStream gets a write stream. -func (c *BigQueryWriteClient) GetWriteStream(ctx context.Context, req *storagepb.GetWriteStreamRequest, opts ...gax.CallOption) (*storagepb.WriteStream, error) { - if _, ok := ctx.Deadline(); !ok && !c.disableDeadlines { - cctx, cancel := context.WithTimeout(ctx, 600000*time.Millisecond) - defer cancel() - ctx = cctx - } - md := metadata.Pairs("x-goog-request-params", fmt.Sprintf("%s=%v", "name", url.QueryEscape(req.GetName()))) - ctx = insertMetadata(ctx, c.xGoogMetadata, md) - opts = append(c.CallOptions.GetWriteStream[0:len(c.CallOptions.GetWriteStream):len(c.CallOptions.GetWriteStream)], opts...) - var resp *storagepb.WriteStream - err := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error { - var err error - resp, err = c.bigQueryWriteClient.GetWriteStream(ctx, req, settings.GRPC...) - return err - }, opts...) - if err != nil { - return nil, err - } - return resp, nil -} - -// FinalizeWriteStream finalize a write stream so that no new data can be appended to the -// stream. Finalize is not supported on the ‘_default’ stream. -func (c *BigQueryWriteClient) FinalizeWriteStream(ctx context.Context, req *storagepb.FinalizeWriteStreamRequest, opts ...gax.CallOption) (*storagepb.FinalizeWriteStreamResponse, error) { - if _, ok := ctx.Deadline(); !ok && !c.disableDeadlines { - cctx, cancel := context.WithTimeout(ctx, 600000*time.Millisecond) - defer cancel() - ctx = cctx - } - md := metadata.Pairs("x-goog-request-params", fmt.Sprintf("%s=%v", "name", url.QueryEscape(req.GetName()))) - ctx = insertMetadata(ctx, c.xGoogMetadata, md) - opts = append(c.CallOptions.FinalizeWriteStream[0:len(c.CallOptions.FinalizeWriteStream):len(c.CallOptions.FinalizeWriteStream)], opts...) - var resp *storagepb.FinalizeWriteStreamResponse - err := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error { - var err error - resp, err = c.bigQueryWriteClient.FinalizeWriteStream(ctx, req, settings.GRPC...) - return err - }, opts...) - if err != nil { - return nil, err - } - return resp, nil -} - -// BatchCommitWriteStreams atomically commits a group of PENDING streams that belong to the same -// parent table. -// Streams must be finalized before commit and cannot be committed multiple -// times. Once a stream is committed, data in the stream becomes available -// for read operations. -func (c *BigQueryWriteClient) BatchCommitWriteStreams(ctx context.Context, req *storagepb.BatchCommitWriteStreamsRequest, opts ...gax.CallOption) (*storagepb.BatchCommitWriteStreamsResponse, error) { - if _, ok := ctx.Deadline(); !ok && !c.disableDeadlines { - cctx, cancel := context.WithTimeout(ctx, 600000*time.Millisecond) - defer cancel() - ctx = cctx - } - md := metadata.Pairs("x-goog-request-params", fmt.Sprintf("%s=%v", "parent", url.QueryEscape(req.GetParent()))) - ctx = insertMetadata(ctx, c.xGoogMetadata, md) - opts = append(c.CallOptions.BatchCommitWriteStreams[0:len(c.CallOptions.BatchCommitWriteStreams):len(c.CallOptions.BatchCommitWriteStreams)], opts...) - var resp *storagepb.BatchCommitWriteStreamsResponse - err := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error { - var err error - resp, err = c.bigQueryWriteClient.BatchCommitWriteStreams(ctx, req, settings.GRPC...) - return err - }, opts...) - if err != nil { - return nil, err - } - return resp, nil -} - -// FlushRows flushes rows to a BUFFERED stream. -// If users are appending rows to BUFFERED stream, flush operation is -// required in order for the rows to become available for reading. A -// Flush operation flushes up to any previously flushed offset in a BUFFERED -// stream, to the offset specified in the request. -// Flush is not supported on the _default stream, since it is not BUFFERED. -func (c *BigQueryWriteClient) FlushRows(ctx context.Context, req *storagepb.FlushRowsRequest, opts ...gax.CallOption) (*storagepb.FlushRowsResponse, error) { - md := metadata.Pairs("x-goog-request-params", fmt.Sprintf("%s=%v", "write_stream", url.QueryEscape(req.GetWriteStream()))) - ctx = insertMetadata(ctx, c.xGoogMetadata, md) - opts = append(c.CallOptions.FlushRows[0:len(c.CallOptions.FlushRows):len(c.CallOptions.FlushRows)], opts...) - var resp *storagepb.FlushRowsResponse - err := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error { - var err error - resp, err = c.bigQueryWriteClient.FlushRows(ctx, req, settings.GRPC...) - return err - }, opts...) - if err != nil { - return nil, err - } - return resp, nil -} diff --git a/bigquery/storage/apiv1alpha2/big_query_write_client_example_test.go b/bigquery/storage/apiv1alpha2/big_query_write_client_example_test.go deleted file mode 100644 index 56840651918..00000000000 --- a/bigquery/storage/apiv1alpha2/big_query_write_client_example_test.go +++ /dev/null @@ -1,171 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// https://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Code generated by protoc-gen-go_gapic. DO NOT EDIT. - -package storage_test - -import ( - "context" - "io" - - storage "cloud.google.com/go/bigquery/storage/apiv1alpha2" - storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" -) - -func ExampleNewBigQueryWriteClient() { - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - // TODO: Use client. - _ = c -} - -func ExampleBigQueryWriteClient_CreateWriteStream() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.CreateWriteStreamRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.CreateWriteStream(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -func ExampleBigQueryWriteClient_AppendRows() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - stream, err := c.AppendRows(ctx) - if err != nil { - // TODO: Handle error. - } - go func() { - reqs := []*storagepb.AppendRowsRequest{ - // TODO: Create requests. - } - for _, req := range reqs { - if err := stream.Send(req); err != nil { - // TODO: Handle error. - } - } - stream.CloseSend() - }() - for { - resp, err := stream.Recv() - if err == io.EOF { - break - } - if err != nil { - // TODO: handle error. - } - // TODO: Use resp. - _ = resp - } -} - -func ExampleBigQueryWriteClient_GetWriteStream() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.GetWriteStreamRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.GetWriteStream(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -func ExampleBigQueryWriteClient_FinalizeWriteStream() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.FinalizeWriteStreamRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.FinalizeWriteStream(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -func ExampleBigQueryWriteClient_BatchCommitWriteStreams() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.BatchCommitWriteStreamsRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.BatchCommitWriteStreams(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -func ExampleBigQueryWriteClient_FlushRows() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.FlushRowsRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.FlushRows(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} diff --git a/bigquery/storage/apiv1alpha2/doc.go b/bigquery/storage/apiv1alpha2/doc.go deleted file mode 100644 index 24a7f0d4818..00000000000 --- a/bigquery/storage/apiv1alpha2/doc.go +++ /dev/null @@ -1,116 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// https://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Code generated by protoc-gen-go_gapic. DO NOT EDIT. - -// Package storage is an auto-generated package for the -// BigQuery Storage API. -// -// Use of Context -// -// The ctx passed to NewClient is used for authentication requests and -// for creating the underlying connection, but is not used for subsequent calls. -// Individual methods on the client use the ctx given to them. -// -// To close the open connection, use the Close() method. -// -// For information about setting deadlines, reusing contexts, and more -// please visit pkg.go.dev/cloud.google.com/go. -package storage // import "cloud.google.com/go/bigquery/storage/apiv1alpha2" - -import ( - "context" - "os" - "runtime" - "strconv" - "strings" - "unicode" - - "google.golang.org/api/option" - "google.golang.org/grpc/metadata" -) - -// For more information on implementing a client constructor hook, see -// https://github.com/googleapis/google-cloud-go/wiki/Customizing-constructors. -type clientHookParams struct{} -type clientHook func(context.Context, clientHookParams) ([]option.ClientOption, error) - -const versionClient = "20210508" - -func insertMetadata(ctx context.Context, mds ...metadata.MD) context.Context { - out, _ := metadata.FromOutgoingContext(ctx) - out = out.Copy() - for _, md := range mds { - for k, v := range md { - out[k] = append(out[k], v...) - } - } - return metadata.NewOutgoingContext(ctx, out) -} - -func checkDisableDeadlines() (bool, error) { - raw, ok := os.LookupEnv("GOOGLE_API_GO_EXPERIMENTAL_DISABLE_DEFAULT_DEADLINE") - if !ok { - return false, nil - } - - b, err := strconv.ParseBool(raw) - return b, err -} - -// DefaultAuthScopes reports the default set of authentication scopes to use with this package. -func DefaultAuthScopes() []string { - return []string{ - "https://www.googleapis.com/auth/bigquery", - "https://www.googleapis.com/auth/bigquery.insertdata", - "https://www.googleapis.com/auth/cloud-platform", - } -} - -// versionGo returns the Go runtime version. The returned string -// has no whitespace, suitable for reporting in header. -func versionGo() string { - const develPrefix = "devel +" - - s := runtime.Version() - if strings.HasPrefix(s, develPrefix) { - s = s[len(develPrefix):] - if p := strings.IndexFunc(s, unicode.IsSpace); p >= 0 { - s = s[:p] - } - return s - } - - notSemverRune := func(r rune) bool { - return !strings.ContainsRune("0123456789.", r) - } - - if strings.HasPrefix(s, "go1") { - s = s[2:] - var prerelease string - if p := strings.IndexFunc(s, notSemverRune); p >= 0 { - s, prerelease = s[:p], s[p:] - } - if strings.HasSuffix(s, ".") { - s += "0" - } else if strings.Count(s, ".") < 2 { - s += ".0" - } - if prerelease != "" { - s += "-" + prerelease - } - return s - } - return "UNKNOWN" -} diff --git a/bigquery/storage/apiv1alpha2/gapic_metadata.json b/bigquery/storage/apiv1alpha2/gapic_metadata.json deleted file mode 100644 index 51625515111..00000000000 --- a/bigquery/storage/apiv1alpha2/gapic_metadata.json +++ /dev/null @@ -1,48 +0,0 @@ -{ - "schema": "1.0", - "comment": "This file maps proto services/RPCs to the corresponding library clients/methods.", - "language": "go", - "protoPackage": "google.cloud.bigquery.storage.v1alpha2", - "libraryPackage": "cloud.google.com/go/bigquery/storage/apiv1alpha2", - "services": { - "BigQueryWrite": { - "clients": { - "grpc": { - "libraryClient": "BigQueryWriteClient", - "rpcs": { - "AppendRows": { - "methods": [ - "AppendRows" - ] - }, - "BatchCommitWriteStreams": { - "methods": [ - "BatchCommitWriteStreams" - ] - }, - "CreateWriteStream": { - "methods": [ - "CreateWriteStream" - ] - }, - "FinalizeWriteStream": { - "methods": [ - "FinalizeWriteStream" - ] - }, - "FlushRows": { - "methods": [ - "FlushRows" - ] - }, - "GetWriteStream": { - "methods": [ - "GetWriteStream" - ] - } - } - } - } - } - } -} diff --git a/internal/gapicgen/generator/config.go b/internal/gapicgen/generator/config.go index 9ff99543a3c..ae9910e1b11 100644 --- a/internal/gapicgen/generator/config.go +++ b/internal/gapicgen/generator/config.go @@ -194,15 +194,6 @@ var microgenGapicConfigs = []*microgenConfig{ apiServiceConfigPath: "google/cloud/bigquery/reservation/v1/bigqueryreservation_v1.yaml", releaseLevel: "ga", }, - { - inputDirectoryPath: "google/cloud/bigquery/storage/v1alpha2", - pkg: "storage", - importPath: "cloud.google.com/go/bigquery/storage/apiv1alpha2", - gRPCServiceConfigPath: "google/cloud/bigquery/storage/v1alpha2/bigquerystorage_grpc_service_config.json", - apiServiceConfigPath: "google/cloud/bigquery/storage/v1alpha2/bigquerystorage_v1alpha2.yaml", - releaseLevel: "alpha", - stopGeneration: true, - }, { inputDirectoryPath: "google/cloud/bigquery/storage/v1beta1", pkg: "storage", diff --git a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/AppendRows/main.go b/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/AppendRows/main.go deleted file mode 100644 index 44c08a84a39..00000000000 --- a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/AppendRows/main.go +++ /dev/null @@ -1,63 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// [START bigquerystorage_v1alpha2_generated_BigQueryWrite_AppendRows_sync] - -package main - -import ( - "context" - "io" - - storage "cloud.google.com/go/bigquery/storage/apiv1alpha2" - storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" -) - -func main() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - stream, err := c.AppendRows(ctx) - if err != nil { - // TODO: Handle error. - } - go func() { - reqs := []*storagepb.AppendRowsRequest{ - // TODO: Create requests. - } - for _, req := range reqs { - if err := stream.Send(req); err != nil { - // TODO: Handle error. - } - } - stream.CloseSend() - }() - for { - resp, err := stream.Recv() - if err == io.EOF { - break - } - if err != nil { - // TODO: handle error. - } - // TODO: Use resp. - _ = resp - } -} - -// [END bigquerystorage_v1alpha2_generated_BigQueryWrite_AppendRows_sync] diff --git a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/BatchCommitWriteStreams/main.go b/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/BatchCommitWriteStreams/main.go deleted file mode 100644 index 683c6c5cb06..00000000000 --- a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/BatchCommitWriteStreams/main.go +++ /dev/null @@ -1,46 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// [START bigquerystorage_v1alpha2_generated_BigQueryWrite_BatchCommitWriteStreams_sync] - -package main - -import ( - "context" - - storage "cloud.google.com/go/bigquery/storage/apiv1alpha2" - storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" -) - -func main() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.BatchCommitWriteStreamsRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.BatchCommitWriteStreams(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -// [END bigquerystorage_v1alpha2_generated_BigQueryWrite_BatchCommitWriteStreams_sync] diff --git a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/CreateWriteStream/main.go b/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/CreateWriteStream/main.go deleted file mode 100644 index 44c580f2e45..00000000000 --- a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/CreateWriteStream/main.go +++ /dev/null @@ -1,46 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// [START bigquerystorage_v1alpha2_generated_BigQueryWrite_CreateWriteStream_sync] - -package main - -import ( - "context" - - storage "cloud.google.com/go/bigquery/storage/apiv1alpha2" - storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" -) - -func main() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.CreateWriteStreamRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.CreateWriteStream(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -// [END bigquerystorage_v1alpha2_generated_BigQueryWrite_CreateWriteStream_sync] diff --git a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/FinalizeWriteStream/main.go b/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/FinalizeWriteStream/main.go deleted file mode 100644 index 33d1665856a..00000000000 --- a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/FinalizeWriteStream/main.go +++ /dev/null @@ -1,46 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// [START bigquerystorage_v1alpha2_generated_BigQueryWrite_FinalizeWriteStream_sync] - -package main - -import ( - "context" - - storage "cloud.google.com/go/bigquery/storage/apiv1alpha2" - storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" -) - -func main() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.FinalizeWriteStreamRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.FinalizeWriteStream(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -// [END bigquerystorage_v1alpha2_generated_BigQueryWrite_FinalizeWriteStream_sync] diff --git a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/FlushRows/main.go b/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/FlushRows/main.go deleted file mode 100644 index 0566a4a85cb..00000000000 --- a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/FlushRows/main.go +++ /dev/null @@ -1,46 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// [START bigquerystorage_v1alpha2_generated_BigQueryWrite_FlushRows_sync] - -package main - -import ( - "context" - - storage "cloud.google.com/go/bigquery/storage/apiv1alpha2" - storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" -) - -func main() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.FlushRowsRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.FlushRows(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -// [END bigquerystorage_v1alpha2_generated_BigQueryWrite_FlushRows_sync] diff --git a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/GetWriteStream/main.go b/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/GetWriteStream/main.go deleted file mode 100644 index 2fdda90d94e..00000000000 --- a/internal/generated/snippets/bigquery/storage/apiv1alpha2/BigQueryWriteClient/GetWriteStream/main.go +++ /dev/null @@ -1,46 +0,0 @@ -// Copyright 2021 Google LLC -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// [START bigquerystorage_v1alpha2_generated_BigQueryWrite_GetWriteStream_sync] - -package main - -import ( - "context" - - storage "cloud.google.com/go/bigquery/storage/apiv1alpha2" - storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" -) - -func main() { - // import storagepb "google.golang.org/genproto/googleapis/cloud/bigquery/storage/v1alpha2" - - ctx := context.Background() - c, err := storage.NewBigQueryWriteClient(ctx) - if err != nil { - // TODO: Handle error. - } - - req := &storagepb.GetWriteStreamRequest{ - // TODO: Fill request struct fields. - } - resp, err := c.GetWriteStream(ctx, req) - if err != nil { - // TODO: Handle error. - } - // TODO: Use resp. - _ = resp -} - -// [END bigquerystorage_v1alpha2_generated_BigQueryWrite_GetWriteStream_sync]