From 466c9a938ee07debe7d6e35550b473c62f9f3992 Mon Sep 17 00:00:00 2001 From: yoshi-code-bot <70984784+yoshi-code-bot@users.noreply.github.com> Date: Mon, 22 Nov 2021 23:24:31 -0800 Subject: [PATCH] chore: Update discovery artifacts (#1609) ## Deleted keys were detected in the following stable discovery artifacts: compute v1 https://github.com/googleapis/google-api-python-client/commit/574d4ace68dc1bd844f01ec13419b7d055fa857e ## Deleted keys were detected in the following pre-stable discovery artifacts: compute alpha https://github.com/googleapis/google-api-python-client/commit/574d4ace68dc1bd844f01ec13419b7d055fa857e compute beta https://github.com/googleapis/google-api-python-client/commit/574d4ace68dc1bd844f01ec13419b7d055fa857e gameservices v1beta https://github.com/googleapis/google-api-python-client/commit/8016e68c4c2ac2f065fc1a0c7e86aac35329cf25 ## Discovery Artifact Change Summary: feat(artifactregistry): update the api https://github.com/googleapis/google-api-python-client/commit/faa6c7f4fab2190989c9a587f35998203517ff2a feat(assuredworkloads): update the api https://github.com/googleapis/google-api-python-client/commit/c0f6be757e47a62337e7963bea576858cbf17341 feat(cloudbuild): update the api https://github.com/googleapis/google-api-python-client/commit/536d4b0a61e15b45439f3f3066410236d50e228b feat(composer): update the api https://github.com/googleapis/google-api-python-client/commit/bc7ce92bf38d4eed3fce35cd080bf40b43855ba4 feat(compute): update the api https://github.com/googleapis/google-api-python-client/commit/574d4ace68dc1bd844f01ec13419b7d055fa857e feat(container): update the api https://github.com/googleapis/google-api-python-client/commit/c4cdd10e83565119457494a21f1abe128ac59083 feat(containeranalysis): update the api https://github.com/googleapis/google-api-python-client/commit/a86a46ed4f6098a6da8d3bc2db8ae1642e4300f3 feat(datastream): update the api https://github.com/googleapis/google-api-python-client/commit/f54ebeac6f8829b81b975a9fcabde5423f769ea9 feat(dialogflow): update the api https://github.com/googleapis/google-api-python-client/commit/2b33ab2f6823916a872f33c23c9da86b993f0fc0 feat(dlp): update the api https://github.com/googleapis/google-api-python-client/commit/0406443bd63e9bf9312c3ac7a7398ca0a60dac51 feat(gameservices): update the api https://github.com/googleapis/google-api-python-client/commit/8016e68c4c2ac2f065fc1a0c7e86aac35329cf25 feat(gkehub): update the api https://github.com/googleapis/google-api-python-client/commit/fbcb756aa75d122f9929a6319b3bdd413cebc293 feat(logging): update the api https://github.com/googleapis/google-api-python-client/commit/8ce3b8583936958cabce8c7a0cd72ad16817de0c feat(ml): update the api https://github.com/googleapis/google-api-python-client/commit/cfe625ae36d61eda2f496115ad515a7000d8973c feat(networkconnectivity): update the api https://github.com/googleapis/google-api-python-client/commit/6348dd518b3be79230e3d23beb4e4f2527d67855 feat(osconfig): update the api https://github.com/googleapis/google-api-python-client/commit/0c6c8148b57f793fab971f2f815acef70fde478b feat(storagetransfer): update the api https://github.com/googleapis/google-api-python-client/commit/5758865ada7f1c87525752f002f82ca8681e927d --- ...roidmanagement_v1.enterprises.devices.html | 2 +- ...ay_v1.projects.locations.apis.configs.html | 2 +- ...apigateway_v1.projects.locations.apis.html | 2 +- ...ateway_v1.projects.locations.gateways.html | 2 +- ...1beta.projects.locations.apis.configs.html | 2 +- ...ateway_v1beta.projects.locations.apis.html | 2 +- ...ay_v1beta.projects.locations.gateways.html | 2 +- docs/dyn/apigee_v1.organizations.apps.html | 4 +- ...pigee_v1.organizations.datacollectors.html | 2 +- ...igee_v1.organizations.developers.apps.html | 18 +- .../apigee_v1.organizations.environments.html | 2 +- docs/dyn/apigee_v1.organizations.html | 10 +- ...s.locations.repositories.dockerImages.html | 30 + ...beta1.projects.locations.repositories.html | 2 +- ...beta2.projects.locations.repositories.html | 2 +- docs/dyn/clouddeploy_v1.html | 111 + docs/dyn/clouddeploy_v1.projects.html | 91 + ....projects.locations.deliveryPipelines.html | 612 +++++ ....locations.deliveryPipelines.releases.html | 539 ++++ ...s.deliveryPipelines.releases.rollouts.html | 292 ++ .../clouddeploy_v1.projects.locations.html | 219 ++ ...ploy_v1.projects.locations.operations.html | 235 ++ ...ddeploy_v1.projects.locations.targets.html | 598 +++++ ...ity_v1beta1.customers.userinvitations.html | 4 +- ....projects.locations.registries.groups.html | 2 +- ...diot_v1.projects.locations.registries.html | 2 +- ...er_v1.projects.locations.environments.html | 8 +- ...beta1.projects.locations.environments.html | 14 +- ...r_v1beta1.projects.locations.clusters.html | 12 + ...ainer_v1beta1.projects.zones.clusters.html | 12 + ...s_v1alpha1.projects.notes.occurrences.html | 83 +- ...nalysis_v1alpha1.projects.occurrences.html | 498 +++- ..._v1alpha1.providers.notes.occurrences.html | 83 +- docs/dyn/datastore_v1.projects.html | 286 +- ...projects.locations.connectionProfiles.html | 653 +++++ .../dyn/datastream_v1.projects.locations.html | 59 + ...projects.locations.privateConnections.html | 311 +++ ...s.locations.privateConnections.routes.html | 269 ++ ...astream_v1.projects.locations.streams.html | 1053 ++++++++ ...v1.projects.locations.streams.objects.html | 457 ++++ ...m_v1alpha1.projects.locations.streams.html | 5 + ...a1.projects.locations.streams.objects.html | 366 +++ ...2.projects.conversations.participants.html | 23 + ...onversations.participants.suggestions.html | 10 + ....locations.conversations.participants.html | 23 + ...onversations.participants.suggestions.html | 10 + ...1.projects.conversations.participants.html | 23 + ...onversations.participants.suggestions.html | 10 + ....locations.conversations.participants.html | 23 + ...onversations.participants.suggestions.html | 10 + ...v3.projects.locations.agents.webhooks.html | 36 + ...a1.projects.locations.agents.webhooks.html | 36 + ..._v2.organizations.deidentifyTemplates.html | 126 + ...zations.locations.deidentifyTemplates.html | 126 + docs/dyn/dlp_v2.projects.content.html | 84 + .../dlp_v2.projects.deidentifyTemplates.html | 126 + .../dlp_v2.projects.locations.content.html | 84 + ...rojects.locations.deidentifyTemplates.html | 126 + ...s.locations.realms.gameServerClusters.html | 24 - ...hub_v1alpha.projects.locations.fleets.html | 218 ++ .../gkehub_v1alpha.projects.locations.html | 5 + docs/dyn/index.md | 4 + .../logging_v2.billingAccounts.buckets.html | 5 + docs/dyn/logging_v2.billingAccounts.html | 24 + ..._v2.billingAccounts.locations.buckets.html | 25 + docs/dyn/logging_v2.folders.html | 24 + .../logging_v2.folders.locations.buckets.html | 30 + docs/dyn/logging_v2.locations.buckets.html | 30 + docs/dyn/logging_v2.organizations.html | 30 +- ...ng_v2.organizations.locations.buckets.html | 30 + docs/dyn/logging_v2.projects.html | 24 + ...logging_v2.projects.locations.buckets.html | 30 + docs/dyn/logging_v2.v2.html | 30 +- docs/dyn/ml_v1.projects.jobs.html | 8 +- docs/dyn/ml_v1.projects.models.html | 2 +- ...ty_v1.projects.locations.global_.hubs.html | 4 +- ...s.locations.global_.policyBasedRoutes.html | 2 +- ...ectivity_v1.projects.locations.spokes.html | 4 +- ...lpha1.projects.locations.global_.hubs.html | 2 +- ...ha1.projects.locations.internalRanges.html | 2 +- ...ty_v1alpha1.projects.locations.spokes.html | 2 +- ...ations.instances.vulnerabilityReports.html | 16 + ...osconfig_v1.projects.patchDeployments.html | 6 + ...ations.instances.vulnerabilityReports.html | 16 + ...nfig_v1beta.projects.patchDeployments.html | 6 + docs/dyn/pubsub_v1.projects.schemas.html | 2 +- docs/dyn/pubsub_v1.projects.snapshots.html | 2 +- .../dyn/pubsub_v1.projects.subscriptions.html | 2 +- docs/dyn/pubsub_v1.projects.topics.html | 2 +- ...pubsub_v1beta2.projects.subscriptions.html | 2 +- docs/dyn/pubsub_v1beta2.projects.topics.html | 2 +- ...rojects.locations.catalogs.placements.html | 4 +- ...l_v2alpha.projects.locations.catalogs.html | 6 +- ...rojects.locations.catalogs.placements.html | 4 +- ...rojects.locations.catalogs.placements.html | 4 +- .../dyn/run_v1.namespaces.configurations.html | 4 +- docs/dyn/run_v1.namespaces.revisions.html | 4 +- docs/dyn/run_v1.namespaces.services.html | 12 +- ..._v1.projects.locations.configurations.html | 4 +- .../run_v1.projects.locations.revisions.html | 4 +- .../run_v1.projects.locations.services.html | 12 +- docs/dyn/run_v1alpha1.namespaces.jobs.html | 8 +- docs/dyn/storagetransfer_v1.transferJobs.html | 54 + .../documents/abusiveexperiencereport.v1.json | 2 +- .../acceleratedmobilepageurl.v1.json | 2 +- .../documents/accesscontextmanager.v1.json | 2 +- .../accesscontextmanager.v1beta.json | 2 +- .../documents/adexchangebuyer2.v2beta1.json | 2 +- .../documents/adexperiencereport.v1.json | 2 +- .../documents/admin.datatransfer_v1.json | 2 +- .../documents/admin.directory_v1.json | 2 +- .../documents/admin.reports_v1.json | 2 +- .../discovery_cache/documents/admob.v1.json | 2 +- .../documents/admob.v1beta.json | 2 +- .../documents/analyticsadmin.v1alpha.json | 2 +- .../documents/analyticsdata.v1beta.json | 2 +- .../documents/analyticsreporting.v4.json | 2 +- .../androiddeviceprovisioning.v1.json | 2 +- .../documents/androidenterprise.v1.json | 2 +- .../documents/androidmanagement.v1.json | 4 +- .../documents/androidpublisher.v3.json | 2 +- .../documents/apigateway.v1.json | 8 +- .../documents/apigateway.v1beta.json | 8 +- .../discovery_cache/documents/apigee.v1.json | 10 +- .../documents/appengine.v1.json | 2 +- .../documents/appengine.v1alpha.json | 2 +- .../documents/appengine.v1beta.json | 2 +- .../documents/area120tables.v1alpha1.json | 2 +- .../documents/artifactregistry.v1.json | 28 +- .../documents/artifactregistry.v1beta1.json | 4 +- .../documents/artifactregistry.v1beta2.json | 4 +- .../documents/assuredworkloads.v1.json | 284 +- .../authorizedbuyersmarketplace.v1.json | 2 +- .../documents/billingbudgets.v1.json | 2 +- .../documents/billingbudgets.v1beta1.json | 2 +- .../documents/binaryauthorization.v1.json | 4 +- .../binaryauthorization.v1beta1.json | 4 +- .../discovery_cache/documents/blogger.v2.json | 2 +- .../discovery_cache/documents/blogger.v3.json | 2 +- .../discovery_cache/documents/books.v1.json | 2 +- .../discovery_cache/documents/chat.v1.json | 2 +- .../documents/chromemanagement.v1.json | 2 +- .../documents/chromepolicy.v1.json | 2 +- .../documents/chromeuxreport.v1.json | 2 +- .../documents/classroom.v1.json | 2 +- .../documents/cloudbuild.v1.json | 23 +- .../documents/cloudbuild.v1alpha1.json | 23 +- .../documents/cloudbuild.v1alpha2.json | 23 +- .../documents/cloudbuild.v1beta1.json | 23 +- .../documents/cloudchannel.v1.json | 2 +- .../documents/clouddebugger.v2.json | 2 +- .../documents/clouddeploy.v1.json | 2351 +++++++++++++++++ .../clouderrorreporting.v1beta1.json | 2 +- .../documents/cloudfunctions.v1.json | 2 +- .../documents/cloudidentity.v1.json | 4 +- .../documents/cloudidentity.v1beta1.json | 4 +- .../documents/cloudiot.v1.json | 4 +- .../documents/cloudprofiler.v2.json | 2 +- .../documents/cloudresourcemanager.v1.json | 2 +- .../cloudresourcemanager.v1beta1.json | 2 +- .../documents/cloudresourcemanager.v2.json | 2 +- .../cloudresourcemanager.v2beta1.json | 2 +- .../documents/cloudresourcemanager.v3.json | 2 +- .../documents/cloudsearch.v1.json | 2 +- .../documents/cloudshell.v1.json | 2 +- .../documents/cloudsupport.v2beta.json | 2 +- .../documents/cloudtrace.v1.json | 2 +- .../documents/cloudtrace.v2.json | 2 +- .../documents/cloudtrace.v2beta1.json | 2 +- .../documents/composer.v1.json | 12 +- .../documents/composer.v1beta1.json | 35 +- .../documents/compute.alpha.json | 371 +-- .../documents/compute.beta.json | 111 +- .../discovery_cache/documents/compute.v1.json | 355 +-- .../documents/contactcenterinsights.v1.json | 2 +- .../documents/container.v1.json | 2 +- .../documents/container.v1beta1.json | 17 +- .../documents/containeranalysis.v1.json | 2 +- .../documents/containeranalysis.v1alpha1.json | 8 +- .../documents/containeranalysis.v1beta1.json | 2 +- .../documents/content.v2.1.json | 2 +- .../discovery_cache/documents/content.v2.json | 2 +- .../documents/customsearch.v1.json | 2 +- .../documents/datapipelines.v1.json | 2 +- .../documents/datastream.v1.json | 2350 ++++++++++++++-- .../documents/datastream.v1alpha1.json | 313 ++- .../documents/dialogflow.v2.json | 41 +- .../documents/dialogflow.v2beta1.json | 35 +- .../documents/dialogflow.v3.json | 10 +- .../documents/dialogflow.v3beta1.json | 10 +- .../documents/digitalassetlinks.v1.json | 2 +- .../discovery_cache/documents/dlp.v2.json | 17 +- .../documents/documentai.v1.json | 2 +- .../documents/documentai.v1beta2.json | 2 +- .../documents/documentai.v1beta3.json | 2 +- .../documents/domainsrdap.v1.json | 2 +- .../discovery_cache/documents/drive.v2.json | 4 +- .../discovery_cache/documents/drive.v3.json | 4 +- .../documents/driveactivity.v2.json | 2 +- .../documents/essentialcontacts.v1.json | 2 +- .../documents/factchecktools.v1alpha1.json | 2 +- .../discovery_cache/documents/fcm.v1.json | 2 +- .../documents/fcmdata.v1beta1.json | 2 +- .../documents/firebase.v1beta1.json | 2 +- .../documents/firebaseappcheck.v1beta.json | 2 +- .../documents/firebasedatabase.v1beta.json | 2 +- .../documents/firebasedynamiclinks.v1.json | 2 +- .../documents/firebasehosting.v1.json | 2 +- .../documents/firebasehosting.v1beta1.json | 2 +- .../documents/firebaseml.v1.json | 2 +- .../documents/firebaseml.v1beta2.json | 2 +- .../discovery_cache/documents/fitness.v1.json | 2 +- .../documents/gameservices.v1.json | 2 +- .../documents/gameservices.v1beta.json | 35 +- .../discovery_cache/documents/gkehub.v1.json | 2 +- .../documents/gkehub.v1alpha.json | 161 +- .../documents/gkehub.v1alpha2.json | 2 +- .../documents/gkehub.v1beta.json | 2 +- .../documents/gkehub.v1beta1.json | 2 +- .../discovery_cache/documents/gmail.v1.json | 2 +- .../documents/gmailpostmastertools.v1.json | 2 +- .../gmailpostmastertools.v1beta1.json | 2 +- .../documents/groupsmigration.v1.json | 2 +- .../documents/groupssettings.v1.json | 2 +- .../documents/healthcare.v1.json | 2 +- .../documents/healthcare.v1beta1.json | 2 +- .../documents/homegraph.v1.json | 2 +- .../documents/iamcredentials.v1.json | 2 +- .../discovery_cache/documents/iap.v1.json | 2 +- .../documents/iap.v1beta1.json | 2 +- .../documents/ideahub.v1alpha.json | 2 +- .../documents/ideahub.v1beta.json | 2 +- .../documents/indexing.v3.json | 2 +- .../discovery_cache/documents/jobs.v3.json | 2 +- .../documents/jobs.v3p1beta1.json | 2 +- .../discovery_cache/documents/jobs.v4.json | 2 +- .../discovery_cache/documents/keep.v1.json | 2 +- .../documents/kgsearch.v1.json | 2 +- .../documents/language.v1.json | 2 +- .../documents/language.v1beta1.json | 2 +- .../documents/language.v1beta2.json | 2 +- .../documents/libraryagent.v1.json | 2 +- .../documents/licensing.v1.json | 2 +- .../documents/localservices.v1.json | 2 +- .../discovery_cache/documents/logging.v2.json | 120 +- .../documents/manufacturers.v1.json | 2 +- .../discovery_cache/documents/ml.v1.json | 12 +- .../documents/monitoring.v1.json | 2 +- .../documents/monitoring.v3.json | 2 +- .../mybusinessaccountmanagement.v1.json | 2 +- .../mybusinessbusinessinformation.v1.json | 2 +- .../documents/mybusinesslodging.v1.json | 2 +- .../documents/mybusinessnotifications.v1.json | 2 +- .../documents/mybusinessplaceactions.v1.json | 2 +- .../documents/mybusinessverifications.v1.json | 2 +- .../documents/networkconnectivity.v1.json | 36 +- .../networkconnectivity.v1alpha1.json | 8 +- .../documents/ondemandscanning.v1.json | 2 +- .../documents/ondemandscanning.v1beta1.json | 2 +- .../documents/orgpolicy.v2.json | 4 +- .../documents/osconfig.v1.json | 43 +- .../documents/osconfig.v1alpha.json | 32 +- .../documents/osconfig.v1beta.json | 13 +- .../discovery_cache/documents/oslogin.v1.json | 2 +- .../documents/oslogin.v1alpha.json | 2 +- .../documents/pagespeedonline.v5.json | 2 +- .../paymentsresellersubscription.v1.json | 2 +- .../discovery_cache/documents/people.v1.json | 2 +- .../documents/playcustomapp.v1.json | 2 +- .../documents/policysimulator.v1.json | 2 +- .../documents/policysimulator.v1beta1.json | 2 +- .../documents/privateca.v1.json | 2 +- .../documents/privateca.v1beta1.json | 2 +- .../documents/prod_tt_sasportal.v1alpha1.json | 2 +- .../discovery_cache/documents/pubsub.v1.json | 10 +- .../documents/pubsub.v1beta1a.json | 2 +- .../documents/pubsub.v1beta2.json | 6 +- .../documents/pubsublite.v1.json | 2 +- .../documents/realtimebidding.v1.json | 2 +- .../documents/realtimebidding.v1alpha.json | 2 +- .../documents/recommender.v1.json | 2 +- .../documents/recommender.v1beta1.json | 2 +- .../discovery_cache/documents/redis.v1.json | 2 +- .../documents/redis.v1beta1.json | 2 +- .../documents/reseller.v1.json | 2 +- .../documents/resourcesettings.v1.json | 2 +- .../discovery_cache/documents/retail.v2.json | 6 +- .../documents/retail.v2alpha.json | 8 +- .../documents/retail.v2beta.json | 6 +- .../discovery_cache/documents/run.v1.json | 4 +- .../documents/run.v1alpha1.json | 4 +- .../documents/runtimeconfig.v1.json | 2 +- .../documents/runtimeconfig.v1beta1.json | 2 +- .../documents/searchconsole.v1.json | 2 +- .../documents/servicecontrol.v1.json | 2 +- .../documents/servicecontrol.v2.json | 2 +- .../documents/servicemanagement.v1.json | 2 +- .../documents/servicenetworking.v1.json | 2 +- .../documents/servicenetworking.v1beta.json | 2 +- .../discovery_cache/documents/slides.v1.json | 2 +- .../documents/storagetransfer.v1.json | 53 +- .../documents/streetviewpublish.v1.json | 2 +- .../documents/tagmanager.v1.json | 2 +- .../documents/tagmanager.v2.json | 2 +- .../discovery_cache/documents/tasks.v1.json | 2 +- .../discovery_cache/documents/testing.v1.json | 2 +- .../documents/toolresults.v1beta3.json | 2 +- .../discovery_cache/documents/tpu.v1.json | 2 +- .../documents/tpu.v1alpha1.json | 2 +- .../documents/tpu.v2alpha1.json | 2 +- .../discovery_cache/documents/vault.v1.json | 2 +- .../documents/versionhistory.v1.json | 2 +- .../discovery_cache/documents/vision.v1.json | 2 +- .../documents/vision.v1p1beta1.json | 2 +- .../documents/vision.v1p2beta1.json | 2 +- .../documents/webfonts.v1.json | 2 +- .../documents/workflowexecutions.v1.json | 2 +- .../documents/workflowexecutions.v1beta.json | 2 +- .../discovery_cache/documents/youtube.v3.json | 2 +- .../documents/youtubeAnalytics.v2.json | 2 +- .../documents/youtubereporting.v1.json | 2 +- 321 files changed, 14478 insertions(+), 1490 deletions(-) create mode 100644 docs/dyn/clouddeploy_v1.html create mode 100644 docs/dyn/clouddeploy_v1.projects.html create mode 100644 docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.html create mode 100644 docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.releases.html create mode 100644 docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.releases.rollouts.html create mode 100644 docs/dyn/clouddeploy_v1.projects.locations.html create mode 100644 docs/dyn/clouddeploy_v1.projects.locations.operations.html create mode 100644 docs/dyn/clouddeploy_v1.projects.locations.targets.html create mode 100644 docs/dyn/datastream_v1.projects.locations.connectionProfiles.html create mode 100644 docs/dyn/datastream_v1.projects.locations.privateConnections.html create mode 100644 docs/dyn/datastream_v1.projects.locations.privateConnections.routes.html create mode 100644 docs/dyn/datastream_v1.projects.locations.streams.html create mode 100644 docs/dyn/datastream_v1.projects.locations.streams.objects.html create mode 100644 docs/dyn/datastream_v1alpha1.projects.locations.streams.objects.html create mode 100644 docs/dyn/gkehub_v1alpha.projects.locations.fleets.html create mode 100644 googleapiclient/discovery_cache/documents/clouddeploy.v1.json diff --git a/docs/dyn/androidmanagement_v1.enterprises.devices.html b/docs/dyn/androidmanagement_v1.enterprises.devices.html index 446fa0a1375..f25710eb0b8 100644 --- a/docs/dyn/androidmanagement_v1.enterprises.devices.html +++ b/docs/dyn/androidmanagement_v1.enterprises.devices.html @@ -396,7 +396,7 @@
Close httplib2 connections.
+ +Gets a docker image.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Lists docker images.
@@ -89,6 +92,33 @@Close httplib2 connections.+
get(name, x__xgafv=None)
+ Gets a docker image. + +Args: + name: string, Required. The name of the docker images. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # DockerImage represents a docker artifact. The following fields are returned as untyped metadata in the Version resource, using camelcase keys (i.e. metadata.imageSizeBytes): * imageSizeBytes * mediaType * buildTime + "buildTime": "A String", # The time this image was built. This field is returned as the 'metadata.buildTime' field in the Version resource. The build time is returned to the client as an RFC 3339 string, which can be easily used with the JavaScript Date constructor. + "imageSizeBytes": "A String", # Calculated size of the image. This field is returned as the 'metadata.imageSizeBytes' field in the Version resource. + "mediaType": "A String", # Media type of this image, e.g. "application/vnd.docker.distribution.manifest.v2+json". This field is returned as the 'metadata.mediaType' field in the Version resource. + "name": "A String", # Required. registry_location, project_id, repository_name and image id forms a unique image name:`projects//locations//repository//dockerImages/`. For example, "projects/test-project/locations/us-west4/repositories/test-repo/dockerImages/ nginx@sha256:e9954c1fc875017be1c3e36eca16be2d9e9bccc4bf072163515467d6a823c7cf", where "us-west4" is the registry_location, "test-project" is the project_id, "test-repo" is the repository_name and "nginx@sha256:e9954c1fc875017be1c3e36eca16be2d9e9bccc4bf072163515467d6a823c7cf" is the image's digest. + "tags": [ # Tags attached to this image. + "A String", + ], + "uploadTime": "A String", # Time the image was uploaded. + "uri": "A String", # Required. URL to access the image. Example: us-west4-docker.pkg.dev/test-project/test-repo/nginx@sha256:e9954c1fc875017be1c3e36eca16be2d9e9bccc4bf072163515467d6a823c7cf +}+
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Lists docker images. diff --git a/docs/dyn/artifactregistry_v1beta1.projects.locations.repositories.html b/docs/dyn/artifactregistry_v1beta1.projects.locations.repositories.html index 8d5d7a17e6e..e8b8c9e5a79 100644 --- a/docs/dyn/artifactregistry_v1beta1.projects.locations.repositories.html +++ b/docs/dyn/artifactregistry_v1beta1.projects.locations.repositories.html @@ -239,7 +239,7 @@Method Details
Args: resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required) - options_requestedPolicyVersion: integer, Optional. The policy format version to be returned. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + options_requestedPolicyVersion: integer, Optional. The maximum policy version that will be used to format the policy. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional role bindings must specify version 3. Policies with no conditional role bindings may specify any valid value or leave the field unset. The policy in the response might use the policy version that you specified, or it might use a lower policy version. For example, if you specify version 3, but the policy has no conditional role bindings, the response uses version 1. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). x__xgafv: string, V1 error format. Allowed values 1 - v1 error format diff --git a/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.html b/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.html index 687a70fd433..691fad9a7a9 100644 --- a/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.html +++ b/docs/dyn/artifactregistry_v1beta2.projects.locations.repositories.html @@ -267,7 +267,7 @@Method Details
Args: resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required) - options_requestedPolicyVersion: integer, Optional. The policy format version to be returned. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + options_requestedPolicyVersion: integer, Optional. The maximum policy version that will be used to format the policy. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional role bindings must specify version 3. Policies with no conditional role bindings may specify any valid value or leave the field unset. The policy in the response might use the policy version that you specified, or it might use a lower policy version. For example, if you specify version 3, but the policy has no conditional role bindings, the response uses version 1. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). x__xgafv: string, V1 error format. Allowed values 1 - v1 error format diff --git a/docs/dyn/clouddeploy_v1.html b/docs/dyn/clouddeploy_v1.html new file mode 100644 index 00000000000..602d5e3827a --- /dev/null +++ b/docs/dyn/clouddeploy_v1.html @@ -0,0 +1,111 @@ + + + +Google Cloud Deploy API
+Instance Methods
++
+projects()
+Returns the projects Resource.
+ ++
+close()
Close httplib2 connections.
+ +Create a BatchHttpRequest object based on the discovery document.
+Method Details
+++ +close()
+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/clouddeploy_v1.projects.html b/docs/dyn/clouddeploy_v1.projects.html new file mode 100644 index 00000000000..f5decb4d6cd --- /dev/null +++ b/docs/dyn/clouddeploy_v1.projects.html @@ -0,0 +1,91 @@ + + + +new_batch_http_request()
+Create a BatchHttpRequest object based on the discovery document. + + Args: + callback: callable, A callback to be called for each response, of the + form callback(id, response, exception). The first parameter is the + request id, and the second is the deserialized response object. The + third is an apiclient.errors.HttpError exception object if an HTTP + error occurred while processing the request, or None if no error + occurred. + + Returns: + A BatchHttpRequest object based on the discovery document. ++Google Cloud Deploy API . projects
+Instance Methods
++
+locations()
+Returns the locations Resource.
+ ++
+close()
Close httplib2 connections.
+Method Details
+++ + \ No newline at end of file diff --git a/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.html b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.html new file mode 100644 index 00000000000..0b8ee9133e2 --- /dev/null +++ b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.html @@ -0,0 +1,612 @@ + + + +close()
+Close httplib2 connections.+Google Cloud Deploy API . projects . locations . deliveryPipelines
+Instance Methods
++
+releases()
+Returns the releases Resource.
+ ++
+close()
Close httplib2 connections.
++
+create(parent, body=None, deliveryPipelineId=None, requestId=None, validateOnly=None, x__xgafv=None)
Creates a new DeliveryPipeline in a given project and location.
+ +Deletes a single DeliveryPipeline.
+ +Gets details of a single DeliveryPipeline.
++
+getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
++
+list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
Lists DeliveryPipelines in a given project and location.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+ +Updates the parameters of a single DeliveryPipeline.
++
+setIamPolicy(resource, body=None, x__xgafv=None)
Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
++
+testIamPermissions(resource, body=None, x__xgafv=None)
Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, deliveryPipelineId=None, requestId=None, validateOnly=None, x__xgafv=None)
+Creates a new DeliveryPipeline in a given project and location. + +Args: + parent: string, Required. The parent collection in which the `DeliveryPipeline` should be created. Format should be projects/{project_id}/locations/{location_name}. (required) + body: object, The request body. + The object takes the form of: + +{ # A `DeliveryPipeline` resource in the Google Cloud Deploy API. A `DeliveryPipeline` defines a pipeline through which a Skaffold configuration can progress. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "condition": { # PipelineCondition contains all conditions relevant to a Delivery Pipeline. # Output only. Information around the state of the Delivery Pipeline. + "pipelineReadyCondition": { # PipelineReadyCondition contains information around the status of the Pipeline. # Details around the Pipeline's overall status. + "status": True or False, # True if the Pipeline is in a valid state. Otherwise at least one condition in `PipelineCondition` is in an invalid state. Iterate over those conditions and see which condition(s) has status = false to find out what is wrong with the Pipeline. + "updateTime": "A String", # Last time the condition was updated. + }, + "targetsPresentCondition": { # TargetsPresentCondition contains information on any Targets defined in the Delivery Pipeline that do not actually exist. # Detalis around targets enumerated in the pipeline. + "missingTargets": [ # The list of Target names that are missing. For example, projects/{project_id}/locations/{location_name}/targets/{target_name}. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + }, + "createTime": "A String", # Output only. Time at which the pipeline was created. + "description": "A String", # Description of the `DeliveryPipeline`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `DeliveryPipeline`. Format is projects/{project}/ locations/{location}/deliveryPipelines/a-z{0,62}. + "serialPipeline": { # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. + "stages": [ # Each stage specifies configuration for a `Target`. The ordering of this list defines the promotion flow. + { # Stage specifies a location to which to deploy. + "profiles": [ # Skaffold profiles to use when rendering the manifest for this stage's `Target`. + "A String", + ], + "targetId": "A String", # The target_id to which this stage points. This field refers exclusively to the last segment of a target name. For example, this field would just be `my-target` (rather than `projects/project/deliveryPipelines/pipeline/targets/my-target`). The parent `DeliveryPipeline` of the `Target` is inferred to be the parent `DeliveryPipeline` of the `Release` in which this `Stage` lives. + }, + ], + }, + "uid": "A String", # Output only. Unique identifier of the `DeliveryPipeline`. + "updateTime": "A String", # Output only. Most recent time at which the pipeline was updated. +} + + deliveryPipelineId: string, Required. ID of the `DeliveryPipeline`. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + validateOnly: boolean, Optional. If set to true, the request is validated and the user is provided with an expected result, but no actual change is made. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, allowMissing=None, etag=None, force=None, requestId=None, validateOnly=None, x__xgafv=None)
+Deletes a single DeliveryPipeline. + +Args: + name: string, Required. The name of the `DeliveryPipeline` to delete. Format should be projects/{project_id}/locations/{location_name}/deliveryPipelines/{pipeline_name}. (required) + allowMissing: boolean, Optional. If set to true, then deleting an already deleted or non-existing `DeliveryPipeline` will succeed. + etag: string, Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + force: boolean, Optional. If set to true, all child resources under this pipeline will also be deleted. Otherwise, the request will only work if the pipeline has no child resources. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes after the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + validateOnly: boolean, Optional. If set, validate the request and preview the review, but do not actually post it. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)
+Gets details of a single DeliveryPipeline. + +Args: + name: string, Required. Name of the `DeliveryPipeline`. Format must be projects/{project_id}/locations/{location_name}/deliveryPipelines/{pipeline_name}. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A `DeliveryPipeline` resource in the Google Cloud Deploy API. A `DeliveryPipeline` defines a pipeline through which a Skaffold configuration can progress. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "condition": { # PipelineCondition contains all conditions relevant to a Delivery Pipeline. # Output only. Information around the state of the Delivery Pipeline. + "pipelineReadyCondition": { # PipelineReadyCondition contains information around the status of the Pipeline. # Details around the Pipeline's overall status. + "status": True or False, # True if the Pipeline is in a valid state. Otherwise at least one condition in `PipelineCondition` is in an invalid state. Iterate over those conditions and see which condition(s) has status = false to find out what is wrong with the Pipeline. + "updateTime": "A String", # Last time the condition was updated. + }, + "targetsPresentCondition": { # TargetsPresentCondition contains information on any Targets defined in the Delivery Pipeline that do not actually exist. # Detalis around targets enumerated in the pipeline. + "missingTargets": [ # The list of Target names that are missing. For example, projects/{project_id}/locations/{location_name}/targets/{target_name}. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + }, + "createTime": "A String", # Output only. Time at which the pipeline was created. + "description": "A String", # Description of the `DeliveryPipeline`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `DeliveryPipeline`. Format is projects/{project}/ locations/{location}/deliveryPipelines/a-z{0,62}. + "serialPipeline": { # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. + "stages": [ # Each stage specifies configuration for a `Target`. The ordering of this list defines the promotion flow. + { # Stage specifies a location to which to deploy. + "profiles": [ # Skaffold profiles to use when rendering the manifest for this stage's `Target`. + "A String", + ], + "targetId": "A String", # The target_id to which this stage points. This field refers exclusively to the last segment of a target name. For example, this field would just be `my-target` (rather than `projects/project/deliveryPipelines/pipeline/targets/my-target`). The parent `DeliveryPipeline` of the `Target` is inferred to be the parent `DeliveryPipeline` of the `Release` in which this `Stage` lives. + }, + ], + }, + "uid": "A String", # Output only. Unique identifier of the `DeliveryPipeline`. + "updateTime": "A String", # Output only. Most recent time at which the pipeline was updated. +}+++ +getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set. + +Args: + resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required) + options_requestedPolicyVersion: integer, Optional. The maximum policy version that will be used to format the policy. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional role bindings must specify version 3. Policies with no conditional role bindings may specify any valid value or leave the field unset. The policy in the response might use the policy version that you specified, or it might use a lower policy version. For example, if you specify version 3, but the policy has no conditional role bindings, the response uses version 1. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members`, or principals, to a single `role`. Principals can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') etag: BwWWja0YfJA= version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). + "auditConfigs": [ # Specifies cloud audit logging configuration for this policy. + { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging. + "auditLogConfigs": [ # The configuration for logging of each type of permission. + { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging. + "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members. + "A String", + ], + "logType": "A String", # The log type that this config enables. + }, + ], + "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services. + }, + ], + "bindings": [ # Associates a list of `members`, or principals, with a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one principal. The `bindings` in a `Policy` can refer to up to 1,500 principals; up to 250 of these principals can be Google groups. Each occurrence of a principal counts towards these limits. For example, if the `bindings` grant 50 different roles to `user:alice@example.com`, and not to any other principal, then you can add another 1,450 principals to the `bindings` in the `Policy`. + { # Associates `members`, or principals, with a `role`. + "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI. + "expression": "A String", # Textual representation of an expression in Common Expression Language syntax. + "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file. + "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression. + }, + "members": [ # Specifies the principals requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. + "A String", + ], + "role": "A String", # Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`. + }, + ], + "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. + "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). +}+++ +list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists DeliveryPipelines in a given project and location. + +Args: + parent: string, Required. The parent, which owns this collection of pipelines. Format must be projects/{project_id}/locations/{location_name}. (required) + filter: string, Filter builds to be returned. See https://google.aip.dev/160 for more details. + orderBy: string, Field to sort by. See https://google.aip.dev/132#ordering for more details. + pageSize: integer, The maximum number of pipelines to return. The service may return fewer than this value. If unspecified, at most 50 pipelines will be returned. The maximum value is 1000; values above 1000 will be set to 1000. + pageToken: string, A page token, received from a previous `ListDeliveryPipelines` call. Provide this to retrieve the subsequent page. When paginating, all other provided parameters match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response object from `ListDeliveryPipelines`. + "deliveryPipelines": [ # The `DeliveryPipeline` objects. + { # A `DeliveryPipeline` resource in the Google Cloud Deploy API. A `DeliveryPipeline` defines a pipeline through which a Skaffold configuration can progress. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "condition": { # PipelineCondition contains all conditions relevant to a Delivery Pipeline. # Output only. Information around the state of the Delivery Pipeline. + "pipelineReadyCondition": { # PipelineReadyCondition contains information around the status of the Pipeline. # Details around the Pipeline's overall status. + "status": True or False, # True if the Pipeline is in a valid state. Otherwise at least one condition in `PipelineCondition` is in an invalid state. Iterate over those conditions and see which condition(s) has status = false to find out what is wrong with the Pipeline. + "updateTime": "A String", # Last time the condition was updated. + }, + "targetsPresentCondition": { # TargetsPresentCondition contains information on any Targets defined in the Delivery Pipeline that do not actually exist. # Detalis around targets enumerated in the pipeline. + "missingTargets": [ # The list of Target names that are missing. For example, projects/{project_id}/locations/{location_name}/targets/{target_name}. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + }, + "createTime": "A String", # Output only. Time at which the pipeline was created. + "description": "A String", # Description of the `DeliveryPipeline`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `DeliveryPipeline`. Format is projects/{project}/ locations/{location}/deliveryPipelines/a-z{0,62}. + "serialPipeline": { # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. + "stages": [ # Each stage specifies configuration for a `Target`. The ordering of this list defines the promotion flow. + { # Stage specifies a location to which to deploy. + "profiles": [ # Skaffold profiles to use when rendering the manifest for this stage's `Target`. + "A String", + ], + "targetId": "A String", # The target_id to which this stage points. This field refers exclusively to the last segment of a target name. For example, this field would just be `my-target` (rather than `projects/project/deliveryPipelines/pipeline/targets/my-target`). The parent `DeliveryPipeline` of the `Target` is inferred to be the parent `DeliveryPipeline` of the `Release` in which this `Stage` lives. + }, + ], + }, + "uid": "A String", # Output only. Unique identifier of the `DeliveryPipeline`. + "updateTime": "A String", # Output only. Most recent time at which the pipeline was updated. + }, + ], + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "unreachable": [ # Locations that could not be reached. + "A String", + ], +}+++ +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ +patch(name, allowMissing=None, body=None, requestId=None, updateMask=None, validateOnly=None, x__xgafv=None)
+Updates the parameters of a single DeliveryPipeline. + +Args: + name: string, Optional. Name of the `DeliveryPipeline`. Format is projects/{project}/ locations/{location}/deliveryPipelines/a-z{0,62}. (required) + body: object, The request body. + The object takes the form of: + +{ # A `DeliveryPipeline` resource in the Google Cloud Deploy API. A `DeliveryPipeline` defines a pipeline through which a Skaffold configuration can progress. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "condition": { # PipelineCondition contains all conditions relevant to a Delivery Pipeline. # Output only. Information around the state of the Delivery Pipeline. + "pipelineReadyCondition": { # PipelineReadyCondition contains information around the status of the Pipeline. # Details around the Pipeline's overall status. + "status": True or False, # True if the Pipeline is in a valid state. Otherwise at least one condition in `PipelineCondition` is in an invalid state. Iterate over those conditions and see which condition(s) has status = false to find out what is wrong with the Pipeline. + "updateTime": "A String", # Last time the condition was updated. + }, + "targetsPresentCondition": { # TargetsPresentCondition contains information on any Targets defined in the Delivery Pipeline that do not actually exist. # Detalis around targets enumerated in the pipeline. + "missingTargets": [ # The list of Target names that are missing. For example, projects/{project_id}/locations/{location_name}/targets/{target_name}. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + }, + "createTime": "A String", # Output only. Time at which the pipeline was created. + "description": "A String", # Description of the `DeliveryPipeline`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `DeliveryPipeline`. Format is projects/{project}/ locations/{location}/deliveryPipelines/a-z{0,62}. + "serialPipeline": { # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. + "stages": [ # Each stage specifies configuration for a `Target`. The ordering of this list defines the promotion flow. + { # Stage specifies a location to which to deploy. + "profiles": [ # Skaffold profiles to use when rendering the manifest for this stage's `Target`. + "A String", + ], + "targetId": "A String", # The target_id to which this stage points. This field refers exclusively to the last segment of a target name. For example, this field would just be `my-target` (rather than `projects/project/deliveryPipelines/pipeline/targets/my-target`). The parent `DeliveryPipeline` of the `Target` is inferred to be the parent `DeliveryPipeline` of the `Release` in which this `Stage` lives. + }, + ], + }, + "uid": "A String", # Output only. Unique identifier of the `DeliveryPipeline`. + "updateTime": "A String", # Output only. Most recent time at which the pipeline was updated. +} + + allowMissing: boolean, Optional. If set to true, updating a `DeliveryPipeline` that does not exist will result in the creation of a new `DeliveryPipeline`. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + updateMask: string, Required. Field mask is used to specify the fields to be overwritten in the `DeliveryPipeline` resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then all fields will be overwritten. + validateOnly: boolean, Optional. If set to true, the request is validated and the user is provided with an expected result, but no actual change is made. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +setIamPolicy(resource, body=None, x__xgafv=None)
+Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors. + +Args: + resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required) + body: object, The request body. + The object takes the form of: + +{ # Request message for `SetIamPolicy` method. + "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members`, or principals, to a single `role`. Principals can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') etag: BwWWja0YfJA= version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the `resource`. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Cloud Platform services (such as Projects) might reject them. + "auditConfigs": [ # Specifies cloud audit logging configuration for this policy. + { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging. + "auditLogConfigs": [ # The configuration for logging of each type of permission. + { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging. + "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members. + "A String", + ], + "logType": "A String", # The log type that this config enables. + }, + ], + "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services. + }, + ], + "bindings": [ # Associates a list of `members`, or principals, with a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one principal. The `bindings` in a `Policy` can refer to up to 1,500 principals; up to 250 of these principals can be Google groups. Each occurrence of a principal counts towards these limits. For example, if the `bindings` grant 50 different roles to `user:alice@example.com`, and not to any other principal, then you can add another 1,450 principals to the `bindings` in the `Policy`. + { # Associates `members`, or principals, with a `role`. + "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI. + "expression": "A String", # Textual representation of an expression in Common Expression Language syntax. + "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file. + "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression. + }, + "members": [ # Specifies the principals requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. + "A String", + ], + "role": "A String", # Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`. + }, + ], + "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. + "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + }, + "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used: `paths: "bindings, etag"` +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members`, or principals, to a single `role`. Principals can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') etag: BwWWja0YfJA= version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). + "auditConfigs": [ # Specifies cloud audit logging configuration for this policy. + { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging. + "auditLogConfigs": [ # The configuration for logging of each type of permission. + { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging. + "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members. + "A String", + ], + "logType": "A String", # The log type that this config enables. + }, + ], + "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services. + }, + ], + "bindings": [ # Associates a list of `members`, or principals, with a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one principal. The `bindings` in a `Policy` can refer to up to 1,500 principals; up to 250 of these principals can be Google groups. Each occurrence of a principal counts towards these limits. For example, if the `bindings` grant 50 different roles to `user:alice@example.com`, and not to any other principal, then you can add another 1,450 principals to the `bindings` in the `Policy`. + { # Associates `members`, or principals, with a `role`. + "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI. + "expression": "A String", # Textual representation of an expression in Common Expression Language syntax. + "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file. + "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression. + }, + "members": [ # Specifies the principals requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. + "A String", + ], + "role": "A String", # Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`. + }, + ], + "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. + "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). +}+++ + \ No newline at end of file diff --git a/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.releases.html b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.releases.html new file mode 100644 index 00000000000..94e4d5e1633 --- /dev/null +++ b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.releases.html @@ -0,0 +1,539 @@ + + + +testIamPermissions(resource, body=None, x__xgafv=None)
+Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning. + +Args: + resource: string, REQUIRED: The resource for which the policy detail is being requested. See the operation documentation for the appropriate value for this field. (required) + body: object, The request body. + The object takes the form of: + +{ # Request message for `TestIamPermissions` method. + "permissions": [ # The set of permissions to check for the `resource`. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions). + "A String", + ], +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response message for `TestIamPermissions` method. + "permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is allowed. + "A String", + ], +}+Google Cloud Deploy API . projects . locations . deliveryPipelines . releases
+Instance Methods
++
+rollouts()
+Returns the rollouts Resource.
+ ++
+close()
Close httplib2 connections.
++
+create(parent, body=None, releaseId=None, requestId=None, validateOnly=None, x__xgafv=None)
Creates a new Release in a given project and location.
+ +Gets details of a single Release.
++
+list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
Lists Releases in a given project and location.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, releaseId=None, requestId=None, validateOnly=None, x__xgafv=None)
+Creates a new Release in a given project and location. + +Args: + parent: string, Required. The parent collection in which the `Release` should be created. Format should be projects/{project_id}/locations/{location_name}/deliveryPipelines/{pipeline_name}. (required) + body: object, The request body. + The object takes the form of: + +{ # A `Release` resource in the Google Cloud Deploy API. A `Release` defines a specific Skaffold configuration instance that can be deployed. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "buildArtifacts": [ # List of artifacts to pass through to Skaffold command. + { # Description of an a image to use during Skaffold rendering. + "image": "A String", # Image name in Skaffold configuration. + "tag": "A String", # Image tag to use. This will generally be the full path to an image, such as "gcr.io/my-project/busybox:1.2.3" or "gcr.io/my-project/busybox@sha256:abc123". + }, + ], + "createTime": "A String", # Output only. Time at which the `Release` was created. + "deliveryPipelineSnapshot": { # A `DeliveryPipeline` resource in the Google Cloud Deploy API. A `DeliveryPipeline` defines a pipeline through which a Skaffold configuration can progress. # Output only. Snapshot of the parent pipeline taken at release creation time. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "condition": { # PipelineCondition contains all conditions relevant to a Delivery Pipeline. # Output only. Information around the state of the Delivery Pipeline. + "pipelineReadyCondition": { # PipelineReadyCondition contains information around the status of the Pipeline. # Details around the Pipeline's overall status. + "status": True or False, # True if the Pipeline is in a valid state. Otherwise at least one condition in `PipelineCondition` is in an invalid state. Iterate over those conditions and see which condition(s) has status = false to find out what is wrong with the Pipeline. + "updateTime": "A String", # Last time the condition was updated. + }, + "targetsPresentCondition": { # TargetsPresentCondition contains information on any Targets defined in the Delivery Pipeline that do not actually exist. # Detalis around targets enumerated in the pipeline. + "missingTargets": [ # The list of Target names that are missing. For example, projects/{project_id}/locations/{location_name}/targets/{target_name}. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + }, + "createTime": "A String", # Output only. Time at which the pipeline was created. + "description": "A String", # Description of the `DeliveryPipeline`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `DeliveryPipeline`. Format is projects/{project}/ locations/{location}/deliveryPipelines/a-z{0,62}. + "serialPipeline": { # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. + "stages": [ # Each stage specifies configuration for a `Target`. The ordering of this list defines the promotion flow. + { # Stage specifies a location to which to deploy. + "profiles": [ # Skaffold profiles to use when rendering the manifest for this stage's `Target`. + "A String", + ], + "targetId": "A String", # The target_id to which this stage points. This field refers exclusively to the last segment of a target name. For example, this field would just be `my-target` (rather than `projects/project/deliveryPipelines/pipeline/targets/my-target`). The parent `DeliveryPipeline` of the `Target` is inferred to be the parent `DeliveryPipeline` of the `Release` in which this `Stage` lives. + }, + ], + }, + "uid": "A String", # Output only. Unique identifier of the `DeliveryPipeline`. + "updateTime": "A String", # Output only. Most recent time at which the pipeline was updated. + }, + "description": "A String", # Description of the `Release`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Release`. Format is projects/{project}/ locations/{location}/deliveryPipelines/{deliveryPipeline}/ releases/a-z{0,62}. + "renderEndTime": "A String", # Output only. Time at which the render completed. + "renderStartTime": "A String", # Output only. Time at which the render began. + "renderState": "A String", # Output only. Current state of the render operation. + "skaffoldConfigPath": "A String", # Filepath of the Skaffold config inside of the config URI. + "skaffoldConfigUri": "A String", # Cloud Storage URI of tar.gz archive containing Skaffold configuration. + "skaffoldVersion": "A String", # The Skaffold version to use when operating on this release, such as "1.20.0". Not all versions are valid; Google Cloud Deploy supports a specific set of versions. If unset, the most recent supported Skaffold version will be used. + "targetArtifacts": { # Output only. Map from target ID to the target artifacts created during the render operation. + "a_key": { # The artifacts produced by a target render operation. + "artifactUri": "A String", # Output only. URI of a directory containing the artifacts. This contains deployment configuration used by Skaffold during a rollout, and all paths are relative to this location. + "manifestPath": "A String", # Output only. File path of the rendered manifest relative to the URI. + "skaffoldConfigPath": "A String", # Output only. File path of the resolved Skaffold configuration relative to the URI. + }, + }, + "targetRenders": { # Output only. Map from target ID to details of the render operation for that target. + "a_key": { # Details of rendering for a single target. + "renderingBuild": "A String", # Output only. The resource name of the Cloud Build `Build` object that is used to render the manifest for this target. Format is `projects/{project}/locations/{location}/builds/{build}`. + "renderingState": "A String", # Output only. Current state of the render operation for this Target. + }, + }, + "targetSnapshots": [ # Output only. Snapshot of the parent pipeline's targets taken at release creation time. + { # A `Target` resource in the Google Cloud Deploy API. A `Target` defines a location to which a Skaffold configuration can be deployed. + "annotations": { # Optional. User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "createTime": "A String", # Output only. Time at which the `Target` was created. + "description": "A String", # Optional. Description of the `Target`. Max length is 255 characters. + "etag": "A String", # Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "executionConfigs": [ # Configurations for all execution that relates to this `Target`. Each `ExecutionEnvironmentUsage` value may only be used in a single configuration; using the same value multiple times is an error. When one or more configurations are specified, they must include the `RENDER` and `DEPLOY` `ExecutionEnvironmentUsage` values. When no configurations are specified, execution will use the default specified in `DefaultPool`. + { # Configuration of the environment to use when calling Skaffold. + "defaultPool": { # Execution using the default Cloud Build pool. # Optional. Use default Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + }, + "privatePool": { # Execution using a private Cloud Build pool. # Optional. Use private Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + "workerPool": "A String", # Required. Resource name of the Cloud Build worker pool to use. The format is `projects/{project}/locations/{location}/workerPools/{pool}`. + }, + "usages": [ # Required. Usages when this configuration should be applied. + "A String", + ], + }, + ], + "gke": { # Information specifying a GKE Cluster. # Information specifying a GKE Cluster. + "cluster": "A String", # Information specifying a GKE Cluster. Format is `projects/{project_id}/locations/{location_id}/clusters/{cluster_id}. + }, + "labels": { # Optional. Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Target`. Format is projects/{project}/locations/{location}/ deliveryPipelines/{deliveryPipeline}/targets/a-z{0,62}. + "requireApproval": True or False, # Optional. Whether or not the `Target` requires approval. + "targetId": "A String", # Output only. Resource id of the `Target`. + "uid": "A String", # Output only. Unique identifier of the `Target`. + "updateTime": "A String", # Output only. Most recent time at which the `Target` was updated. + }, + ], + "uid": "A String", # Output only. Unique identifier of the `Release`. +} + + releaseId: string, Required. ID of the `Release`. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + validateOnly: boolean, Optional. If set to true, the request is validated and the user is provided with an expected result, but no actual change is made. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)
+Gets details of a single Release. + +Args: + name: string, Required. Name of the `Release`. Format must be projects/{project_id}/locations/{location_name}/deliveryPipelines/{pipeline_name}/releases/{release_name}. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A `Release` resource in the Google Cloud Deploy API. A `Release` defines a specific Skaffold configuration instance that can be deployed. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "buildArtifacts": [ # List of artifacts to pass through to Skaffold command. + { # Description of an a image to use during Skaffold rendering. + "image": "A String", # Image name in Skaffold configuration. + "tag": "A String", # Image tag to use. This will generally be the full path to an image, such as "gcr.io/my-project/busybox:1.2.3" or "gcr.io/my-project/busybox@sha256:abc123". + }, + ], + "createTime": "A String", # Output only. Time at which the `Release` was created. + "deliveryPipelineSnapshot": { # A `DeliveryPipeline` resource in the Google Cloud Deploy API. A `DeliveryPipeline` defines a pipeline through which a Skaffold configuration can progress. # Output only. Snapshot of the parent pipeline taken at release creation time. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "condition": { # PipelineCondition contains all conditions relevant to a Delivery Pipeline. # Output only. Information around the state of the Delivery Pipeline. + "pipelineReadyCondition": { # PipelineReadyCondition contains information around the status of the Pipeline. # Details around the Pipeline's overall status. + "status": True or False, # True if the Pipeline is in a valid state. Otherwise at least one condition in `PipelineCondition` is in an invalid state. Iterate over those conditions and see which condition(s) has status = false to find out what is wrong with the Pipeline. + "updateTime": "A String", # Last time the condition was updated. + }, + "targetsPresentCondition": { # TargetsPresentCondition contains information on any Targets defined in the Delivery Pipeline that do not actually exist. # Detalis around targets enumerated in the pipeline. + "missingTargets": [ # The list of Target names that are missing. For example, projects/{project_id}/locations/{location_name}/targets/{target_name}. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + }, + "createTime": "A String", # Output only. Time at which the pipeline was created. + "description": "A String", # Description of the `DeliveryPipeline`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `DeliveryPipeline`. Format is projects/{project}/ locations/{location}/deliveryPipelines/a-z{0,62}. + "serialPipeline": { # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. + "stages": [ # Each stage specifies configuration for a `Target`. The ordering of this list defines the promotion flow. + { # Stage specifies a location to which to deploy. + "profiles": [ # Skaffold profiles to use when rendering the manifest for this stage's `Target`. + "A String", + ], + "targetId": "A String", # The target_id to which this stage points. This field refers exclusively to the last segment of a target name. For example, this field would just be `my-target` (rather than `projects/project/deliveryPipelines/pipeline/targets/my-target`). The parent `DeliveryPipeline` of the `Target` is inferred to be the parent `DeliveryPipeline` of the `Release` in which this `Stage` lives. + }, + ], + }, + "uid": "A String", # Output only. Unique identifier of the `DeliveryPipeline`. + "updateTime": "A String", # Output only. Most recent time at which the pipeline was updated. + }, + "description": "A String", # Description of the `Release`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Release`. Format is projects/{project}/ locations/{location}/deliveryPipelines/{deliveryPipeline}/ releases/a-z{0,62}. + "renderEndTime": "A String", # Output only. Time at which the render completed. + "renderStartTime": "A String", # Output only. Time at which the render began. + "renderState": "A String", # Output only. Current state of the render operation. + "skaffoldConfigPath": "A String", # Filepath of the Skaffold config inside of the config URI. + "skaffoldConfigUri": "A String", # Cloud Storage URI of tar.gz archive containing Skaffold configuration. + "skaffoldVersion": "A String", # The Skaffold version to use when operating on this release, such as "1.20.0". Not all versions are valid; Google Cloud Deploy supports a specific set of versions. If unset, the most recent supported Skaffold version will be used. + "targetArtifacts": { # Output only. Map from target ID to the target artifacts created during the render operation. + "a_key": { # The artifacts produced by a target render operation. + "artifactUri": "A String", # Output only. URI of a directory containing the artifacts. This contains deployment configuration used by Skaffold during a rollout, and all paths are relative to this location. + "manifestPath": "A String", # Output only. File path of the rendered manifest relative to the URI. + "skaffoldConfigPath": "A String", # Output only. File path of the resolved Skaffold configuration relative to the URI. + }, + }, + "targetRenders": { # Output only. Map from target ID to details of the render operation for that target. + "a_key": { # Details of rendering for a single target. + "renderingBuild": "A String", # Output only. The resource name of the Cloud Build `Build` object that is used to render the manifest for this target. Format is `projects/{project}/locations/{location}/builds/{build}`. + "renderingState": "A String", # Output only. Current state of the render operation for this Target. + }, + }, + "targetSnapshots": [ # Output only. Snapshot of the parent pipeline's targets taken at release creation time. + { # A `Target` resource in the Google Cloud Deploy API. A `Target` defines a location to which a Skaffold configuration can be deployed. + "annotations": { # Optional. User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "createTime": "A String", # Output only. Time at which the `Target` was created. + "description": "A String", # Optional. Description of the `Target`. Max length is 255 characters. + "etag": "A String", # Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "executionConfigs": [ # Configurations for all execution that relates to this `Target`. Each `ExecutionEnvironmentUsage` value may only be used in a single configuration; using the same value multiple times is an error. When one or more configurations are specified, they must include the `RENDER` and `DEPLOY` `ExecutionEnvironmentUsage` values. When no configurations are specified, execution will use the default specified in `DefaultPool`. + { # Configuration of the environment to use when calling Skaffold. + "defaultPool": { # Execution using the default Cloud Build pool. # Optional. Use default Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + }, + "privatePool": { # Execution using a private Cloud Build pool. # Optional. Use private Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + "workerPool": "A String", # Required. Resource name of the Cloud Build worker pool to use. The format is `projects/{project}/locations/{location}/workerPools/{pool}`. + }, + "usages": [ # Required. Usages when this configuration should be applied. + "A String", + ], + }, + ], + "gke": { # Information specifying a GKE Cluster. # Information specifying a GKE Cluster. + "cluster": "A String", # Information specifying a GKE Cluster. Format is `projects/{project_id}/locations/{location_id}/clusters/{cluster_id}. + }, + "labels": { # Optional. Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Target`. Format is projects/{project}/locations/{location}/ deliveryPipelines/{deliveryPipeline}/targets/a-z{0,62}. + "requireApproval": True or False, # Optional. Whether or not the `Target` requires approval. + "targetId": "A String", # Output only. Resource id of the `Target`. + "uid": "A String", # Output only. Unique identifier of the `Target`. + "updateTime": "A String", # Output only. Most recent time at which the `Target` was updated. + }, + ], + "uid": "A String", # Output only. Unique identifier of the `Release`. +}+++ +list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists Releases in a given project and location. + +Args: + parent: string, Required. The `DeliveryPipeline` which owns this collection of `Release` objects. (required) + filter: string, Optional. Filter builds to be returned. See https://google.aip.dev/160 for more details. + orderBy: string, Optional. Field to sort by. See https://google.aip.dev/132#ordering for more details. + pageSize: integer, Optional. The maximum number of `Release` objects to return. The service may return fewer than this value. If unspecified, at most 50 `Release` objects will be returned. The maximum value is 1000; values above 1000 will be set to 1000. + pageToken: string, Optional. A page token, received from a previous `ListReleases` call. Provide this to retrieve the subsequent page. When paginating, all other provided parameters match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response object from `ListReleases`. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "releases": [ # The `Release` objects. + { # A `Release` resource in the Google Cloud Deploy API. A `Release` defines a specific Skaffold configuration instance that can be deployed. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "buildArtifacts": [ # List of artifacts to pass through to Skaffold command. + { # Description of an a image to use during Skaffold rendering. + "image": "A String", # Image name in Skaffold configuration. + "tag": "A String", # Image tag to use. This will generally be the full path to an image, such as "gcr.io/my-project/busybox:1.2.3" or "gcr.io/my-project/busybox@sha256:abc123". + }, + ], + "createTime": "A String", # Output only. Time at which the `Release` was created. + "deliveryPipelineSnapshot": { # A `DeliveryPipeline` resource in the Google Cloud Deploy API. A `DeliveryPipeline` defines a pipeline through which a Skaffold configuration can progress. # Output only. Snapshot of the parent pipeline taken at release creation time. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "condition": { # PipelineCondition contains all conditions relevant to a Delivery Pipeline. # Output only. Information around the state of the Delivery Pipeline. + "pipelineReadyCondition": { # PipelineReadyCondition contains information around the status of the Pipeline. # Details around the Pipeline's overall status. + "status": True or False, # True if the Pipeline is in a valid state. Otherwise at least one condition in `PipelineCondition` is in an invalid state. Iterate over those conditions and see which condition(s) has status = false to find out what is wrong with the Pipeline. + "updateTime": "A String", # Last time the condition was updated. + }, + "targetsPresentCondition": { # TargetsPresentCondition contains information on any Targets defined in the Delivery Pipeline that do not actually exist. # Detalis around targets enumerated in the pipeline. + "missingTargets": [ # The list of Target names that are missing. For example, projects/{project_id}/locations/{location_name}/targets/{target_name}. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + }, + "createTime": "A String", # Output only. Time at which the pipeline was created. + "description": "A String", # Description of the `DeliveryPipeline`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `DeliveryPipeline`. Format is projects/{project}/ locations/{location}/deliveryPipelines/a-z{0,62}. + "serialPipeline": { # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. # SerialPipeline defines a sequential set of stages for a `DeliveryPipeline`. + "stages": [ # Each stage specifies configuration for a `Target`. The ordering of this list defines the promotion flow. + { # Stage specifies a location to which to deploy. + "profiles": [ # Skaffold profiles to use when rendering the manifest for this stage's `Target`. + "A String", + ], + "targetId": "A String", # The target_id to which this stage points. This field refers exclusively to the last segment of a target name. For example, this field would just be `my-target` (rather than `projects/project/deliveryPipelines/pipeline/targets/my-target`). The parent `DeliveryPipeline` of the `Target` is inferred to be the parent `DeliveryPipeline` of the `Release` in which this `Stage` lives. + }, + ], + }, + "uid": "A String", # Output only. Unique identifier of the `DeliveryPipeline`. + "updateTime": "A String", # Output only. Most recent time at which the pipeline was updated. + }, + "description": "A String", # Description of the `Release`. Max length is 255 characters. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Release`. Format is projects/{project}/ locations/{location}/deliveryPipelines/{deliveryPipeline}/ releases/a-z{0,62}. + "renderEndTime": "A String", # Output only. Time at which the render completed. + "renderStartTime": "A String", # Output only. Time at which the render began. + "renderState": "A String", # Output only. Current state of the render operation. + "skaffoldConfigPath": "A String", # Filepath of the Skaffold config inside of the config URI. + "skaffoldConfigUri": "A String", # Cloud Storage URI of tar.gz archive containing Skaffold configuration. + "skaffoldVersion": "A String", # The Skaffold version to use when operating on this release, such as "1.20.0". Not all versions are valid; Google Cloud Deploy supports a specific set of versions. If unset, the most recent supported Skaffold version will be used. + "targetArtifacts": { # Output only. Map from target ID to the target artifacts created during the render operation. + "a_key": { # The artifacts produced by a target render operation. + "artifactUri": "A String", # Output only. URI of a directory containing the artifacts. This contains deployment configuration used by Skaffold during a rollout, and all paths are relative to this location. + "manifestPath": "A String", # Output only. File path of the rendered manifest relative to the URI. + "skaffoldConfigPath": "A String", # Output only. File path of the resolved Skaffold configuration relative to the URI. + }, + }, + "targetRenders": { # Output only. Map from target ID to details of the render operation for that target. + "a_key": { # Details of rendering for a single target. + "renderingBuild": "A String", # Output only. The resource name of the Cloud Build `Build` object that is used to render the manifest for this target. Format is `projects/{project}/locations/{location}/builds/{build}`. + "renderingState": "A String", # Output only. Current state of the render operation for this Target. + }, + }, + "targetSnapshots": [ # Output only. Snapshot of the parent pipeline's targets taken at release creation time. + { # A `Target` resource in the Google Cloud Deploy API. A `Target` defines a location to which a Skaffold configuration can be deployed. + "annotations": { # Optional. User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "createTime": "A String", # Output only. Time at which the `Target` was created. + "description": "A String", # Optional. Description of the `Target`. Max length is 255 characters. + "etag": "A String", # Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "executionConfigs": [ # Configurations for all execution that relates to this `Target`. Each `ExecutionEnvironmentUsage` value may only be used in a single configuration; using the same value multiple times is an error. When one or more configurations are specified, they must include the `RENDER` and `DEPLOY` `ExecutionEnvironmentUsage` values. When no configurations are specified, execution will use the default specified in `DefaultPool`. + { # Configuration of the environment to use when calling Skaffold. + "defaultPool": { # Execution using the default Cloud Build pool. # Optional. Use default Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + }, + "privatePool": { # Execution using a private Cloud Build pool. # Optional. Use private Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + "workerPool": "A String", # Required. Resource name of the Cloud Build worker pool to use. The format is `projects/{project}/locations/{location}/workerPools/{pool}`. + }, + "usages": [ # Required. Usages when this configuration should be applied. + "A String", + ], + }, + ], + "gke": { # Information specifying a GKE Cluster. # Information specifying a GKE Cluster. + "cluster": "A String", # Information specifying a GKE Cluster. Format is `projects/{project_id}/locations/{location_id}/clusters/{cluster_id}. + }, + "labels": { # Optional. Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Target`. Format is projects/{project}/locations/{location}/ deliveryPipelines/{deliveryPipeline}/targets/a-z{0,62}. + "requireApproval": True or False, # Optional. Whether or not the `Target` requires approval. + "targetId": "A String", # Output only. Resource id of the `Target`. + "uid": "A String", # Output only. Unique identifier of the `Target`. + "updateTime": "A String", # Output only. Most recent time at which the `Target` was updated. + }, + ], + "uid": "A String", # Output only. Unique identifier of the `Release`. + }, + ], + "unreachable": [ # Locations that could not be reached. + "A String", + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.releases.rollouts.html b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.releases.rollouts.html new file mode 100644 index 00000000000..5a25981dd7f --- /dev/null +++ b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.releases.rollouts.html @@ -0,0 +1,292 @@ + + + +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Google Cloud Deploy API . projects . locations . deliveryPipelines . releases . rollouts
+Instance Methods
++
+approve(name, body=None, x__xgafv=None)
Approves a Rollout.
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, requestId=None, rolloutId=None, validateOnly=None, x__xgafv=None)
Creates a new Rollout in a given project and location.
+ +Gets details of a single Rollout.
++
+list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
Lists Rollouts in a given project and location.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+Method Details
+++ +approve(name, body=None, x__xgafv=None)
+Approves a Rollout. + +Args: + name: string, Required. Name of the Rollout. Format is projects/{project}/locations/{location}/deliveryPipelines/{deliveryPipeline}/ releases/{release}/rollouts/{rollout}. (required) + body: object, The request body. + The object takes the form of: + +{ # The request object used by `ApproveRollout`. + "approved": True or False, # Required. True = approve; false = reject +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response object from `ApproveRollout`. +}+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, requestId=None, rolloutId=None, validateOnly=None, x__xgafv=None)
+Creates a new Rollout in a given project and location. + +Args: + parent: string, Required. The parent collection in which the `Rollout` should be created. Format should be projects/{project_id}/locations/{location_name}/deliveryPipelines/{pipeline_name}/releases/{release_name}. (required) + body: object, The request body. + The object takes the form of: + +{ # A `Rollout` resource in the Google Cloud Deploy API. A `Rollout` contains information around a specific deployment to a `Target`. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "approvalState": "A String", # Output only. Approval state of the `Rollout`. + "approveTime": "A String", # Output only. Time at which the `Rollout` was approved. + "createTime": "A String", # Output only. Time at which the `Rollout` was created. + "deployEndTime": "A String", # Output only. Time at which the `Rollout` finished deploying. + "deployStartTime": "A String", # Output only. Time at which the `Rollout` started deploying. + "deployingBuild": "A String", # Output only. The resource name of the Cloud Build `Build` object that is used to deploy the Rollout. Format is `projects/{project}/locations/{location}/builds/{build}`. + "description": "A String", # Description of the `Rollout` for user purposes. Max length is 255 characters. + "enqueueTime": "A String", # Output only. Time at which the `Rollout` was enqueued. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "failureReason": "A String", # Output only. Reason the build failed. Empty if the build succeeded. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Rollout`. Format is projects/{project}/ locations/{location}/deliveryPipelines/{deliveryPipeline}/ releases/{release}/rollouts/a-z{0,62}. + "state": "A String", # Output only. Current state of the `Rollout`. + "targetId": "A String", # Required. The ID of Target to which this `Rollout` is deploying. + "uid": "A String", # Output only. Unique identifier of the `Rollout`. +} + + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + rolloutId: string, Required. ID of the `Rollout`. + validateOnly: boolean, Optional. If set to true, the request is validated and the user is provided with an expected result, but no actual change is made. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)
+Gets details of a single Rollout. + +Args: + name: string, Required. Name of the `Rollout`. Format must be projects/{project_id}/locations/{location_name}/deliveryPipelines/{pipeline_name}/releases/{release_name}/rollouts/{rollout_name}. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A `Rollout` resource in the Google Cloud Deploy API. A `Rollout` contains information around a specific deployment to a `Target`. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "approvalState": "A String", # Output only. Approval state of the `Rollout`. + "approveTime": "A String", # Output only. Time at which the `Rollout` was approved. + "createTime": "A String", # Output only. Time at which the `Rollout` was created. + "deployEndTime": "A String", # Output only. Time at which the `Rollout` finished deploying. + "deployStartTime": "A String", # Output only. Time at which the `Rollout` started deploying. + "deployingBuild": "A String", # Output only. The resource name of the Cloud Build `Build` object that is used to deploy the Rollout. Format is `projects/{project}/locations/{location}/builds/{build}`. + "description": "A String", # Description of the `Rollout` for user purposes. Max length is 255 characters. + "enqueueTime": "A String", # Output only. Time at which the `Rollout` was enqueued. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "failureReason": "A String", # Output only. Reason the build failed. Empty if the build succeeded. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Rollout`. Format is projects/{project}/ locations/{location}/deliveryPipelines/{deliveryPipeline}/ releases/{release}/rollouts/a-z{0,62}. + "state": "A String", # Output only. Current state of the `Rollout`. + "targetId": "A String", # Required. The ID of Target to which this `Rollout` is deploying. + "uid": "A String", # Output only. Unique identifier of the `Rollout`. +}+++ +list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists Rollouts in a given project and location. + +Args: + parent: string, Required. The `Release` which owns this collection of `Rollout` objects. (required) + filter: string, Optional. Filter builds to be returned. See https://google.aip.dev/160 for more details. + orderBy: string, Optional. Field to sort by. See https://google.aip.dev/132#ordering for more details. + pageSize: integer, Optional. The maximum number of `Rollout` objects to return. The service may return fewer than this value. If unspecified, at most 50 `Rollout` objects will be returned. The maximum value is 1000; values above 1000 will be set to 1000. + pageToken: string, Optional. A page token, received from a previous `ListRollouts` call. Provide this to retrieve the subsequent page. When paginating, all other provided parameters match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # ListRolloutsResponse is the response object reutrned by `ListRollouts`. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "rollouts": [ # The `Rollout` objects. + { # A `Rollout` resource in the Google Cloud Deploy API. A `Rollout` contains information around a specific deployment to a `Target`. + "annotations": { # User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "approvalState": "A String", # Output only. Approval state of the `Rollout`. + "approveTime": "A String", # Output only. Time at which the `Rollout` was approved. + "createTime": "A String", # Output only. Time at which the `Rollout` was created. + "deployEndTime": "A String", # Output only. Time at which the `Rollout` finished deploying. + "deployStartTime": "A String", # Output only. Time at which the `Rollout` started deploying. + "deployingBuild": "A String", # Output only. The resource name of the Cloud Build `Build` object that is used to deploy the Rollout. Format is `projects/{project}/locations/{location}/builds/{build}`. + "description": "A String", # Description of the `Rollout` for user purposes. Max length is 255 characters. + "enqueueTime": "A String", # Output only. Time at which the `Rollout` was enqueued. + "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "failureReason": "A String", # Output only. Reason the build failed. Empty if the build succeeded. + "labels": { # Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Rollout`. Format is projects/{project}/ locations/{location}/deliveryPipelines/{deliveryPipeline}/ releases/{release}/rollouts/a-z{0,62}. + "state": "A String", # Output only. Current state of the `Rollout`. + "targetId": "A String", # Required. The ID of Target to which this `Rollout` is deploying. + "uid": "A String", # Output only. Unique identifier of the `Rollout`. + }, + ], + "unreachable": [ # Locations that could not be reached. + "A String", + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/clouddeploy_v1.projects.locations.html b/docs/dyn/clouddeploy_v1.projects.locations.html new file mode 100644 index 00000000000..1d3e812609e --- /dev/null +++ b/docs/dyn/clouddeploy_v1.projects.locations.html @@ -0,0 +1,219 @@ + + + +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Google Cloud Deploy API . projects . locations
+Instance Methods
+ +Returns the deliveryPipelines Resource.
+ ++
+operations()
+Returns the operations Resource.
+ ++
+targets()
+Returns the targets Resource.
+ ++
+close()
Close httplib2 connections.
+ +Gets information about a location.
++
+getConfig(name, x__xgafv=None)
Gets the configuration for a location.
++
+list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
Lists information about the supported locations for this service.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +get(name, x__xgafv=None)
+Gets information about a location. + +Args: + name: string, Resource name for the location. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A resource that represents Google Cloud Platform location. + "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". + "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} + "a_key": "A String", + }, + "locationId": "A String", # The canonical id for this location. For example: `"us-east1"`. + "metadata": { # Service-specific metadata. For example the available capacity at the given location. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # Resource name for the location, which may vary between implementations. For example: `"projects/example-project/locations/us-east1"` +}+++ +getConfig(name, x__xgafv=None)
+Gets the configuration for a location. + +Args: + name: string, Required. Name of requested configuration. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Service-wide configuration. + "defaultSkaffoldVersion": "A String", # Output only. Default Skaffold version that is assigned when a Release is created without specifying a Skaffold version. + "name": "A String", # Name of the configuration. + "supportedVersions": [ # Output only. All supported versions of Skaffold. + { # Details of a supported Skaffold version. + "supportEndDate": { # Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values * A month and day value, with a zero year, such as an anniversary * A year on its own, with zero month and day values * A year and month value, with a zero day, such as a credit card expiration date Related types are google.type.TimeOfDay and `google.protobuf.Timestamp`. # Date when this version is expected to no longer be supported. + "day": 42, # Day of a month. Must be from 1 to 31 and valid for the year and month, or 0 to specify a year by itself or a year and month where the day isn't significant. + "month": 42, # Month of a year. Must be from 1 to 12, or 0 to specify a year without a month and day. + "year": 42, # Year of the date. Must be from 1 to 9999, or 0 to specify a date without a year. + }, + "version": "A String", # Release version number. For example, "1.20.3". + }, + ], +}+++ +list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists information about the supported locations for this service. + +Args: + name: string, The resource that owns the locations collection, if applicable. (required) + filter: string, A filter to narrow down results to a preferred subset. The filtering language accepts strings like "displayName=tokyo", and is documented in more detail in [AIP-160](https://google.aip.dev/160). + pageSize: integer, The maximum number of results to return. If not set, the service selects a default. + pageToken: string, A page token received from the `next_page_token` field in the response. Send that page token to receive the subsequent page. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response message for Locations.ListLocations. + "locations": [ # A list of locations that matches the specified filter in the request. + { # A resource that represents Google Cloud Platform location. + "displayName": "A String", # The friendly name for this location, typically a nearby city name. For example, "Tokyo". + "labels": { # Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"} + "a_key": "A String", + }, + "locationId": "A String", # The canonical id for this location. For example: `"us-east1"`. + "metadata": { # Service-specific metadata. For example the available capacity at the given location. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # Resource name for the location, which may vary between implementations. For example: `"projects/example-project/locations/us-east1"` + }, + ], + "nextPageToken": "A String", # The standard List next-page token. +}+++ + \ No newline at end of file diff --git a/docs/dyn/clouddeploy_v1.projects.locations.operations.html b/docs/dyn/clouddeploy_v1.projects.locations.operations.html new file mode 100644 index 00000000000..f672e017cd6 --- /dev/null +++ b/docs/dyn/clouddeploy_v1.projects.locations.operations.html @@ -0,0 +1,235 @@ + + + +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Google Cloud Deploy API . projects . locations . operations
+Instance Methods
++
+cancel(name, body=None, x__xgafv=None)
Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
++
+close()
Close httplib2 connections.
+ +Deletes a long-running operation. This method indicates that the client is no longer interested in the operation result. It does not cancel the operation. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`.
+ +Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.
++
+list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+Method Details
+++ +cancel(name, body=None, x__xgafv=None)
+Starts asynchronous cancellation on a long-running operation. The server makes a best effort to cancel the operation, but success is not guaranteed. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. Clients can use Operations.GetOperation or other methods to check whether the cancellation succeeded or whether the operation completed despite cancellation. On successful cancellation, the operation is not deleted; instead, it becomes an operation with an Operation.error value with a google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`. + +Args: + name: string, The name of the operation resource to be cancelled. (required) + body: object, The request body. + The object takes the form of: + +{ # The request message for Operations.CancelOperation. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`. +}+++ +close()
+Close httplib2 connections.+++ +delete(name, x__xgafv=None)
+Deletes a long-running operation. This method indicates that the client is no longer interested in the operation result. It does not cancel the operation. If the server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`. + +Args: + name: string, The name of the operation resource to be deleted. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`. +}+++ +get(name, x__xgafv=None)
+Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service. + +Args: + name: string, The name of the operation resource. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +list(name, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists operations that match the specified filter in the request. If the server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `"/v1/{name=users/*}/operations"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id. + +Args: + name: string, The name of the operation's parent resource. (required) + filter: string, The standard list filter. + pageSize: integer, The standard list page size. + pageToken: string, The standard list page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response message for Operations.ListOperations. + "nextPageToken": "A String", # The standard List next-page token. + "operations": [ # A list of operations that matches the specified filter in the request. + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + }, + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/clouddeploy_v1.projects.locations.targets.html b/docs/dyn/clouddeploy_v1.projects.locations.targets.html new file mode 100644 index 00000000000..92cba912647 --- /dev/null +++ b/docs/dyn/clouddeploy_v1.projects.locations.targets.html @@ -0,0 +1,598 @@ + + + +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Google Cloud Deploy API . projects . locations . targets
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, requestId=None, targetId=None, validateOnly=None, x__xgafv=None)
Creates a new Target in a given project and location.
++
+delete(name, allowMissing=None, etag=None, requestId=None, validateOnly=None, x__xgafv=None)
Deletes a single Target.
+ +Gets details of a single Target.
++
+getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
++
+list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
Lists Targets in a given project and location.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+ +Updates the parameters of a single Target.
++
+setIamPolicy(resource, body=None, x__xgafv=None)
Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
++
+testIamPermissions(resource, body=None, x__xgafv=None)
Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, requestId=None, targetId=None, validateOnly=None, x__xgafv=None)
+Creates a new Target in a given project and location. + +Args: + parent: string, Required. The parent collection in which the `Target` should be created. Format should be projects/{project_id}/locations/{location_name}. (required) + body: object, The request body. + The object takes the form of: + +{ # A `Target` resource in the Google Cloud Deploy API. A `Target` defines a location to which a Skaffold configuration can be deployed. + "annotations": { # Optional. User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "createTime": "A String", # Output only. Time at which the `Target` was created. + "description": "A String", # Optional. Description of the `Target`. Max length is 255 characters. + "etag": "A String", # Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "executionConfigs": [ # Configurations for all execution that relates to this `Target`. Each `ExecutionEnvironmentUsage` value may only be used in a single configuration; using the same value multiple times is an error. When one or more configurations are specified, they must include the `RENDER` and `DEPLOY` `ExecutionEnvironmentUsage` values. When no configurations are specified, execution will use the default specified in `DefaultPool`. + { # Configuration of the environment to use when calling Skaffold. + "defaultPool": { # Execution using the default Cloud Build pool. # Optional. Use default Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + }, + "privatePool": { # Execution using a private Cloud Build pool. # Optional. Use private Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + "workerPool": "A String", # Required. Resource name of the Cloud Build worker pool to use. The format is `projects/{project}/locations/{location}/workerPools/{pool}`. + }, + "usages": [ # Required. Usages when this configuration should be applied. + "A String", + ], + }, + ], + "gke": { # Information specifying a GKE Cluster. # Information specifying a GKE Cluster. + "cluster": "A String", # Information specifying a GKE Cluster. Format is `projects/{project_id}/locations/{location_id}/clusters/{cluster_id}. + }, + "labels": { # Optional. Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Target`. Format is projects/{project}/locations/{location}/ deliveryPipelines/{deliveryPipeline}/targets/a-z{0,62}. + "requireApproval": True or False, # Optional. Whether or not the `Target` requires approval. + "targetId": "A String", # Output only. Resource id of the `Target`. + "uid": "A String", # Output only. Unique identifier of the `Target`. + "updateTime": "A String", # Output only. Most recent time at which the `Target` was updated. +} + + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + targetId: string, Required. ID of the `Target`. + validateOnly: boolean, Optional. If set to true, the request is validated and the user is provided with an expected result, but no actual change is made. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, allowMissing=None, etag=None, requestId=None, validateOnly=None, x__xgafv=None)
+Deletes a single Target. + +Args: + name: string, Required. The name of the `Target` to delete. Format should be projects/{project_id}/locations/{location_name}/targets/{target_name}. (required) + allowMissing: boolean, Optional. If set to true, then deleting an already deleted or non-existing DeliveryPipeline will succeed. + etag: string, Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes after the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + validateOnly: boolean, Optional. If set, validate the request and preview the review, but do not actually post it. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)
+Gets details of a single Target. + +Args: + name: string, Required. Name of the `Target`. Format must be projects/{project_id}/locations/{location_name}/targets/{target_name}. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A `Target` resource in the Google Cloud Deploy API. A `Target` defines a location to which a Skaffold configuration can be deployed. + "annotations": { # Optional. User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "createTime": "A String", # Output only. Time at which the `Target` was created. + "description": "A String", # Optional. Description of the `Target`. Max length is 255 characters. + "etag": "A String", # Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "executionConfigs": [ # Configurations for all execution that relates to this `Target`. Each `ExecutionEnvironmentUsage` value may only be used in a single configuration; using the same value multiple times is an error. When one or more configurations are specified, they must include the `RENDER` and `DEPLOY` `ExecutionEnvironmentUsage` values. When no configurations are specified, execution will use the default specified in `DefaultPool`. + { # Configuration of the environment to use when calling Skaffold. + "defaultPool": { # Execution using the default Cloud Build pool. # Optional. Use default Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + }, + "privatePool": { # Execution using a private Cloud Build pool. # Optional. Use private Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + "workerPool": "A String", # Required. Resource name of the Cloud Build worker pool to use. The format is `projects/{project}/locations/{location}/workerPools/{pool}`. + }, + "usages": [ # Required. Usages when this configuration should be applied. + "A String", + ], + }, + ], + "gke": { # Information specifying a GKE Cluster. # Information specifying a GKE Cluster. + "cluster": "A String", # Information specifying a GKE Cluster. Format is `projects/{project_id}/locations/{location_id}/clusters/{cluster_id}. + }, + "labels": { # Optional. Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Target`. Format is projects/{project}/locations/{location}/ deliveryPipelines/{deliveryPipeline}/targets/a-z{0,62}. + "requireApproval": True or False, # Optional. Whether or not the `Target` requires approval. + "targetId": "A String", # Output only. Resource id of the `Target`. + "uid": "A String", # Output only. Unique identifier of the `Target`. + "updateTime": "A String", # Output only. Most recent time at which the `Target` was updated. +}+++ +getIamPolicy(resource, options_requestedPolicyVersion=None, x__xgafv=None)
+Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set. + +Args: + resource: string, REQUIRED: The resource for which the policy is being requested. See the operation documentation for the appropriate value for this field. (required) + options_requestedPolicyVersion: integer, Optional. The maximum policy version that will be used to format the policy. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional role bindings must specify version 3. Policies with no conditional role bindings may specify any valid value or leave the field unset. The policy in the response might use the policy version that you specified, or it might use a lower policy version. For example, if you specify version 3, but the policy has no conditional role bindings, the response uses version 1. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members`, or principals, to a single `role`. Principals can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') etag: BwWWja0YfJA= version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). + "auditConfigs": [ # Specifies cloud audit logging configuration for this policy. + { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging. + "auditLogConfigs": [ # The configuration for logging of each type of permission. + { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging. + "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members. + "A String", + ], + "logType": "A String", # The log type that this config enables. + }, + ], + "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services. + }, + ], + "bindings": [ # Associates a list of `members`, or principals, with a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one principal. The `bindings` in a `Policy` can refer to up to 1,500 principals; up to 250 of these principals can be Google groups. Each occurrence of a principal counts towards these limits. For example, if the `bindings` grant 50 different roles to `user:alice@example.com`, and not to any other principal, then you can add another 1,450 principals to the `bindings` in the `Policy`. + { # Associates `members`, or principals, with a `role`. + "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI. + "expression": "A String", # Textual representation of an expression in Common Expression Language syntax. + "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file. + "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression. + }, + "members": [ # Specifies the principals requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. + "A String", + ], + "role": "A String", # Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`. + }, + ], + "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. + "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). +}+++ +list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
+Lists Targets in a given project and location. + +Args: + parent: string, Required. The parent, which owns this collection of targets. Format must be projects/{project_id}/locations/{location_name}. (required) + filter: string, Optional. Filter builds to be returned. See https://google.aip.dev/160 for more details. + orderBy: string, Optional. Field to sort by. See https://google.aip.dev/132#ordering for more details. + pageSize: integer, Optional. The maximum number of `Target` objects to return. The service may return fewer than this value. If unspecified, at most 50 `Target` objects will be returned. The maximum value is 1000; values above 1000 will be set to 1000. + pageToken: string, Optional. A page token, received from a previous `ListTargets` call. Provide this to retrieve the subsequent page. When paginating, all other provided parameters match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response object from `ListTargets`. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "targets": [ # The `Target` objects. + { # A `Target` resource in the Google Cloud Deploy API. A `Target` defines a location to which a Skaffold configuration can be deployed. + "annotations": { # Optional. User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "createTime": "A String", # Output only. Time at which the `Target` was created. + "description": "A String", # Optional. Description of the `Target`. Max length is 255 characters. + "etag": "A String", # Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "executionConfigs": [ # Configurations for all execution that relates to this `Target`. Each `ExecutionEnvironmentUsage` value may only be used in a single configuration; using the same value multiple times is an error. When one or more configurations are specified, they must include the `RENDER` and `DEPLOY` `ExecutionEnvironmentUsage` values. When no configurations are specified, execution will use the default specified in `DefaultPool`. + { # Configuration of the environment to use when calling Skaffold. + "defaultPool": { # Execution using the default Cloud Build pool. # Optional. Use default Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + }, + "privatePool": { # Execution using a private Cloud Build pool. # Optional. Use private Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + "workerPool": "A String", # Required. Resource name of the Cloud Build worker pool to use. The format is `projects/{project}/locations/{location}/workerPools/{pool}`. + }, + "usages": [ # Required. Usages when this configuration should be applied. + "A String", + ], + }, + ], + "gke": { # Information specifying a GKE Cluster. # Information specifying a GKE Cluster. + "cluster": "A String", # Information specifying a GKE Cluster. Format is `projects/{project_id}/locations/{location_id}/clusters/{cluster_id}. + }, + "labels": { # Optional. Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Target`. Format is projects/{project}/locations/{location}/ deliveryPipelines/{deliveryPipeline}/targets/a-z{0,62}. + "requireApproval": True or False, # Optional. Whether or not the `Target` requires approval. + "targetId": "A String", # Output only. Resource id of the `Target`. + "uid": "A String", # Output only. Unique identifier of the `Target`. + "updateTime": "A String", # Output only. Most recent time at which the `Target` was updated. + }, + ], + "unreachable": [ # Locations that could not be reached. + "A String", + ], +}+++ +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ +patch(name, allowMissing=None, body=None, requestId=None, updateMask=None, validateOnly=None, x__xgafv=None)
+Updates the parameters of a single Target. + +Args: + name: string, Optional. Name of the `Target`. Format is projects/{project}/locations/{location}/ deliveryPipelines/{deliveryPipeline}/targets/a-z{0,62}. (required) + body: object, The request body. + The object takes the form of: + +{ # A `Target` resource in the Google Cloud Deploy API. A `Target` defines a location to which a Skaffold configuration can be deployed. + "annotations": { # Optional. User annotations. These attributes can only be set and used by the user, and not by Google Cloud Deploy. See https://google.aip.dev/128#annotations for more details such as format and size limitations. + "a_key": "A String", + }, + "createTime": "A String", # Output only. Time at which the `Target` was created. + "description": "A String", # Optional. Description of the `Target`. Max length is 255 characters. + "etag": "A String", # Optional. This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. + "executionConfigs": [ # Configurations for all execution that relates to this `Target`. Each `ExecutionEnvironmentUsage` value may only be used in a single configuration; using the same value multiple times is an error. When one or more configurations are specified, they must include the `RENDER` and `DEPLOY` `ExecutionEnvironmentUsage` values. When no configurations are specified, execution will use the default specified in `DefaultPool`. + { # Configuration of the environment to use when calling Skaffold. + "defaultPool": { # Execution using the default Cloud Build pool. # Optional. Use default Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + }, + "privatePool": { # Execution using a private Cloud Build pool. # Optional. Use private Cloud Build pool. + "artifactStorage": "A String", # Optional. Cloud Storage location where execution outputs should be stored. This can either be a bucket ("gs://my-bucket") or a path within a bucket ("gs://my-bucket/my-dir"). If unspecified, a default bucket located in the same region will be used. + "serviceAccount": "A String", # Optional. Google service account to use for execution. If unspecified, the project execution service account (-compute@developer.gserviceaccount.com) will be used. + "workerPool": "A String", # Required. Resource name of the Cloud Build worker pool to use. The format is `projects/{project}/locations/{location}/workerPools/{pool}`. + }, + "usages": [ # Required. Usages when this configuration should be applied. + "A String", + ], + }, + ], + "gke": { # Information specifying a GKE Cluster. # Information specifying a GKE Cluster. + "cluster": "A String", # Information specifying a GKE Cluster. Format is `projects/{project_id}/locations/{location_id}/clusters/{cluster_id}. + }, + "labels": { # Optional. Labels are attributes that can be set and used by both the user and by Google Cloud Deploy. Labels must meet the following constraints: * Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. * All characters must use UTF-8 encoding, and international characters are allowed. * Keys must start with a lowercase letter or international character. * Each resource is limited to a maximum of 64 labels. Both keys and values are additionally constrained to be <= 128 bytes. + "a_key": "A String", + }, + "name": "A String", # Optional. Name of the `Target`. Format is projects/{project}/locations/{location}/ deliveryPipelines/{deliveryPipeline}/targets/a-z{0,62}. + "requireApproval": True or False, # Optional. Whether or not the `Target` requires approval. + "targetId": "A String", # Output only. Resource id of the `Target`. + "uid": "A String", # Output only. Unique identifier of the `Target`. + "updateTime": "A String", # Output only. Most recent time at which the `Target` was updated. +} + + allowMissing: boolean, Optional. If set to true, updating a `Target` that does not exist will result in the creation of a new `Target`. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + updateMask: string, Required. Field mask is used to specify the fields to be overwritten in the Target resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then all fields will be overwritten. + validateOnly: boolean, Optional. If set to true, the request is validated and the user is provided with an expected result, but no actual change is made. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +setIamPolicy(resource, body=None, x__xgafv=None)
+Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors. + +Args: + resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required) + body: object, The request body. + The object takes the form of: + +{ # Request message for `SetIamPolicy` method. + "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members`, or principals, to a single `role`. Principals can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') etag: BwWWja0YfJA= version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the `resource`. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Cloud Platform services (such as Projects) might reject them. + "auditConfigs": [ # Specifies cloud audit logging configuration for this policy. + { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging. + "auditLogConfigs": [ # The configuration for logging of each type of permission. + { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging. + "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members. + "A String", + ], + "logType": "A String", # The log type that this config enables. + }, + ], + "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services. + }, + ], + "bindings": [ # Associates a list of `members`, or principals, with a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one principal. The `bindings` in a `Policy` can refer to up to 1,500 principals; up to 250 of these principals can be Google groups. Each occurrence of a principal counts towards these limits. For example, if the `bindings` grant 50 different roles to `user:alice@example.com`, and not to any other principal, then you can add another 1,450 principals to the `bindings` in the `Policy`. + { # Associates `members`, or principals, with a `role`. + "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI. + "expression": "A String", # Textual representation of an expression in Common Expression Language syntax. + "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file. + "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression. + }, + "members": [ # Specifies the principals requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. + "A String", + ], + "role": "A String", # Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`. + }, + ], + "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. + "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + }, + "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used: `paths: "bindings, etag"` +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members`, or principals, to a single `role`. Principals can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } **YAML example:** bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') etag: BwWWja0YfJA= version: 3 For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). + "auditConfigs": [ # Specifies cloud audit logging configuration for this policy. + { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts jose@example.com from DATA_READ logging, and aliya@example.com from DATA_WRITE logging. + "auditLogConfigs": [ # The configuration for logging of each type of permission. + { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging. + "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members. + "A String", + ], + "logType": "A String", # The log type that this config enables. + }, + ], + "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services. + }, + ], + "bindings": [ # Associates a list of `members`, or principals, with a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one principal. The `bindings` in a `Policy` can refer to up to 1,500 principals; up to 250 of these principals can be Google groups. Each occurrence of a principal counts towards these limits. For example, if the `bindings` grant 50 different roles to `user:alice@example.com`, and not to any other principal, then you can add another 1,450 principals to the `bindings` in the `Policy`. + { # Associates `members`, or principals, with a `role`. + "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI. + "expression": "A String", # Textual representation of an expression in Common Expression Language syntax. + "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file. + "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression. + }, + "members": [ # Specifies the principals requesting access for a Cloud Platform resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. + "A String", + ], + "role": "A String", # Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`. + }, + ], + "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. + "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). +}+++ + \ No newline at end of file diff --git a/docs/dyn/cloudidentity_v1beta1.customers.userinvitations.html b/docs/dyn/cloudidentity_v1beta1.customers.userinvitations.html index b2494e4e5ec..346c9392be8 100644 --- a/docs/dyn/cloudidentity_v1beta1.customers.userinvitations.html +++ b/docs/dyn/cloudidentity_v1beta1.customers.userinvitations.html @@ -156,7 +156,7 @@testIamPermissions(resource, body=None, x__xgafv=None)
+Returns permissions that a caller has on the specified resource. If the resource does not exist, this will return an empty set of permissions, not a `NOT_FOUND` error. Note: This operation is designed to be used for building permission-aware UIs and command-line tools, not for authorization checking. This operation may "fail open" without warning. + +Args: + resource: string, REQUIRED: The resource for which the policy detail is being requested. See the operation documentation for the appropriate value for this field. (required) + body: object, The request body. + The object takes the form of: + +{ # Request message for `TestIamPermissions` method. + "permissions": [ # The set of permissions to check for the `resource`. Permissions with wildcards (such as '*' or 'storage.*') are not allowed. For more information see [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions). + "A String", + ], +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response message for `TestIamPermissions` method. + "permissions": [ # A subset of `TestPermissionsRequest.permissions` that the caller is allowed. + "A String", + ], +}+Method Details
Returns: An object of the form: - { # The `UserInvitation` resource represents an email that can be sent to an unmanaged user account inviting them to join the customer’s Google Workspace or Cloud Identity account. An unmanaged account shares an email address domain with the Google Workspace or Cloud Identity account but is not managed by it yet. If the user accepts the `UserInvitation`, the user account will become managed. + { # The `UserInvitation` resource represents an email that can be sent to an unmanaged user account inviting them to join the customer's Google Workspace or Cloud Identity account. An unmanaged account shares an email address domain with the Google Workspace or Cloud Identity account but is not managed by it yet. If the user accepts the `UserInvitation`, the user account will become managed. "mailsSentCount": "A String", # Number of invitation emails sent to the user. "name": "A String", # Shall be of the form `customers/{customer}/userinvitations/{user_email_address}`. "state": "A String", # State of the `UserInvitation`. @@ -204,7 +204,7 @@Method Details
{ # Response message for UserInvitation listing request. "nextPageToken": "A String", # The token for the next page. If not empty, indicates that there may be more `UserInvitation` resources that match the listing request; this value can be used in a subsequent ListUserInvitationsRequest to get continued results with the current list call. "userInvitations": [ # The list of UserInvitation resources. - { # The `UserInvitation` resource represents an email that can be sent to an unmanaged user account inviting them to join the customer’s Google Workspace or Cloud Identity account. An unmanaged account shares an email address domain with the Google Workspace or Cloud Identity account but is not managed by it yet. If the user accepts the `UserInvitation`, the user account will become managed. + { # The `UserInvitation` resource represents an email that can be sent to an unmanaged user account inviting them to join the customer's Google Workspace or Cloud Identity account. An unmanaged account shares an email address domain with the Google Workspace or Cloud Identity account but is not managed by it yet. If the user accepts the `UserInvitation`, the user account will become managed. "mailsSentCount": "A String", # Number of invitation emails sent to the user. "name": "A String", # Shall be of the form `customers/{customer}/userinvitations/{user_email_address}`. "state": "A String", # State of the `UserInvitation`. diff --git a/docs/dyn/cloudiot_v1.projects.locations.registries.groups.html b/docs/dyn/cloudiot_v1.projects.locations.registries.groups.html index 0249e01c14f..7fd304eca58 100644 --- a/docs/dyn/cloudiot_v1.projects.locations.registries.groups.html +++ b/docs/dyn/cloudiot_v1.projects.locations.registries.groups.html @@ -108,7 +108,7 @@Method Details
{ # Request message for `GetIamPolicy` method. "options": { # Encapsulates settings provided to GetIamPolicy. # OPTIONAL: A `GetPolicyOptions` object for specifying options to `GetIamPolicy`. - "requestedPolicyVersion": 42, # Optional. The policy format version to be returned. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + "requestedPolicyVersion": 42, # Optional. The maximum policy version that will be used to format the policy. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional role bindings must specify version 3. Policies with no conditional role bindings may specify any valid value or leave the field unset. The policy in the response might use the policy version that you specified, or it might use a lower policy version. For example, if you specify version 3, but the policy has no conditional role bindings, the response uses version 1. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). }, } diff --git a/docs/dyn/cloudiot_v1.projects.locations.registries.html b/docs/dyn/cloudiot_v1.projects.locations.registries.html index e7049e72e7b..d84fcb89aaa 100644 --- a/docs/dyn/cloudiot_v1.projects.locations.registries.html +++ b/docs/dyn/cloudiot_v1.projects.locations.registries.html @@ -325,7 +325,7 @@Method Details
{ # Request message for `GetIamPolicy` method. "options": { # Encapsulates settings provided to GetIamPolicy. # OPTIONAL: A `GetPolicyOptions` object for specifying options to `GetIamPolicy`. - "requestedPolicyVersion": 42, # Optional. The policy format version to be returned. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional bindings must specify version 3. Policies without any conditional bindings may specify any valid value or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). + "requestedPolicyVersion": 42, # Optional. The maximum policy version that will be used to format the policy. Valid values are 0, 1, and 3. Requests specifying an invalid value will be rejected. Requests for policies with any conditional role bindings must specify version 3. Policies with no conditional role bindings may specify any valid value or leave the field unset. The policy in the response might use the policy version that you specified, or it might use a lower policy version. For example, if you specify version 3, but the policy has no conditional role bindings, the response uses version 1. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). }, } diff --git a/docs/dyn/composer_v1.projects.locations.environments.html b/docs/dyn/composer_v1.projects.locations.environments.html index c89f06e5b61..138f56b9b77 100644 --- a/docs/dyn/composer_v1.projects.locations.environments.html +++ b/docs/dyn/composer_v1.projects.locations.environments.html @@ -174,7 +174,7 @@Method Details
"webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.* # Optional. The configuration settings for the Airflow web server App Engine instance. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -355,7 +355,7 @@Method Details
"webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.* # Optional. The configuration settings for the Airflow web server App Engine instance. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -477,7 +477,7 @@Method Details
"webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.* # Optional. The configuration settings for the Airflow web server App Engine instance. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -607,7 +607,7 @@Method Details
"webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.* # Optional. The configuration settings for the Airflow web server App Engine instance. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. diff --git a/docs/dyn/composer_v1beta1.projects.locations.environments.html b/docs/dyn/composer_v1beta1.projects.locations.environments.html index d4453e71fa2..1c2fbacf399 100644 --- a/docs/dyn/composer_v1beta1.projects.locations.environments.html +++ b/docs/dyn/composer_v1beta1.projects.locations.environments.html @@ -209,6 +209,7 @@Method Details
}, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudSqlIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range in tenant project will be reserved for Cloud SQL. Needs to be disjoint from web_server_ipv4_cidr_block @@ -239,7 +240,7 @@Method Details
"webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for the Airflow web server App Engine instance. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -407,6 +408,7 @@Method Details
}, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudSqlIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range in tenant project will be reserved for Cloud SQL. Needs to be disjoint from web_server_ipv4_cidr_block @@ -437,7 +439,7 @@Method Details
"webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for the Airflow web server App Engine instance. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -546,6 +548,7 @@Method Details
}, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudSqlIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range in tenant project will be reserved for Cloud SQL. Needs to be disjoint from web_server_ipv4_cidr_block @@ -576,7 +579,7 @@Method Details
"webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for the Airflow web server App Engine instance. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -693,6 +696,7 @@Method Details
}, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudSqlIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range in tenant project will be reserved for Cloud SQL. Needs to be disjoint from web_server_ipv4_cidr_block @@ -723,7 +727,7 @@Method Details
"webServerConfig": { # The configuration settings for the Airflow web server App Engine instance. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The configuration settings for the Airflow web server App Engine instance. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. "machineType": "A String", # Optional. Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. If not specified, composer-n1-webserver-2 will be used. Value custom is returned only in response, if Airflow web server parameters were manually changed to a non-standard values. }, - "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. + "webServerNetworkAccessControl": { # Network-level access control policy for the Airflow web server. # Optional. The network-level access control policy for the Airflow web server. If unspecified, no network-level access restrictions will be applied. "allowedIpRanges": [ # A collection of allowed IP ranges with descriptions. { # Allowed IP range with user-provided description. "description": "A String", # Optional. User-provided description. It must contain at most 300 characters. @@ -762,7 +766,7 @@Method Details
"uuid": "A String", # Output only. The UUID (Universally Unique IDentifier) associated with this environment. This value is generated when the environment is created. } - updateMask: string, Required. A comma-separated list of paths, relative to `Environment`, of fields to update. For example, to set the version of scikit-learn to install in the environment to 0.19.0 and to remove an existing installation of argparse, the `updateMask` parameter would include the following two `paths` values: "config.softwareConfig.pypiPackages.scikit-learn" and "config.softwareConfig.pypiPackages.argparse". The included patch environment would specify the scikit-learn version as follows: { "config":{ "softwareConfig":{ "pypiPackages":{ "scikit-learn":"==0.19.0" } } } } Note that in the above example, any existing PyPI packages other than scikit-learn and argparse will be unaffected. Only one update type may be included in a single request's `updateMask`. For example, one cannot update both the PyPI packages and labels in the same request. However, it is possible to update multiple members of a map field simultaneously in the same request. For example, to set the labels "label1" and "label2" while clearing "label3" (assuming it already exists), one can provide the paths "labels.label1", "labels.label2", and "labels.label3" and populate the patch environment as follows: { "labels":{ "label1":"new-label1-value" "label2":"new-label2-value" } } Note that in the above example, any existing labels that are not included in the `updateMask` will be unaffected. It is also possible to replace an entire map field by providing the map field's path in the `updateMask`. The new value of the field will be that which is provided in the patch environment. For example, to delete all pre-existing user-specified PyPI packages and install botocore at version 1.7.14, the `updateMask` would contain the path "config.softwareConfig.pypiPackages", and the patch environment would be the following: { "config":{ "softwareConfig":{ "pypiPackages":{ "botocore":"==1.7.14" } } } } **Note:** Only the following fields can be updated: * `config.softwareConfig.pypiPackages` * Replace all custom custom PyPI packages. If a replacement package map is not included in `environment`, all custom PyPI packages are cleared. It is an error to provide both this mask and a mask specifying an individual package. * `config.softwareConfig.pypiPackages.`packagename * Update the custom PyPI package *packagename*, preserving other packages. To delete the package, include it in `updateMask`, and omit the mapping for it in `environment.config.softwareConfig.pypiPackages`. It is an error to provide both a mask of this form and the `config.softwareConfig.pypiPackages` mask. * `labels` * Replace all environment labels. If a replacement labels map is not included in `environment`, all labels are cleared. It is an error to provide both this mask and a mask specifying one or more individual labels. * `labels.`labelName * Set the label named *labelName*, while preserving other labels. To delete the label, include it in `updateMask` and omit its mapping in `environment.labels`. It is an error to provide both a mask of this form and the `labels` mask. * `config.nodeCount` * Horizontally scale the number of nodes in the environment. An integer greater than or equal to 3 must be provided in the `config.nodeCount` field. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.webServerNetworkAccessControl` * Replace the environment's current WebServerNetworkAccessControl. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.softwareConfig.airflowConfigOverrides` * Replace all Apache Airflow config overrides. If a replacement config overrides map is not included in `environment`, all config overrides are cleared. It is an error to provide both this mask and a mask specifying one or more individual config overrides. * `config.softwareConfig.airflowConfigOverrides.`section-name * Override the Apache Airflow config property *name* in the section named *section*, preserving other properties. To delete the property override, include it in `updateMask` and omit its mapping in `environment.config.softwareConfig.airflowConfigOverrides`. It is an error to provide both a mask of this form and the `config.softwareConfig.airflowConfigOverrides` mask. * `config.softwareConfig.envVariables` * Replace all environment variables. If a replacement environment variable map is not included in `environment`, all custom environment variables are cleared. It is an error to provide both this mask and a mask specifying one or more individual environment variables. * `config.softwareConfig.imageVersion` * Upgrade the version of the environment in-place. Refer to `SoftwareConfig.image_version` for information on how to format the new image version. Additionally, the new image version cannot effect a version downgrade and must match the current image version's Composer major version and Airflow major and minor versions. Consult the [Cloud Composer Version List](https://cloud.google.com/composer/docs/concepts/versioning/composer-versions) for valid values. * `config.softwareConfig.schedulerCount` * Horizontally scale the number of schedulers in Airflow. A positive integer not greater than the number of nodes must be provided in the `config.softwareConfig.schedulerCount` field. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-2.*.*. * `config.databaseConfig.machineType` * Cloud SQL machine type used by Airflow database. It has to be one of: db-n1-standard-2, db-n1-standard-4, db-n1-standard-8 or db-n1-standard-16. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.webServerConfig.machineType` * Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.maintenanceWindow` * Maintenance window during which Cloud Composer components may be under maintenance. * `config.workloadsConfig` * The workloads configuration settings for the GKE cluster associated with the Cloud Composer environment. Supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. * `config.environmentSize` * The size of the Cloud Composer environment. Supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. + updateMask: string, Required. A comma-separated list of paths, relative to `Environment`, of fields to update. For example, to set the version of scikit-learn to install in the environment to 0.19.0 and to remove an existing installation of argparse, the `updateMask` parameter would include the following two `paths` values: "config.softwareConfig.pypiPackages.scikit-learn" and "config.softwareConfig.pypiPackages.argparse". The included patch environment would specify the scikit-learn version as follows: { "config":{ "softwareConfig":{ "pypiPackages":{ "scikit-learn":"==0.19.0" } } } } Note that in the above example, any existing PyPI packages other than scikit-learn and argparse will be unaffected. Only one update type may be included in a single request's `updateMask`. For example, one cannot update both the PyPI packages and labels in the same request. However, it is possible to update multiple members of a map field simultaneously in the same request. For example, to set the labels "label1" and "label2" while clearing "label3" (assuming it already exists), one can provide the paths "labels.label1", "labels.label2", and "labels.label3" and populate the patch environment as follows: { "labels":{ "label1":"new-label1-value" "label2":"new-label2-value" } } Note that in the above example, any existing labels that are not included in the `updateMask` will be unaffected. It is also possible to replace an entire map field by providing the map field's path in the `updateMask`. The new value of the field will be that which is provided in the patch environment. For example, to delete all pre-existing user-specified PyPI packages and install botocore at version 1.7.14, the `updateMask` would contain the path "config.softwareConfig.pypiPackages", and the patch environment would be the following: { "config":{ "softwareConfig":{ "pypiPackages":{ "botocore":"==1.7.14" } } } } **Note:** Only the following fields can be updated: * `config.softwareConfig.pypiPackages` * Replace all custom custom PyPI packages. If a replacement package map is not included in `environment`, all custom PyPI packages are cleared. It is an error to provide both this mask and a mask specifying an individual package. * `config.softwareConfig.pypiPackages.`packagename * Update the custom PyPI package *packagename*, preserving other packages. To delete the package, include it in `updateMask`, and omit the mapping for it in `environment.config.softwareConfig.pypiPackages`. It is an error to provide both a mask of this form and the `config.softwareConfig.pypiPackages` mask. * `labels` * Replace all environment labels. If a replacement labels map is not included in `environment`, all labels are cleared. It is an error to provide both this mask and a mask specifying one or more individual labels. * `labels.`labelName * Set the label named *labelName*, while preserving other labels. To delete the label, include it in `updateMask` and omit its mapping in `environment.labels`. It is an error to provide both a mask of this form and the `labels` mask. * `config.nodeCount` * Horizontally scale the number of nodes in the environment. An integer greater than or equal to 3 must be provided in the `config.nodeCount` field. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.webServerNetworkAccessControl` * Replace the environment's current WebServerNetworkAccessControl. * `config.softwareConfig.airflowConfigOverrides` * Replace all Apache Airflow config overrides. If a replacement config overrides map is not included in `environment`, all config overrides are cleared. It is an error to provide both this mask and a mask specifying one or more individual config overrides. * `config.softwareConfig.airflowConfigOverrides.`section-name * Override the Apache Airflow config property *name* in the section named *section*, preserving other properties. To delete the property override, include it in `updateMask` and omit its mapping in `environment.config.softwareConfig.airflowConfigOverrides`. It is an error to provide both a mask of this form and the `config.softwareConfig.airflowConfigOverrides` mask. * `config.softwareConfig.envVariables` * Replace all environment variables. If a replacement environment variable map is not included in `environment`, all custom environment variables are cleared. It is an error to provide both this mask and a mask specifying one or more individual environment variables. * `config.softwareConfig.imageVersion` * Upgrade the version of the environment in-place. Refer to `SoftwareConfig.image_version` for information on how to format the new image version. Additionally, the new image version cannot effect a version downgrade and must match the current image version's Composer major version and Airflow major and minor versions. Consult the [Cloud Composer Version List](https://cloud.google.com/composer/docs/concepts/versioning/composer-versions) for valid values. * `config.softwareConfig.schedulerCount` * Horizontally scale the number of schedulers in Airflow. A positive integer not greater than the number of nodes must be provided in the `config.softwareConfig.schedulerCount` field. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-2.*.*. * `config.databaseConfig.machineType` * Cloud SQL machine type used by Airflow database. It has to be one of: db-n1-standard-2, db-n1-standard-4, db-n1-standard-8 or db-n1-standard-16. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.webServerConfig.machineType` * Machine type on which Airflow web server is running. It has to be one of: composer-n1-webserver-2, composer-n1-webserver-4 or composer-n1-webserver-8. Supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. * `config.maintenanceWindow` * Maintenance window during which Cloud Composer components may be under maintenance. * `config.workloadsConfig` * The workloads configuration settings for the GKE cluster associated with the Cloud Composer environment. Supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. * `config.environmentSize` * The size of the Cloud Composer environment. Supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format diff --git a/docs/dyn/container_v1beta1.projects.locations.clusters.html b/docs/dyn/container_v1beta1.projects.locations.clusters.html index aee34ae466d..a7abc5d4def 100644 --- a/docs/dyn/container_v1beta1.projects.locations.clusters.html +++ b/docs/dyn/container_v1beta1.projects.locations.clusters.html @@ -447,6 +447,9 @@Method Details
"A String", ], }, + "managedPrometheusConfig": { # ManagedPrometheusConfig defines the configuration for Google Cloud Managed Service for Prometheus. # Enable Google Cloud Managed Service for Prometheus in the cluster. + "enabled": True or False, # Enable Managed Collection. + }, }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. @@ -1121,6 +1124,9 @@Method Details
"A String", ], }, + "managedPrometheusConfig": { # ManagedPrometheusConfig defines the configuration for Google Cloud Managed Service for Prometheus. # Enable Google Cloud Managed Service for Prometheus in the cluster. + "enabled": True or False, # Enable Managed Collection. + }, }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. @@ -1698,6 +1704,9 @@Method Details
"A String", ], }, + "managedPrometheusConfig": { # ManagedPrometheusConfig defines the configuration for Google Cloud Managed Service for Prometheus. # Enable Google Cloud Managed Service for Prometheus in the cluster. + "enabled": True or False, # Enable Managed Collection. + }, }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. @@ -3028,6 +3037,9 @@Method Details
"A String", ], }, + "managedPrometheusConfig": { # ManagedPrometheusConfig defines the configuration for Google Cloud Managed Service for Prometheus. # Enable Google Cloud Managed Service for Prometheus in the cluster. + "enabled": True or False, # Enable Managed Collection. + }, }, "desiredMonitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "desiredNodePoolAutoscaling": { # NodePoolAutoscaling contains information required by cluster autoscaler to adjust the size of the node pool to the current cluster usage. # Autoscaler configuration for the node pool specified in desired_node_pool_id. If there is only one pool in the cluster and desired_node_pool_id is not provided then the change applies to that single node pool. diff --git a/docs/dyn/container_v1beta1.projects.zones.clusters.html b/docs/dyn/container_v1beta1.projects.zones.clusters.html index 10d894f2085..c52c73a65de 100644 --- a/docs/dyn/container_v1beta1.projects.zones.clusters.html +++ b/docs/dyn/container_v1beta1.projects.zones.clusters.html @@ -558,6 +558,9 @@Method Details
"A String", ], }, + "managedPrometheusConfig": { # ManagedPrometheusConfig defines the configuration for Google Cloud Managed Service for Prometheus. # Enable Google Cloud Managed Service for Prometheus in the cluster. + "enabled": True or False, # Enable Managed Collection. + }, }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. @@ -1232,6 +1235,9 @@Method Details
"A String", ], }, + "managedPrometheusConfig": { # ManagedPrometheusConfig defines the configuration for Google Cloud Managed Service for Prometheus. # Enable Google Cloud Managed Service for Prometheus in the cluster. + "enabled": True or False, # Enable Managed Collection. + }, }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. @@ -1853,6 +1859,9 @@Method Details
"A String", ], }, + "managedPrometheusConfig": { # ManagedPrometheusConfig defines the configuration for Google Cloud Managed Service for Prometheus. # Enable Google Cloud Managed Service for Prometheus in the cluster. + "enabled": True or False, # Enable Managed Collection. + }, }, "monitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "name": "A String", # The name of this cluster. The name must be unique within this project and location (e.g. zone or region), and can be up to 40 characters with the following restrictions: * Lowercase letters, numbers, and hyphens only. * Must start with a letter. * Must end with a number or a letter. @@ -3089,6 +3098,9 @@Method Details
"A String", ], }, + "managedPrometheusConfig": { # ManagedPrometheusConfig defines the configuration for Google Cloud Managed Service for Prometheus. # Enable Google Cloud Managed Service for Prometheus in the cluster. + "enabled": True or False, # Enable Managed Collection. + }, }, "desiredMonitoringService": "A String", # The monitoring service the cluster should use to write metrics. Currently available options: * "monitoring.googleapis.com/kubernetes" - The Cloud Monitoring service with a Kubernetes-native resource model * `monitoring.googleapis.com` - The legacy Cloud Monitoring service (no longer available as of GKE 1.15). * `none` - No metrics will be exported from the cluster. If left as an empty string,`monitoring.googleapis.com/kubernetes` will be used for GKE 1.14+ or `monitoring.googleapis.com` for earlier versions. "desiredNodePoolAutoscaling": { # NodePoolAutoscaling contains information required by cluster autoscaler to adjust the size of the node pool to the current cluster usage. # Autoscaler configuration for the node pool specified in desired_node_pool_id. If there is only one pool in the cluster and desired_node_pool_id is not provided then the change applies to that single node pool. diff --git a/docs/dyn/containeranalysis_v1alpha1.projects.notes.occurrences.html b/docs/dyn/containeranalysis_v1alpha1.projects.notes.occurrences.html index c40b88200a7..a3e637f5733 100644 --- a/docs/dyn/containeranalysis_v1alpha1.projects.notes.occurrences.html +++ b/docs/dyn/containeranalysis_v1alpha1.projects.notes.occurrences.html @@ -118,7 +118,7 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. - "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "intotoProvenance": { # Deprecated. See InTotoStatement for the replacement. In-toto Provenance representation as defined in spec. "builderConfig": { # required "id": "A String", }, @@ -152,6 +152,87 @@Method Details
"type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. }, }, + "intotoStatement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". # In-toto Statement representation as defined in spec. The intoto_statement can contain any type of provenance. The serialized payload of the statement can be stored and signed in the Occurrence's envelope. + "_type": "A String", # Always "https://in-toto.io/Statement/v0.1". + "predicateType": "A String", # "https://slsa.dev/provenance/v0.1" for SlsaProvenance. + "provenance": { # provenance is a predicate of type intotoprovenance + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": [ # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "slsaProvenance": { # SlsaProvenance is the slsa provenance as defined by the slsa spec. # slsa_provenance is a predicate of type slsaProvenance + "builder": { # SlsaBuilder encapsulates the identity of the builder of this provenance. # builder is the builder of this provenance + "id": "A String", # id is the id of the slsa provenance builder + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + { # Material is a material used in the generation of the provenance + "digest": { # digest is a map from a hash algorithm (e.g. sha256) to the value in the material + "a_key": "A String", + }, + "uri": "A String", # uri is the uri of the material + }, + ], + "metadata": { # Other properties of the build. # metadata is the metadata of the provenance + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). + "arguments": { # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ # subject is the subjects of the intoto statement + { # Subject refers to the subject of the intoto statement + "digest": { # "": "" Algorithms can be e.g. sha256, sha512 See https://github.com/in-toto/attestation/blob/main/spec/field_types.md#DigestSet + "a_key": "A String", + }, + "name": "A String", # name is the name of the Subject used here + }, + ], + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", diff --git a/docs/dyn/containeranalysis_v1alpha1.projects.occurrences.html b/docs/dyn/containeranalysis_v1alpha1.projects.occurrences.html index 72932c1d248..0c15db9ebfa 100644 --- a/docs/dyn/containeranalysis_v1alpha1.projects.occurrences.html +++ b/docs/dyn/containeranalysis_v1alpha1.projects.occurrences.html @@ -134,7 +134,7 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. - "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "intotoProvenance": { # Deprecated. See InTotoStatement for the replacement. In-toto Provenance representation as defined in spec. "builderConfig": { # required "id": "A String", }, @@ -168,6 +168,87 @@Method Details
"type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. }, }, + "intotoStatement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". # In-toto Statement representation as defined in spec. The intoto_statement can contain any type of provenance. The serialized payload of the statement can be stored and signed in the Occurrence's envelope. + "_type": "A String", # Always "https://in-toto.io/Statement/v0.1". + "predicateType": "A String", # "https://slsa.dev/provenance/v0.1" for SlsaProvenance. + "provenance": { # provenance is a predicate of type intotoprovenance + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": [ # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "slsaProvenance": { # SlsaProvenance is the slsa provenance as defined by the slsa spec. # slsa_provenance is a predicate of type slsaProvenance + "builder": { # SlsaBuilder encapsulates the identity of the builder of this provenance. # builder is the builder of this provenance + "id": "A String", # id is the id of the slsa provenance builder + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + { # Material is a material used in the generation of the provenance + "digest": { # digest is a map from a hash algorithm (e.g. sha256) to the value in the material + "a_key": "A String", + }, + "uri": "A String", # uri is the uri of the material + }, + ], + "metadata": { # Other properties of the build. # metadata is the metadata of the provenance + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). + "arguments": { # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ # subject is the subjects of the intoto statement + { # Subject refers to the subject of the intoto statement + "digest": { # "": "" Algorithms can be e.g. sha256, sha512 See https://github.com/in-toto/attestation/blob/main/spec/field_types.md#DigestSet + "a_key": "A String", + }, + "name": "A String", # name is the name of the Subject used here + }, + ], + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -638,7 +719,7 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. - "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "intotoProvenance": { # Deprecated. See InTotoStatement for the replacement. In-toto Provenance representation as defined in spec. "builderConfig": { # required "id": "A String", }, @@ -672,6 +753,87 @@Method Details
"type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. }, }, + "intotoStatement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". # In-toto Statement representation as defined in spec. The intoto_statement can contain any type of provenance. The serialized payload of the statement can be stored and signed in the Occurrence's envelope. + "_type": "A String", # Always "https://in-toto.io/Statement/v0.1". + "predicateType": "A String", # "https://slsa.dev/provenance/v0.1" for SlsaProvenance. + "provenance": { # provenance is a predicate of type intotoprovenance + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": [ # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "slsaProvenance": { # SlsaProvenance is the slsa provenance as defined by the slsa spec. # slsa_provenance is a predicate of type slsaProvenance + "builder": { # SlsaBuilder encapsulates the identity of the builder of this provenance. # builder is the builder of this provenance + "id": "A String", # id is the id of the slsa provenance builder + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + { # Material is a material used in the generation of the provenance + "digest": { # digest is a map from a hash algorithm (e.g. sha256) to the value in the material + "a_key": "A String", + }, + "uri": "A String", # uri is the uri of the material + }, + ], + "metadata": { # Other properties of the build. # metadata is the metadata of the provenance + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). + "arguments": { # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ # subject is the subjects of the intoto statement + { # Subject refers to the subject of the intoto statement + "digest": { # "": "" Algorithms can be e.g. sha256, sha512 See https://github.com/in-toto/attestation/blob/main/spec/field_types.md#DigestSet + "a_key": "A String", + }, + "name": "A String", # name is the name of the Subject used here + }, + ], + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -1166,7 +1328,7 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. - "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "intotoProvenance": { # Deprecated. See InTotoStatement for the replacement. In-toto Provenance representation as defined in spec. "builderConfig": { # required "id": "A String", }, @@ -1200,6 +1362,87 @@Method Details
"type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. }, }, + "intotoStatement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". # In-toto Statement representation as defined in spec. The intoto_statement can contain any type of provenance. The serialized payload of the statement can be stored and signed in the Occurrence's envelope. + "_type": "A String", # Always "https://in-toto.io/Statement/v0.1". + "predicateType": "A String", # "https://slsa.dev/provenance/v0.1" for SlsaProvenance. + "provenance": { # provenance is a predicate of type intotoprovenance + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": [ # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "slsaProvenance": { # SlsaProvenance is the slsa provenance as defined by the slsa spec. # slsa_provenance is a predicate of type slsaProvenance + "builder": { # SlsaBuilder encapsulates the identity of the builder of this provenance. # builder is the builder of this provenance + "id": "A String", # id is the id of the slsa provenance builder + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + { # Material is a material used in the generation of the provenance + "digest": { # digest is a map from a hash algorithm (e.g. sha256) to the value in the material + "a_key": "A String", + }, + "uri": "A String", # uri is the uri of the material + }, + ], + "metadata": { # Other properties of the build. # metadata is the metadata of the provenance + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). + "arguments": { # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ # subject is the subjects of the intoto statement + { # Subject refers to the subject of the intoto statement + "digest": { # "": "" Algorithms can be e.g. sha256, sha512 See https://github.com/in-toto/attestation/blob/main/spec/field_types.md#DigestSet + "a_key": "A String", + }, + "name": "A String", # name is the name of the Subject used here + }, + ], + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -1978,7 +2221,7 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. - "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "intotoProvenance": { # Deprecated. See InTotoStatement for the replacement. In-toto Provenance representation as defined in spec. "builderConfig": { # required "id": "A String", }, @@ -2012,6 +2255,87 @@Method Details
"type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. }, }, + "intotoStatement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". # In-toto Statement representation as defined in spec. The intoto_statement can contain any type of provenance. The serialized payload of the statement can be stored and signed in the Occurrence's envelope. + "_type": "A String", # Always "https://in-toto.io/Statement/v0.1". + "predicateType": "A String", # "https://slsa.dev/provenance/v0.1" for SlsaProvenance. + "provenance": { # provenance is a predicate of type intotoprovenance + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": [ # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "slsaProvenance": { # SlsaProvenance is the slsa provenance as defined by the slsa spec. # slsa_provenance is a predicate of type slsaProvenance + "builder": { # SlsaBuilder encapsulates the identity of the builder of this provenance. # builder is the builder of this provenance + "id": "A String", # id is the id of the slsa provenance builder + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + { # Material is a material used in the generation of the provenance + "digest": { # digest is a map from a hash algorithm (e.g. sha256) to the value in the material + "a_key": "A String", + }, + "uri": "A String", # uri is the uri of the material + }, + ], + "metadata": { # Other properties of the build. # metadata is the metadata of the provenance + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). + "arguments": { # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ # subject is the subjects of the intoto statement + { # Subject refers to the subject of the intoto statement + "digest": { # "": "" Algorithms can be e.g. sha256, sha512 See https://github.com/in-toto/attestation/blob/main/spec/field_types.md#DigestSet + "a_key": "A String", + }, + "name": "A String", # name is the name of the Subject used here + }, + ], + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -2499,7 +2823,7 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. - "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "intotoProvenance": { # Deprecated. See InTotoStatement for the replacement. In-toto Provenance representation as defined in spec. "builderConfig": { # required "id": "A String", }, @@ -2533,6 +2857,87 @@Method Details
"type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. }, }, + "intotoStatement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". # In-toto Statement representation as defined in spec. The intoto_statement can contain any type of provenance. The serialized payload of the statement can be stored and signed in the Occurrence's envelope. + "_type": "A String", # Always "https://in-toto.io/Statement/v0.1". + "predicateType": "A String", # "https://slsa.dev/provenance/v0.1" for SlsaProvenance. + "provenance": { # provenance is a predicate of type intotoprovenance + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": [ # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "slsaProvenance": { # SlsaProvenance is the slsa provenance as defined by the slsa spec. # slsa_provenance is a predicate of type slsaProvenance + "builder": { # SlsaBuilder encapsulates the identity of the builder of this provenance. # builder is the builder of this provenance + "id": "A String", # id is the id of the slsa provenance builder + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + { # Material is a material used in the generation of the provenance + "digest": { # digest is a map from a hash algorithm (e.g. sha256) to the value in the material + "a_key": "A String", + }, + "uri": "A String", # uri is the uri of the material + }, + ], + "metadata": { # Other properties of the build. # metadata is the metadata of the provenance + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). + "arguments": { # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ # subject is the subjects of the intoto statement + { # Subject refers to the subject of the intoto statement + "digest": { # "": "" Algorithms can be e.g. sha256, sha512 See https://github.com/in-toto/attestation/blob/main/spec/field_types.md#DigestSet + "a_key": "A String", + }, + "name": "A String", # name is the name of the Subject used here + }, + ], + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", @@ -3003,7 +3408,7 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. - "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "intotoProvenance": { # Deprecated. See InTotoStatement for the replacement. In-toto Provenance representation as defined in spec. "builderConfig": { # required "id": "A String", }, @@ -3037,6 +3442,87 @@Method Details
"type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. }, }, + "intotoStatement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". # In-toto Statement representation as defined in spec. The intoto_statement can contain any type of provenance. The serialized payload of the statement can be stored and signed in the Occurrence's envelope. + "_type": "A String", # Always "https://in-toto.io/Statement/v0.1". + "predicateType": "A String", # "https://slsa.dev/provenance/v0.1" for SlsaProvenance. + "provenance": { # provenance is a predicate of type intotoprovenance + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": [ # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "slsaProvenance": { # SlsaProvenance is the slsa provenance as defined by the slsa spec. # slsa_provenance is a predicate of type slsaProvenance + "builder": { # SlsaBuilder encapsulates the identity of the builder of this provenance. # builder is the builder of this provenance + "id": "A String", # id is the id of the slsa provenance builder + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + { # Material is a material used in the generation of the provenance + "digest": { # digest is a map from a hash algorithm (e.g. sha256) to the value in the material + "a_key": "A String", + }, + "uri": "A String", # uri is the uri of the material + }, + ], + "metadata": { # Other properties of the build. # metadata is the metadata of the provenance + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). + "arguments": { # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ # subject is the subjects of the intoto statement + { # Subject refers to the subject of the intoto statement + "digest": { # "": "" Algorithms can be e.g. sha256, sha512 See https://github.com/in-toto/attestation/blob/main/spec/field_types.md#DigestSet + "a_key": "A String", + }, + "name": "A String", # name is the name of the Subject used here + }, + ], + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", diff --git a/docs/dyn/containeranalysis_v1alpha1.providers.notes.occurrences.html b/docs/dyn/containeranalysis_v1alpha1.providers.notes.occurrences.html index 0c6450438c0..2bb37bf33b1 100644 --- a/docs/dyn/containeranalysis_v1alpha1.providers.notes.occurrences.html +++ b/docs/dyn/containeranalysis_v1alpha1.providers.notes.occurrences.html @@ -118,7 +118,7 @@Method Details
}, }, "buildDetails": { # Message encapsulating build provenance details. # Build details for a verifiable build. - "intotoProvenance": { # In-toto Provenance representation as defined in spec. + "intotoProvenance": { # Deprecated. See InTotoStatement for the replacement. In-toto Provenance representation as defined in spec. "builderConfig": { # required "id": "A String", }, @@ -152,6 +152,87 @@Method Details
"type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. }, }, + "intotoStatement": { # Spec defined at https://github.com/in-toto/attestation/tree/main/spec#statement The serialized InTotoStatement will be stored as Envelope.payload. Envelope.payloadType is always "application/vnd.in-toto+json". # In-toto Statement representation as defined in spec. The intoto_statement can contain any type of provenance. The serialized payload of the statement can be stored and signed in the Occurrence's envelope. + "_type": "A String", # Always "https://in-toto.io/Statement/v0.1". + "predicateType": "A String", # "https://slsa.dev/provenance/v0.1" for SlsaProvenance. + "provenance": { # provenance is a predicate of type intotoprovenance + "builderConfig": { # required + "id": "A String", + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + "A String", + ], + "metadata": { # Other properties of the build. + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). required + "arguments": [ # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": [ # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "slsaProvenance": { # SlsaProvenance is the slsa provenance as defined by the slsa spec. # slsa_provenance is a predicate of type slsaProvenance + "builder": { # SlsaBuilder encapsulates the identity of the builder of this provenance. # builder is the builder of this provenance + "id": "A String", # id is the id of the slsa provenance builder + }, + "materials": [ # The collection of artifacts that influenced the build including sources, dependencies, build tools, base images, and so on. This is considered to be incomplete unless metadata.completeness.materials is true. Unset or null is equivalent to empty. + { # Material is a material used in the generation of the provenance + "digest": { # digest is a map from a hash algorithm (e.g. sha256) to the value in the material + "a_key": "A String", + }, + "uri": "A String", # uri is the uri of the material + }, + ], + "metadata": { # Other properties of the build. # metadata is the metadata of the provenance + "buildFinishedOn": "A String", # The timestamp of when the build completed. + "buildInvocationId": "A String", # Identifies the particular build invocation, which can be useful for finding associated logs or other ad-hoc analysis. The value SHOULD be globally unique, per in-toto Provenance spec. + "buildStartedOn": "A String", # The timestamp of when the build started. + "completeness": { # Indicates that the builder claims certain fields in this message to be complete. # Indicates that the builder claims certain fields in this message to be complete. + "arguments": True or False, # If true, the builder claims that recipe.arguments is complete, meaning that all external inputs are properly captured in the recipe. + "environment": True or False, # If true, the builder claims that recipe.environment is claimed to be complete. + "materials": True or False, # If true, the builder claims that materials are complete, usually through some controls to prevent network access. Sometimes called "hermetic". + }, + "reproducible": True or False, # If true, the builder claims that running the recipe on materials will produce bit-for-bit identical output. + }, + "recipe": { # Steps taken to build the artifact. For a TaskRun, typically each container corresponds to one step in the recipe. # Identifies the configuration used for the build. When combined with materials, this SHOULD fully describe the build, such that re-running this recipe results in bit-for-bit identical output (if the build is reproducible). + "arguments": { # Collection of all external inputs that influenced the build on top of recipe.definedInMaterial and recipe.entryPoint. For example, if the recipe type were "make", then this might be the flags passed to make aside from the target, which is captured in recipe.entryPoint. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "definedInMaterial": "A String", # Index in materials containing the recipe steps that are not implied by recipe.type. For example, if the recipe type were "make", then this would point to the source containing the Makefile, not the make program itself. Set to -1 if the recipe doesn't come from a material, as zero is default unset value for int64. + "entryPoint": "A String", # String identifying the entry point into the build. This is often a path to a configuration file and/or a target label within that file. The syntax and meaning are defined by recipe.type. For example, if the recipe type were "make", then this would reference the directory in which to run make as well as which target to use. + "environment": { # Any other builder-controlled inputs necessary for correctly evaluating the recipe. Usually only needed for reproducing the build but not evaluated as part of policy. Depending on the recipe Type, the structure may be different. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "type": "A String", # URI indicating what type of recipe was performed. It determines the meaning of recipe.entryPoint, recipe.arguments, recipe.environment, and materials. + }, + }, + "subject": [ # subject is the subjects of the intoto statement + { # Subject refers to the subject of the intoto statement + "digest": { # "": "" Algorithms can be e.g. sha256, sha512 See https://github.com/in-toto/attestation/blob/main/spec/field_types.md#DigestSet + "a_key": "A String", + }, + "name": "A String", # name is the name of the Subject used here + }, + ], + }, "provenance": { # Provenance of a build. Contains all information needed to verify the full details about the build from source to completion. # The actual provenance "buildOptions": { # Special options applied to this build. This is a catch-all field where build providers can enter any desired additional details. "a_key": "A String", diff --git a/docs/dyn/datastore_v1.projects.html b/docs/dyn/datastore_v1.projects.html index 4f71a3fd441..96acf9cb890 100644 --- a/docs/dyn/datastore_v1.projects.html +++ b/docs/dyn/datastore_v1.projects.html @@ -248,7 +248,40 @@Method Details
], }, "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value + "a_key": { # A message that can hold any of the supported value types and associated metadata. + "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`. + "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'. + # Object with schema name: Value + ], + }, + "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. + "booleanValue": True or False, # A boolean value. + "doubleValue": 3.14, # A double value. + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. + "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. + "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. + "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. + "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0]. + }, + "integerValue": "A String", # An integer value. + "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value. + "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. + "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. + "projectId": "A String", # The ID of the project to which the entities belong. + }, + "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. + { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. + "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. + "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + }, + ], + }, + "meaning": 42, # The `meaning` field should only be populated for backwards compatibility. + "nullValue": "A String", # A null value. + "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes. + "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down. + }, }, }, "update": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # The entity to update. The entity must already exist. Must have a complete key path. @@ -266,7 +299,40 @@Method Details
], }, "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value + "a_key": { # A message that can hold any of the supported value types and associated metadata. + "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`. + "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'. + # Object with schema name: Value + ], + }, + "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. + "booleanValue": True or False, # A boolean value. + "doubleValue": 3.14, # A double value. + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. + "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. + "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. + "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. + "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0]. + }, + "integerValue": "A String", # An integer value. + "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value. + "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. + "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. + "projectId": "A String", # The ID of the project to which the entities belong. + }, + "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. + { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. + "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. + "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + }, + ], + }, + "meaning": 42, # The `meaning` field should only be populated for backwards compatibility. + "nullValue": "A String", # A null value. + "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes. + "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down. + }, }, }, "upsert": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # The entity to upsert. The entity may or may not already exist. The entity key's final path element may be incomplete. @@ -284,7 +350,40 @@Method Details
], }, "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value + "a_key": { # A message that can hold any of the supported value types and associated metadata. + "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`. + "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'. + # Object with schema name: Value + ], + }, + "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. + "booleanValue": True or False, # A boolean value. + "doubleValue": 3.14, # A double value. + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. + "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. + "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. + "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. + "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0]. + }, + "integerValue": "A String", # An integer value. + "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value. + "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. + "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. + "projectId": "A String", # The ID of the project to which the entities belong. + }, + "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. + { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. + "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. + "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + }, + ], + }, + "meaning": 42, # The `meaning` field should only be populated for backwards compatibility. + "nullValue": "A String", # A null value. + "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes. + "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down. + }, }, }, }, @@ -503,7 +602,40 @@Method Details
], }, "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value + "a_key": { # A message that can hold any of the supported value types and associated metadata. + "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`. + "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'. + # Object with schema name: Value + ], + }, + "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. + "booleanValue": True or False, # A boolean value. + "doubleValue": 3.14, # A double value. + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. + "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. + "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. + "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. + "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0]. + }, + "integerValue": "A String", # An integer value. + "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value. + "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. + "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. + "projectId": "A String", # The ID of the project to which the entities belong. + }, + "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. + { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. + "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. + "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + }, + ], + }, + "meaning": 42, # The `meaning` field should only be populated for backwards compatibility. + "nullValue": "A String", # A null value. + "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes. + "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down. + }, }, }, "version": "A String", # The version of the entity, a strictly positive number that monotonically increases with changes to the entity. This field is set for `FULL` entity results. For missing entities in `LookupResponse`, this is the version of the snapshot that was used to look up the entity, and it is always set except for eventually consistent reads. @@ -527,7 +659,40 @@Method Details
], }, "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value + "a_key": { # A message that can hold any of the supported value types and associated metadata. + "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`. + "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'. + # Object with schema name: Value + ], + }, + "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. + "booleanValue": True or False, # A boolean value. + "doubleValue": 3.14, # A double value. + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. + "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. + "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. + "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. + "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0]. + }, + "integerValue": "A String", # An integer value. + "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value. + "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. + "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. + "projectId": "A String", # The ID of the project to which the entities belong. + }, + "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. + { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. + "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. + "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + }, + ], + }, + "meaning": 42, # The `meaning` field should only be populated for backwards compatibility. + "nullValue": "A String", # A null value. + "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes. + "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down. + }, }, }, "version": "A String", # The version of the entity, a strictly positive number that monotonically increases with changes to the entity. This field is set for `FULL` entity results. For missing entities in `LookupResponse`, this is the version of the snapshot that was used to look up the entity, and it is always set except for eventually consistent reads. @@ -625,24 +790,7 @@Method Details
"blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. "booleanValue": True or False, # A boolean value. "doubleValue": 3.14, # A double value. - "entityValue": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. - "key": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # The entity's key. An entity must have a key, unless otherwise documented (for example, an entity in `Value.entity_value` may have no key). An entity's kind is its key path's last element's kind, or null if it has no key. - "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. - "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. - "projectId": "A String", # The ID of the project to which the entities belong. - }, - "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. - { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. - "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. - "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. - "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. - }, - ], - }, - "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value - }, - }, + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. @@ -681,24 +829,7 @@Method Details
"blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. "booleanValue": True or False, # A boolean value. "doubleValue": 3.14, # A double value. - "entityValue": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. - "key": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # The entity's key. An entity must have a key, unless otherwise documented (for example, an entity in `Value.entity_value` may have no key). An entity's kind is its key path's last element's kind, or null if it has no key. - "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. - "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. - "projectId": "A String", # The ID of the project to which the entities belong. - }, - "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. - { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. - "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. - "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. - "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. - }, - ], - }, - "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value - }, - }, + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. @@ -759,24 +890,7 @@Method Details
"blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. "booleanValue": True or False, # A boolean value. "doubleValue": 3.14, # A double value. - "entityValue": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. - "key": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # The entity's key. An entity must have a key, unless otherwise documented (for example, an entity in `Value.entity_value` may have no key). An entity's kind is its key path's last element's kind, or null if it has no key. - "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. - "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. - "projectId": "A String", # The ID of the project to which the entities belong. - }, - "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. - { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. - "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. - "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. - "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. - }, - ], - }, - "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value - }, - }, + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. @@ -863,7 +977,40 @@Method Details
], }, "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value + "a_key": { # A message that can hold any of the supported value types and associated metadata. + "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`. + "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'. + # Object with schema name: Value + ], + }, + "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. + "booleanValue": True or False, # A boolean value. + "doubleValue": 3.14, # A double value. + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. + "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. + "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. + "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. + "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0]. + }, + "integerValue": "A String", # An integer value. + "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value. + "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. + "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. + "projectId": "A String", # The ID of the project to which the entities belong. + }, + "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. + { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. + "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. + "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. + }, + ], + }, + "meaning": 42, # The `meaning` field should only be populated for backwards compatibility. + "nullValue": "A String", # A null value. + "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes. + "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down. + }, }, }, "version": "A String", # The version of the entity, a strictly positive number that monotonically increases with changes to the entity. This field is set for `FULL` entity results. For missing entities in `LookupResponse`, this is the version of the snapshot that was used to look up the entity, and it is always set except for eventually consistent reads. @@ -902,24 +1049,7 @@Method Details
"blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded. "booleanValue": True or False, # A boolean value. "doubleValue": 3.14, # A double value. - "entityValue": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. - "key": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # The entity's key. An entity must have a key, unless otherwise documented (for example, an entity in `Value.entity_value` may have no key). An entity's kind is its key path's last element's kind, or null if it has no key. - "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition. - "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong. - "projectId": "A String", # The ID of the project to which the entities belong. - }, - "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements. - { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete. - "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future. - "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`. - "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`. - }, - ], - }, - "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`. - "a_key": # Object with schema name: Value - }, - }, + "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key. "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly. "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth. "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0]. diff --git a/docs/dyn/datastream_v1.projects.locations.connectionProfiles.html b/docs/dyn/datastream_v1.projects.locations.connectionProfiles.html new file mode 100644 index 00000000000..e70a0903dc0 --- /dev/null +++ b/docs/dyn/datastream_v1.projects.locations.connectionProfiles.html @@ -0,0 +1,653 @@ + + + +Datastream API . projects . locations . connectionProfiles
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, connectionProfileId=None, force=None, requestId=None, x__xgafv=None)
Use this method to create a connection profile in a project and location.
++
+delete(name, requestId=None, x__xgafv=None)
Use this method to delete a connection profile..
++
+discover(parent, body=None, x__xgafv=None)
Use this method to discover a connection profile. The discover API call exposes the data objects and metadata belonging to the profile. Typically, a request returns children data objects under a parent data object that's optionally supplied in the request.
+ +Use this method to get details about a connection profile.
++
+list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
Use this method to list connection profiles created in a project and location.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
++
+patch(name, body=None, force=None, requestId=None, updateMask=None, x__xgafv=None)
Use this method to update the parameters of a connection profile.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, connectionProfileId=None, force=None, requestId=None, x__xgafv=None)
+Use this method to create a connection profile in a project and location. + +Args: + parent: string, Required. The parent that owns the collection of ConnectionProfiles. (required) + body: object, The request body. + The object takes the form of: + +{ # A set of reusable connection configurations to be used as a source or destination for a stream. + "createTime": "A String", # Output only. The create time of the resource. + "displayName": "A String", # Required. Display name. + "forwardSshConnectivity": { # Forward SSH Tunnel connectivity. # Forward SSH tunnel connectivity. + "hostname": "A String", # Required. Hostname for the SSH tunnel. + "password": "A String", # Input only. SSH password. + "port": 42, # Port for the SSH tunnel, default value is 22. + "privateKey": "A String", # Input only. SSH private key. + "username": "A String", # Required. Username for the SSH tunnel. + }, + "gcsProfile": { # Cloud Storage bucket profile. # Cloud Storage ConnectionProfile configuration. + "bucket": "A String", # Required. The Cloud Storage bucket name. + "rootPath": "A String", # The root path inside the Cloud Storage bucket. + }, + "labels": { # Labels. + "a_key": "A String", + }, + "mysqlProfile": { # MySQL database profile. # MySQL ConnectionProfile configuration. + "hostname": "A String", # Required. Hostname for the MySQL connection. + "password": "A String", # Required. Input only. Password for the MySQL connection. + "port": 42, # Port for the MySQL connection, default value is 3306. + "sslConfig": { # MySQL SSL configuration information. # SSL configuration for the MySQL connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field is set. + "clientCertificate": "A String", # Input only. PEM-encoded certificate that will be used by the replica to authenticate against the source database server. If this field is used then the 'client_key' and the 'ca_certificate' fields are mandatory. + "clientCertificateSet": True or False, # Output only. Indicates whether the client_certificate field is set. + "clientKey": "A String", # Input only. PEM-encoded private key associated with the Client Certificate. If this field is used then the 'client_certificate' and the 'ca_certificate' fields are mandatory. + "clientKeySet": True or False, # Output only. Indicates whether the client_key field is set. + }, + "username": "A String", # Required. Username for the MySQL connection. + }, + "name": "A String", # Output only. The resource's name. + "oracleProfile": { # Oracle database profile. # Oracle ConnectionProfile configuration. + "connectionAttributes": { # Connection string attributes + "a_key": "A String", + }, + "databaseService": "A String", # Required. Database for the Oracle connection. + "hostname": "A String", # Required. Hostname for the Oracle connection. + "password": "A String", # Required. Password for the Oracle connection. + "port": 42, # Port for the Oracle connection, default value is 1521. + "username": "A String", # Required. Username for the Oracle connection. + }, + "privateConnectivity": { # Private Connectivity # Private connectivity. + "privateConnection": "A String", # Required. A reference to a private connection resource. Format: `projects/{project}/locations/{location}/privateConnections/{name}` + }, + "staticServiceIpConnectivity": { # Static IP address connectivity. # Static Service IP connectivity. + }, + "updateTime": "A String", # Output only. The update time of the resource. +} + + connectionProfileId: string, Required. The connection profile identifier. + force: boolean, Optional. Create the connection profile without validating it. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, requestId=None, x__xgafv=None)
+Use this method to delete a connection profile.. + +Args: + name: string, Required. The name of the connection profile resource to delete. (required) + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes after the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +discover(parent, body=None, x__xgafv=None)
+Use this method to discover a connection profile. The discover API call exposes the data objects and metadata belonging to the profile. Typically, a request returns children data objects under a parent data object that's optionally supplied in the request. + +Args: + parent: string, Required. The parent resource of the connection profile type. Must be in the format `projects/*/locations/*`. (required) + body: object, The request body. + The object takes the form of: + +{ # Request message for 'discover' ConnectionProfile request. + "connectionProfile": { # A set of reusable connection configurations to be used as a source or destination for a stream. # An ad-hoc connection profile configuration. + "createTime": "A String", # Output only. The create time of the resource. + "displayName": "A String", # Required. Display name. + "forwardSshConnectivity": { # Forward SSH Tunnel connectivity. # Forward SSH tunnel connectivity. + "hostname": "A String", # Required. Hostname for the SSH tunnel. + "password": "A String", # Input only. SSH password. + "port": 42, # Port for the SSH tunnel, default value is 22. + "privateKey": "A String", # Input only. SSH private key. + "username": "A String", # Required. Username for the SSH tunnel. + }, + "gcsProfile": { # Cloud Storage bucket profile. # Cloud Storage ConnectionProfile configuration. + "bucket": "A String", # Required. The Cloud Storage bucket name. + "rootPath": "A String", # The root path inside the Cloud Storage bucket. + }, + "labels": { # Labels. + "a_key": "A String", + }, + "mysqlProfile": { # MySQL database profile. # MySQL ConnectionProfile configuration. + "hostname": "A String", # Required. Hostname for the MySQL connection. + "password": "A String", # Required. Input only. Password for the MySQL connection. + "port": 42, # Port for the MySQL connection, default value is 3306. + "sslConfig": { # MySQL SSL configuration information. # SSL configuration for the MySQL connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field is set. + "clientCertificate": "A String", # Input only. PEM-encoded certificate that will be used by the replica to authenticate against the source database server. If this field is used then the 'client_key' and the 'ca_certificate' fields are mandatory. + "clientCertificateSet": True or False, # Output only. Indicates whether the client_certificate field is set. + "clientKey": "A String", # Input only. PEM-encoded private key associated with the Client Certificate. If this field is used then the 'client_certificate' and the 'ca_certificate' fields are mandatory. + "clientKeySet": True or False, # Output only. Indicates whether the client_key field is set. + }, + "username": "A String", # Required. Username for the MySQL connection. + }, + "name": "A String", # Output only. The resource's name. + "oracleProfile": { # Oracle database profile. # Oracle ConnectionProfile configuration. + "connectionAttributes": { # Connection string attributes + "a_key": "A String", + }, + "databaseService": "A String", # Required. Database for the Oracle connection. + "hostname": "A String", # Required. Hostname for the Oracle connection. + "password": "A String", # Required. Password for the Oracle connection. + "port": 42, # Port for the Oracle connection, default value is 1521. + "username": "A String", # Required. Username for the Oracle connection. + }, + "privateConnectivity": { # Private Connectivity # Private connectivity. + "privateConnection": "A String", # Required. A reference to a private connection resource. Format: `projects/{project}/locations/{location}/privateConnections/{name}` + }, + "staticServiceIpConnectivity": { # Static IP address connectivity. # Static Service IP connectivity. + }, + "updateTime": "A String", # Output only. The update time of the resource. + }, + "connectionProfileName": "A String", # A reference to an existing connection profile. + "fullHierarchy": True or False, # Whether to retrieve the full hierarchy of data objects (TRUE) or only the current level (FALSE). + "hierarchyDepth": 42, # The number of hierarchy levels below the current level to be retrieved. + "mysqlRdbms": { # MySQL database structure # MySQL RDBMS to enrich with child data objects and metadata. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "oracleRdbms": { # Oracle database structure. # Oracle RDBMS to enrich with child data objects and metadata. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response from a discover request. + "mysqlRdbms": { # MySQL database structure # Enriched MySQL RDBMS object. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "oracleRdbms": { # Oracle database structure. # Enriched Oracle RDBMS object. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, +}+++ +get(name, x__xgafv=None)
+Use this method to get details about a connection profile. + +Args: + name: string, Required. The name of the connection profile resource to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A set of reusable connection configurations to be used as a source or destination for a stream. + "createTime": "A String", # Output only. The create time of the resource. + "displayName": "A String", # Required. Display name. + "forwardSshConnectivity": { # Forward SSH Tunnel connectivity. # Forward SSH tunnel connectivity. + "hostname": "A String", # Required. Hostname for the SSH tunnel. + "password": "A String", # Input only. SSH password. + "port": 42, # Port for the SSH tunnel, default value is 22. + "privateKey": "A String", # Input only. SSH private key. + "username": "A String", # Required. Username for the SSH tunnel. + }, + "gcsProfile": { # Cloud Storage bucket profile. # Cloud Storage ConnectionProfile configuration. + "bucket": "A String", # Required. The Cloud Storage bucket name. + "rootPath": "A String", # The root path inside the Cloud Storage bucket. + }, + "labels": { # Labels. + "a_key": "A String", + }, + "mysqlProfile": { # MySQL database profile. # MySQL ConnectionProfile configuration. + "hostname": "A String", # Required. Hostname for the MySQL connection. + "password": "A String", # Required. Input only. Password for the MySQL connection. + "port": 42, # Port for the MySQL connection, default value is 3306. + "sslConfig": { # MySQL SSL configuration information. # SSL configuration for the MySQL connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field is set. + "clientCertificate": "A String", # Input only. PEM-encoded certificate that will be used by the replica to authenticate against the source database server. If this field is used then the 'client_key' and the 'ca_certificate' fields are mandatory. + "clientCertificateSet": True or False, # Output only. Indicates whether the client_certificate field is set. + "clientKey": "A String", # Input only. PEM-encoded private key associated with the Client Certificate. If this field is used then the 'client_certificate' and the 'ca_certificate' fields are mandatory. + "clientKeySet": True or False, # Output only. Indicates whether the client_key field is set. + }, + "username": "A String", # Required. Username for the MySQL connection. + }, + "name": "A String", # Output only. The resource's name. + "oracleProfile": { # Oracle database profile. # Oracle ConnectionProfile configuration. + "connectionAttributes": { # Connection string attributes + "a_key": "A String", + }, + "databaseService": "A String", # Required. Database for the Oracle connection. + "hostname": "A String", # Required. Hostname for the Oracle connection. + "password": "A String", # Required. Password for the Oracle connection. + "port": 42, # Port for the Oracle connection, default value is 1521. + "username": "A String", # Required. Username for the Oracle connection. + }, + "privateConnectivity": { # Private Connectivity # Private connectivity. + "privateConnection": "A String", # Required. A reference to a private connection resource. Format: `projects/{project}/locations/{location}/privateConnections/{name}` + }, + "staticServiceIpConnectivity": { # Static IP address connectivity. # Static Service IP connectivity. + }, + "updateTime": "A String", # Output only. The update time of the resource. +}+++ +list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
+Use this method to list connection profiles created in a project and location. + +Args: + parent: string, Required. The parent that owns the collection of connection profiles. (required) + filter: string, Filter request. + orderBy: string, Order by fields for the result. + pageSize: integer, Maximum number of connection profiles to return. If unspecified, at most 50 connection profiles will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000. + pageToken: string, Page token received from a previous `ListConnectionProfiles` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListConnectionProfiles` must match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response message for listing connection profiles. + "connectionProfiles": [ # List of connection profiles. + { # A set of reusable connection configurations to be used as a source or destination for a stream. + "createTime": "A String", # Output only. The create time of the resource. + "displayName": "A String", # Required. Display name. + "forwardSshConnectivity": { # Forward SSH Tunnel connectivity. # Forward SSH tunnel connectivity. + "hostname": "A String", # Required. Hostname for the SSH tunnel. + "password": "A String", # Input only. SSH password. + "port": 42, # Port for the SSH tunnel, default value is 22. + "privateKey": "A String", # Input only. SSH private key. + "username": "A String", # Required. Username for the SSH tunnel. + }, + "gcsProfile": { # Cloud Storage bucket profile. # Cloud Storage ConnectionProfile configuration. + "bucket": "A String", # Required. The Cloud Storage bucket name. + "rootPath": "A String", # The root path inside the Cloud Storage bucket. + }, + "labels": { # Labels. + "a_key": "A String", + }, + "mysqlProfile": { # MySQL database profile. # MySQL ConnectionProfile configuration. + "hostname": "A String", # Required. Hostname for the MySQL connection. + "password": "A String", # Required. Input only. Password for the MySQL connection. + "port": 42, # Port for the MySQL connection, default value is 3306. + "sslConfig": { # MySQL SSL configuration information. # SSL configuration for the MySQL connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field is set. + "clientCertificate": "A String", # Input only. PEM-encoded certificate that will be used by the replica to authenticate against the source database server. If this field is used then the 'client_key' and the 'ca_certificate' fields are mandatory. + "clientCertificateSet": True or False, # Output only. Indicates whether the client_certificate field is set. + "clientKey": "A String", # Input only. PEM-encoded private key associated with the Client Certificate. If this field is used then the 'client_certificate' and the 'ca_certificate' fields are mandatory. + "clientKeySet": True or False, # Output only. Indicates whether the client_key field is set. + }, + "username": "A String", # Required. Username for the MySQL connection. + }, + "name": "A String", # Output only. The resource's name. + "oracleProfile": { # Oracle database profile. # Oracle ConnectionProfile configuration. + "connectionAttributes": { # Connection string attributes + "a_key": "A String", + }, + "databaseService": "A String", # Required. Database for the Oracle connection. + "hostname": "A String", # Required. Hostname for the Oracle connection. + "password": "A String", # Required. Password for the Oracle connection. + "port": 42, # Port for the Oracle connection, default value is 1521. + "username": "A String", # Required. Username for the Oracle connection. + }, + "privateConnectivity": { # Private Connectivity # Private connectivity. + "privateConnection": "A String", # Required. A reference to a private connection resource. Format: `projects/{project}/locations/{location}/privateConnections/{name}` + }, + "staticServiceIpConnectivity": { # Static IP address connectivity. # Static Service IP connectivity. + }, + "updateTime": "A String", # Output only. The update time of the resource. + }, + ], + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "unreachable": [ # Locations that could not be reached. + "A String", + ], +}+++ +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ + \ No newline at end of file diff --git a/docs/dyn/datastream_v1.projects.locations.html b/docs/dyn/datastream_v1.projects.locations.html index 6b2699671c0..79f66792a30 100644 --- a/docs/dyn/datastream_v1.projects.locations.html +++ b/docs/dyn/datastream_v1.projects.locations.html @@ -74,14 +74,35 @@patch(name, body=None, force=None, requestId=None, updateMask=None, x__xgafv=None)
+Use this method to update the parameters of a connection profile. + +Args: + name: string, Output only. The resource's name. (required) + body: object, The request body. + The object takes the form of: + +{ # A set of reusable connection configurations to be used as a source or destination for a stream. + "createTime": "A String", # Output only. The create time of the resource. + "displayName": "A String", # Required. Display name. + "forwardSshConnectivity": { # Forward SSH Tunnel connectivity. # Forward SSH tunnel connectivity. + "hostname": "A String", # Required. Hostname for the SSH tunnel. + "password": "A String", # Input only. SSH password. + "port": 42, # Port for the SSH tunnel, default value is 22. + "privateKey": "A String", # Input only. SSH private key. + "username": "A String", # Required. Username for the SSH tunnel. + }, + "gcsProfile": { # Cloud Storage bucket profile. # Cloud Storage ConnectionProfile configuration. + "bucket": "A String", # Required. The Cloud Storage bucket name. + "rootPath": "A String", # The root path inside the Cloud Storage bucket. + }, + "labels": { # Labels. + "a_key": "A String", + }, + "mysqlProfile": { # MySQL database profile. # MySQL ConnectionProfile configuration. + "hostname": "A String", # Required. Hostname for the MySQL connection. + "password": "A String", # Required. Input only. Password for the MySQL connection. + "port": 42, # Port for the MySQL connection, default value is 3306. + "sslConfig": { # MySQL SSL configuration information. # SSL configuration for the MySQL connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field is set. + "clientCertificate": "A String", # Input only. PEM-encoded certificate that will be used by the replica to authenticate against the source database server. If this field is used then the 'client_key' and the 'ca_certificate' fields are mandatory. + "clientCertificateSet": True or False, # Output only. Indicates whether the client_certificate field is set. + "clientKey": "A String", # Input only. PEM-encoded private key associated with the Client Certificate. If this field is used then the 'client_certificate' and the 'ca_certificate' fields are mandatory. + "clientKeySet": True or False, # Output only. Indicates whether the client_key field is set. + }, + "username": "A String", # Required. Username for the MySQL connection. + }, + "name": "A String", # Output only. The resource's name. + "oracleProfile": { # Oracle database profile. # Oracle ConnectionProfile configuration. + "connectionAttributes": { # Connection string attributes + "a_key": "A String", + }, + "databaseService": "A String", # Required. Database for the Oracle connection. + "hostname": "A String", # Required. Hostname for the Oracle connection. + "password": "A String", # Required. Password for the Oracle connection. + "port": 42, # Port for the Oracle connection, default value is 1521. + "username": "A String", # Required. Username for the Oracle connection. + }, + "privateConnectivity": { # Private Connectivity # Private connectivity. + "privateConnection": "A String", # Required. A reference to a private connection resource. Format: `projects/{project}/locations/{location}/privateConnections/{name}` + }, + "staticServiceIpConnectivity": { # Static IP address connectivity. # Static Service IP connectivity. + }, + "updateTime": "A String", # Output only. The update time of the resource. +} + + force: boolean, Optional. Execute the update without validating it. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + updateMask: string, Optional. Field mask is used to specify the fields to be overwritten in the ConnectionProfile resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then all fields will be overwritten. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+Datastream API . projects . locations
Instance Methods
+ +Returns the connectionProfiles Resource.
+Returns the operations Resource.
+ +Returns the privateConnections Resource.
+ ++
+streams()
+Returns the streams Resource.
+Close httplib2 connections.
++
+fetchStaticIps(name, pageSize=None, pageToken=None, x__xgafv=None)
The FetchStaticIps API call exposes the static ips used by Datastream. Typically, a request returns children data objects under a parent data object that's optionally supplied in the request.
++
+fetchStaticIps_next(previous_request, previous_response)
Retrieves the next page of results.
Gets information about a location.
@@ -97,6 +118,44 @@Method Details
Close httplib2 connections.
fetchStaticIps(name, pageSize=None, pageToken=None, x__xgafv=None)
+ The FetchStaticIps API call exposes the static ips used by Datastream. Typically, a request returns children data objects under a parent data object that's optionally supplied in the request. + +Args: + name: string, Required. The name resource of the Response type. Must be in the format `projects/*/locations/*`. (required) + pageSize: integer, Maximum number of Ips to return, will likely not be specified. + pageToken: string, A page token, received from a previous `ListStaticIps` call. will likely not be specified. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response message for a 'FetchStaticIps' response. + "nextPageToken": "A String", # A token that can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "staticIps": [ # list of static ips by account + "A String", + ], +}+
fetchStaticIps_next(previous_request, previous_response)
+ Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++
get(name, x__xgafv=None)
Gets information about a location. diff --git a/docs/dyn/datastream_v1.projects.locations.privateConnections.html b/docs/dyn/datastream_v1.projects.locations.privateConnections.html new file mode 100644 index 00000000000..2ca8b802af3 --- /dev/null +++ b/docs/dyn/datastream_v1.projects.locations.privateConnections.html @@ -0,0 +1,311 @@ + + + +Datastream API . projects . locations . privateConnections
+Instance Methods
++
+routes()
+Returns the routes Resource.
+ ++
+close()
Close httplib2 connections.
++
+create(parent, body=None, privateConnectionId=None, requestId=None, x__xgafv=None)
Use this method to create a private connectivity configuration.
++
+delete(name, force=None, requestId=None, x__xgafv=None)
Use this method to delete a private connectivity configuration.
+ +Use this method to get details about a private connectivity configuration.
++
+list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
Use this method to list private connectivity configurations in a project and location.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, privateConnectionId=None, requestId=None, x__xgafv=None)
+Use this method to create a private connectivity configuration. + +Args: + parent: string, Required. The parent that owns the collection of PrivateConnections. (required) + body: object, The request body. + The object takes the form of: + +{ # The PrivateConnection resource is used to establish private connectivity between Datastream and a customer's network. + "createTime": "A String", # Output only. The create time of the resource. + "displayName": "A String", # Required. Display name. + "error": { # Represent a user-facing Error. # Output only. In case of error, the details of the error in a user-friendly format. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The resource's name. + "state": "A String", # Output only. The state of the Private Connection. + "updateTime": "A String", # Output only. The update time of the resource. + "vpcPeeringConfig": { # The VPC Peering configuration is used to create VPC peering between Datastream and the consumer's VPC. # VPC Peering Config. + "subnet": "A String", # Required. A free subnet for peering. (CIDR of /29) TODO(b/172995841) add validators. + "vpc": "A String", # Required. Fully qualified name of the VPC that Datastream will peer to. Format: `projects/{project}/global/{networks}/{name}` + }, +} + + privateConnectionId: string, Required. The private connectivity identifier. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, force=None, requestId=None, x__xgafv=None)
+Use this method to delete a private connectivity configuration. + +Args: + name: string, Required. The name of the private connectivity configuration to delete. (required) + force: boolean, Optional. If set to true, any child routes that belong to this PrivateConnection will also be deleted. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes after the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)
+Use this method to get details about a private connectivity configuration. + +Args: + name: string, Required. The name of the private connectivity configuration to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The PrivateConnection resource is used to establish private connectivity between Datastream and a customer's network. + "createTime": "A String", # Output only. The create time of the resource. + "displayName": "A String", # Required. Display name. + "error": { # Represent a user-facing Error. # Output only. In case of error, the details of the error in a user-friendly format. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The resource's name. + "state": "A String", # Output only. The state of the Private Connection. + "updateTime": "A String", # Output only. The update time of the resource. + "vpcPeeringConfig": { # The VPC Peering configuration is used to create VPC peering between Datastream and the consumer's VPC. # VPC Peering Config. + "subnet": "A String", # Required. A free subnet for peering. (CIDR of /29) TODO(b/172995841) add validators. + "vpc": "A String", # Required. Fully qualified name of the VPC that Datastream will peer to. Format: `projects/{project}/global/{networks}/{name}` + }, +}+++ +list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
+Use this method to list private connectivity configurations in a project and location. + +Args: + parent: string, Required. The parent that owns the collection of private connectivity configurations. (required) + filter: string, Filter request. + orderBy: string, Order by fields for the result. + pageSize: integer, Maximum number of private connectivity configurations to return. If unspecified, at most 50 private connectivity configurations that will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000. + pageToken: string, Page token received from a previous `ListPrivateConnections` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListPrivateConnections` must match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response containing a list of private connection configurations. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "privateConnections": [ # List of private connectivity configurations. + { # The PrivateConnection resource is used to establish private connectivity between Datastream and a customer's network. + "createTime": "A String", # Output only. The create time of the resource. + "displayName": "A String", # Required. Display name. + "error": { # Represent a user-facing Error. # Output only. In case of error, the details of the error in a user-friendly format. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The resource's name. + "state": "A String", # Output only. The state of the Private Connection. + "updateTime": "A String", # Output only. The update time of the resource. + "vpcPeeringConfig": { # The VPC Peering configuration is used to create VPC peering between Datastream and the consumer's VPC. # VPC Peering Config. + "subnet": "A String", # Required. A free subnet for peering. (CIDR of /29) TODO(b/172995841) add validators. + "vpc": "A String", # Required. Fully qualified name of the VPC that Datastream will peer to. Format: `projects/{project}/global/{networks}/{name}` + }, + }, + ], + "unreachable": [ # Locations that could not be reached. + "A String", + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/datastream_v1.projects.locations.privateConnections.routes.html b/docs/dyn/datastream_v1.projects.locations.privateConnections.routes.html new file mode 100644 index 00000000000..c4a41a6bd87 --- /dev/null +++ b/docs/dyn/datastream_v1.projects.locations.privateConnections.routes.html @@ -0,0 +1,269 @@ + + + +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Datastream API . projects . locations . privateConnections . routes
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, requestId=None, routeId=None, x__xgafv=None)
Use this method to create a route for a private connectivity in a project and location.
++
+delete(name, requestId=None, x__xgafv=None)
Use this method to delete a route.
+ +Use this method to get details about a route.
++
+list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
Use this method to list routes created for a private connectivity in a project and location.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, requestId=None, routeId=None, x__xgafv=None)
+Use this method to create a route for a private connectivity in a project and location. + +Args: + parent: string, Required. The parent that owns the collection of Routes. (required) + body: object, The request body. + The object takes the form of: + +{ # The route resource is the child of the private connection resource, used for defining a route for a private connection. + "createTime": "A String", # Output only. The create time of the resource. + "destinationAddress": "A String", # Required. Destination address for connection + "destinationPort": 42, # Destination port for connection + "displayName": "A String", # Required. Display name. + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The resource's name. + "updateTime": "A String", # Output only. The update time of the resource. +} + + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + routeId: string, Required. The Route identifier. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, requestId=None, x__xgafv=None)
+Use this method to delete a route. + +Args: + name: string, Required. The name of the Route resource to delete. (required) + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes after the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)
+Use this method to get details about a route. + +Args: + name: string, Required. The name of the Route resource to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The route resource is the child of the private connection resource, used for defining a route for a private connection. + "createTime": "A String", # Output only. The create time of the resource. + "destinationAddress": "A String", # Required. Destination address for connection + "destinationPort": 42, # Destination port for connection + "displayName": "A String", # Required. Display name. + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The resource's name. + "updateTime": "A String", # Output only. The update time of the resource. +}+++ +list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
+Use this method to list routes created for a private connectivity in a project and location. + +Args: + parent: string, Required. The parent that owns the collection of Routess. (required) + filter: string, Filter request. + orderBy: string, Order by fields for the result. + pageSize: integer, Maximum number of Routes to return. The service may return fewer than this value. If unspecified, at most 50 Routes will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000. + pageToken: string, Page token received from a previous `ListRoutes` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListRoutes` must match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Route list response. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "routes": [ # List of Routes. + { # The route resource is the child of the private connection resource, used for defining a route for a private connection. + "createTime": "A String", # Output only. The create time of the resource. + "destinationAddress": "A String", # Required. Destination address for connection + "destinationPort": 42, # Destination port for connection + "displayName": "A String", # Required. Display name. + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The resource's name. + "updateTime": "A String", # Output only. The update time of the resource. + }, + ], + "unreachable": [ # Locations that could not be reached. + "A String", + ], +}+++ + \ No newline at end of file diff --git a/docs/dyn/datastream_v1.projects.locations.streams.html b/docs/dyn/datastream_v1.projects.locations.streams.html new file mode 100644 index 00000000000..e677b2dcbd1 --- /dev/null +++ b/docs/dyn/datastream_v1.projects.locations.streams.html @@ -0,0 +1,1053 @@ + + + +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Datastream API . projects . locations . streams
+Instance Methods
++
+objects()
+Returns the objects Resource.
+ ++
+close()
Close httplib2 connections.
+ +Use this method to create a stream.
++
+delete(name, requestId=None, x__xgafv=None)
Use this method to delete a stream.
+ +Use this method to get details about a stream.
++
+list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
Use this method to list streams in a project and location.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
+ +Use this method to update the configuration of a stream.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, force=None, requestId=None, streamId=None, validateOnly=None, x__xgafv=None)
+Use this method to create a stream. + +Args: + parent: string, Required. The parent that owns the collection of streams. (required) + body: object, The request body. + The object takes the form of: + +{ # A resource representing streaming data from a source to a destination. + "backfillAll": { # Backfill strategy to automatically backfill the Stream's objects. Specific objects can be excluded. # Automatically backfill objects included in the stream source configuration. Specific objects can be excluded. + "mysqlExcludedObjects": { # MySQL database structure # MySQL data source objects to avoid backfilling. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "oracleExcludedObjects": { # Oracle database structure. # Oracle data source objects to avoid backfilling. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + }, + "backfillNone": { # Backfill strategy to disable automatic backfill for the Stream's objects. # Do not automatically backfill any objects. + }, + "createTime": "A String", # Output only. The creation time of the stream. + "customerManagedEncryptionKey": "A String", # Immutable. A reference to a KMS encryption key. If provided, it will be used to encrypt the data. If left blank, data will be encrypted using an internal Stream-specific encryption key provisioned through KMS. + "destinationConfig": { # The configuration of the stream destination. # Required. Destination connection profile configuration. + "destinationConnectionProfile": "A String", # Required. Destination connection profile resource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` + "gcsDestinationConfig": { # Google Cloud Storage destination configuration # A configuration for how data should be loaded to Cloud Storage. + "avroFileFormat": { # AVRO file format configuration. # AVRO file format configuration. + }, + "fileRotationInterval": "A String", # The maximum duration for which new events are added before a file is closed and a new file is created. + "fileRotationMb": 42, # The maximum file size to be saved in the bucket. + "jsonFileFormat": { # JSON file format configuration. # JSON file format configuration. + "compression": "A String", # Compression of the loaded JSON file. + "schemaFileFormat": "A String", # The schema file format along JSON data files. + }, + "path": "A String", # Path inside the Cloud Storage bucket to write data to. + }, + }, + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Errors on the Stream. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The stream's name. + "sourceConfig": { # The configuration of the stream source. # Required. Source connection profile configuration. + "mysqlSourceConfig": { # MySQL source configuration # MySQL data source configuration + "excludeObjects": { # MySQL database structure # MySQL objects to exclude from the stream. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "includeObjects": { # MySQL database structure # MySQL objects to retrieve from the source. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + }, + "oracleSourceConfig": { # Oracle data source configuration # Oracle data source configuration + "excludeObjects": { # Oracle database structure. # Oracle objects to exclude from the stream. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + "includeObjects": { # Oracle database structure. # Oracle objects to include in the stream. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + }, + "sourceConnectionProfile": "A String", # Required. Source connection profile resoource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` + }, + "state": "A String", # The state of the stream. + "updateTime": "A String", # Output only. The last update time of the stream. +} + + force: boolean, Optional. Create the stream without validating it. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + streamId: string, Required. The stream identifier. + validateOnly: boolean, Optional. Only validate the stream, but do not create any resources. The default is false. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +delete(name, requestId=None, x__xgafv=None)
+Use this method to delete a stream. + +Args: + name: string, Required. The name of the stream resource to delete. (required) + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes after the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)
+Use this method to get details about a stream. + +Args: + name: string, Required. The name of the stream resource to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A resource representing streaming data from a source to a destination. + "backfillAll": { # Backfill strategy to automatically backfill the Stream's objects. Specific objects can be excluded. # Automatically backfill objects included in the stream source configuration. Specific objects can be excluded. + "mysqlExcludedObjects": { # MySQL database structure # MySQL data source objects to avoid backfilling. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "oracleExcludedObjects": { # Oracle database structure. # Oracle data source objects to avoid backfilling. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + }, + "backfillNone": { # Backfill strategy to disable automatic backfill for the Stream's objects. # Do not automatically backfill any objects. + }, + "createTime": "A String", # Output only. The creation time of the stream. + "customerManagedEncryptionKey": "A String", # Immutable. A reference to a KMS encryption key. If provided, it will be used to encrypt the data. If left blank, data will be encrypted using an internal Stream-specific encryption key provisioned through KMS. + "destinationConfig": { # The configuration of the stream destination. # Required. Destination connection profile configuration. + "destinationConnectionProfile": "A String", # Required. Destination connection profile resource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` + "gcsDestinationConfig": { # Google Cloud Storage destination configuration # A configuration for how data should be loaded to Cloud Storage. + "avroFileFormat": { # AVRO file format configuration. # AVRO file format configuration. + }, + "fileRotationInterval": "A String", # The maximum duration for which new events are added before a file is closed and a new file is created. + "fileRotationMb": 42, # The maximum file size to be saved in the bucket. + "jsonFileFormat": { # JSON file format configuration. # JSON file format configuration. + "compression": "A String", # Compression of the loaded JSON file. + "schemaFileFormat": "A String", # The schema file format along JSON data files. + }, + "path": "A String", # Path inside the Cloud Storage bucket to write data to. + }, + }, + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Errors on the Stream. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The stream's name. + "sourceConfig": { # The configuration of the stream source. # Required. Source connection profile configuration. + "mysqlSourceConfig": { # MySQL source configuration # MySQL data source configuration + "excludeObjects": { # MySQL database structure # MySQL objects to exclude from the stream. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "includeObjects": { # MySQL database structure # MySQL objects to retrieve from the source. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + }, + "oracleSourceConfig": { # Oracle data source configuration # Oracle data source configuration + "excludeObjects": { # Oracle database structure. # Oracle objects to exclude from the stream. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + "includeObjects": { # Oracle database structure. # Oracle objects to include in the stream. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + }, + "sourceConnectionProfile": "A String", # Required. Source connection profile resoource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` + }, + "state": "A String", # The state of the stream. + "updateTime": "A String", # Output only. The last update time of the stream. +}+++ +list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
+Use this method to list streams in a project and location. + +Args: + parent: string, Required. The parent that owns the collection of streams. (required) + filter: string, Filter request. + orderBy: string, Order by fields for the result. + pageSize: integer, Maximum number of streams to return. If unspecified, at most 50 streams will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000. + pageToken: string, Page token received from a previous `ListStreams` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListStreams` must match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response message for listing streams. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. + "streams": [ # List of streams + { # A resource representing streaming data from a source to a destination. + "backfillAll": { # Backfill strategy to automatically backfill the Stream's objects. Specific objects can be excluded. # Automatically backfill objects included in the stream source configuration. Specific objects can be excluded. + "mysqlExcludedObjects": { # MySQL database structure # MySQL data source objects to avoid backfilling. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "oracleExcludedObjects": { # Oracle database structure. # Oracle data source objects to avoid backfilling. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + }, + "backfillNone": { # Backfill strategy to disable automatic backfill for the Stream's objects. # Do not automatically backfill any objects. + }, + "createTime": "A String", # Output only. The creation time of the stream. + "customerManagedEncryptionKey": "A String", # Immutable. A reference to a KMS encryption key. If provided, it will be used to encrypt the data. If left blank, data will be encrypted using an internal Stream-specific encryption key provisioned through KMS. + "destinationConfig": { # The configuration of the stream destination. # Required. Destination connection profile configuration. + "destinationConnectionProfile": "A String", # Required. Destination connection profile resource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` + "gcsDestinationConfig": { # Google Cloud Storage destination configuration # A configuration for how data should be loaded to Cloud Storage. + "avroFileFormat": { # AVRO file format configuration. # AVRO file format configuration. + }, + "fileRotationInterval": "A String", # The maximum duration for which new events are added before a file is closed and a new file is created. + "fileRotationMb": 42, # The maximum file size to be saved in the bucket. + "jsonFileFormat": { # JSON file format configuration. # JSON file format configuration. + "compression": "A String", # Compression of the loaded JSON file. + "schemaFileFormat": "A String", # The schema file format along JSON data files. + }, + "path": "A String", # Path inside the Cloud Storage bucket to write data to. + }, + }, + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Errors on the Stream. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The stream's name. + "sourceConfig": { # The configuration of the stream source. # Required. Source connection profile configuration. + "mysqlSourceConfig": { # MySQL source configuration # MySQL data source configuration + "excludeObjects": { # MySQL database structure # MySQL objects to exclude from the stream. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "includeObjects": { # MySQL database structure # MySQL objects to retrieve from the source. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + }, + "oracleSourceConfig": { # Oracle data source configuration # Oracle data source configuration + "excludeObjects": { # Oracle database structure. # Oracle objects to exclude from the stream. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + "includeObjects": { # Oracle database structure. # Oracle objects to include in the stream. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + }, + "sourceConnectionProfile": "A String", # Required. Source connection profile resoource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` + }, + "state": "A String", # The state of the stream. + "updateTime": "A String", # Output only. The last update time of the stream. + }, + ], + "unreachable": [ # Locations that could not be reached. + "A String", + ], +}+++ +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ + \ No newline at end of file diff --git a/docs/dyn/datastream_v1.projects.locations.streams.objects.html b/docs/dyn/datastream_v1.projects.locations.streams.objects.html new file mode 100644 index 00000000000..775cb638130 --- /dev/null +++ b/docs/dyn/datastream_v1.projects.locations.streams.objects.html @@ -0,0 +1,457 @@ + + + +patch(name, body=None, force=None, requestId=None, updateMask=None, validateOnly=None, x__xgafv=None)
+Use this method to update the configuration of a stream. + +Args: + name: string, Output only. The stream's name. (required) + body: object, The request body. + The object takes the form of: + +{ # A resource representing streaming data from a source to a destination. + "backfillAll": { # Backfill strategy to automatically backfill the Stream's objects. Specific objects can be excluded. # Automatically backfill objects included in the stream source configuration. Specific objects can be excluded. + "mysqlExcludedObjects": { # MySQL database structure # MySQL data source objects to avoid backfilling. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "oracleExcludedObjects": { # Oracle database structure. # Oracle data source objects to avoid backfilling. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + }, + "backfillNone": { # Backfill strategy to disable automatic backfill for the Stream's objects. # Do not automatically backfill any objects. + }, + "createTime": "A String", # Output only. The creation time of the stream. + "customerManagedEncryptionKey": "A String", # Immutable. A reference to a KMS encryption key. If provided, it will be used to encrypt the data. If left blank, data will be encrypted using an internal Stream-specific encryption key provisioned through KMS. + "destinationConfig": { # The configuration of the stream destination. # Required. Destination connection profile configuration. + "destinationConnectionProfile": "A String", # Required. Destination connection profile resource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` + "gcsDestinationConfig": { # Google Cloud Storage destination configuration # A configuration for how data should be loaded to Cloud Storage. + "avroFileFormat": { # AVRO file format configuration. # AVRO file format configuration. + }, + "fileRotationInterval": "A String", # The maximum duration for which new events are added before a file is closed and a new file is created. + "fileRotationMb": 42, # The maximum file size to be saved in the bucket. + "jsonFileFormat": { # JSON file format configuration. # JSON file format configuration. + "compression": "A String", # Compression of the loaded JSON file. + "schemaFileFormat": "A String", # The schema file format along JSON data files. + }, + "path": "A String", # Path inside the Cloud Storage bucket to write data to. + }, + }, + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Errors on the Stream. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "labels": { # Labels. + "a_key": "A String", + }, + "name": "A String", # Output only. The stream's name. + "sourceConfig": { # The configuration of the stream source. # Required. Source connection profile configuration. + "mysqlSourceConfig": { # MySQL source configuration # MySQL data source configuration + "excludeObjects": { # MySQL database structure # MySQL objects to exclude from the stream. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + "includeObjects": { # MySQL database structure # MySQL objects to retrieve from the source. + "mysqlDatabases": [ # Mysql databases on the server + { # MySQL database. + "database": "A String", # Database name. + "mysqlTables": [ # Tables in the database. + { # MySQL table. + "mysqlColumns": [ # MySQL columns in the database. When unspecified as part of include/exclude lists, includes/excludes everything. + { # MySQL Column. + "collation": "A String", # Column collation. + "column": "A String", # Column name. + "dataType": "A String", # The MySQL data type. Full data types list can be found here: https://dev.mysql.com/doc/refman/8.0/en/data-types.html + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "primaryKey": True or False, # Whether or not the column represents a primary key. + }, + ], + "table": "A String", # Table name. + }, + ], + }, + ], + }, + }, + "oracleSourceConfig": { # Oracle data source configuration # Oracle data source configuration + "excludeObjects": { # Oracle database structure. # Oracle objects to exclude from the stream. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + "includeObjects": { # Oracle database structure. # Oracle objects to include in the stream. + "oracleSchemas": [ # Oracle schemas/databases in the database server. + { # Oracle schema. + "oracleTables": [ # Tables in the schema. + { # Oracle table. + "oracleColumns": [ # Oracle columns in the schema. When unspecified as part of inclue/exclude lists, includes/excludes everything. + { # Oracle Column. + "column": "A String", # Column name. + "dataType": "A String", # The Oracle data type. + "encoding": "A String", # Column encoding. + "length": 42, # Column length. + "nullable": True or False, # Whether or not the column can accept a null value. + "ordinalPosition": 42, # The ordinal position of the column in the table. + "precision": 42, # Column precision. + "primaryKey": True or False, # Whether or not the column represents a primary key. + "scale": 42, # Column scale. + }, + ], + "table": "A String", # Table name. + }, + ], + "schema": "A String", # Schema name. + }, + ], + }, + }, + "sourceConnectionProfile": "A String", # Required. Source connection profile resoource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` + }, + "state": "A String", # The state of the stream. + "updateTime": "A String", # Output only. The last update time of the stream. +} + + force: boolean, Optional. Create the stream without validating it. + requestId: string, Optional. A request ID to identify requests. Specify a unique request ID so that if you must retry your request, the server will know to ignore the request if it has already been completed. The server will guarantee that for at least 60 minutes since the first request. For example, consider a situation where you make an initial request and the request times out. If you make the request again with the same request ID, the server can check if original operation with the same request ID was received, and if so, will ignore the second request. This prevents clients from accidentally creating duplicate commitments. The request ID must be a valid UUID with the exception that zero UUID is not supported (00000000-0000-0000-0000-000000000000). + updateMask: string, Optional. Field mask is used to specify the fields to be overwritten in the stream resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then all fields will be overwritten. + validateOnly: boolean, Optional. Only validate the stream with the changes, without actually updating it. The default is false. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+Datastream API . projects . locations . streams . objects
+Instance Methods
++
+close()
Close httplib2 connections.
+ +Use this method to get details about a stream object.
++
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Use this method to list the objects of a specific stream.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
++
+lookup(parent, body=None, x__xgafv=None)
Use this method to look up a stream object by its source object identifier.
++
+startBackfillJob(object, body=None, x__xgafv=None)
Starts backfill job for the specified stream object.
++
+stopBackfillJob(object, body=None, x__xgafv=None)
Stops the backfill job for the specified stream object.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +get(name, x__xgafv=None)
+Use this method to get details about a stream object. + +Args: + name: string, Required. The name of the stream object resource to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A specific stream object (e.g a specific DB table). + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object resource's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. +}+++ +list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Use this method to list the objects of a specific stream. + +Args: + parent: string, Required. The parent stream that owns the collection of objects. (required) + pageSize: integer, Maximum number of objects to return. Default is 50. The maximum value is 1000; values above 1000 will be coerced to 1000. + pageToken: string, Page token received from a previous `ListStreamObjectsRequest` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListStreamObjectsRequest` must match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response containing the objects for a stream. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. + "streamObjects": [ # List of stream objects. + { # A specific stream object (e.g a specific DB table). + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object resource's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. + }, + ], +}+++ +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ +lookup(parent, body=None, x__xgafv=None)
+Use this method to look up a stream object by its source object identifier. + +Args: + parent: string, Required. The parent stream that owns the collection of objects. (required) + body: object, The request body. + The object takes the form of: + +{ # Request for looking up a specific stream object by its source object identifier. + "sourceObjectIdentifier": { # Represents an identifier of an object in the data source. # Required. The source object identifier which maps to the stream object. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A specific stream object (e.g a specific DB table). + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object resource's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. +}+++ +startBackfillJob(object, body=None, x__xgafv=None)
+Starts backfill job for the specified stream object. + +Args: + object: string, Required. The name of the stream object resource to start a backfill job for. (required) + body: object, The request body. + The object takes the form of: + +{ # Request for manually initiating a backfill job for a specific stream object. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response for manually initiating a backfill job for a specific stream object. + "object": { # A specific stream object (e.g a specific DB table). # The stream object resource a backfill job was started for. + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object resource's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. + }, +}+++ + \ No newline at end of file diff --git a/docs/dyn/datastream_v1alpha1.projects.locations.streams.html b/docs/dyn/datastream_v1alpha1.projects.locations.streams.html index a85efb880e8..c457658d3ce 100644 --- a/docs/dyn/datastream_v1alpha1.projects.locations.streams.html +++ b/docs/dyn/datastream_v1alpha1.projects.locations.streams.html @@ -74,6 +74,11 @@stopBackfillJob(object, body=None, x__xgafv=None)
+Stops the backfill job for the specified stream object. + +Args: + object: string, Required. The name of the stream object resource to stop the backfill job for. (required) + body: object, The request body. + The object takes the form of: + +{ # Request for manually stopping a running backfill job for a specific stream object. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response for manually stop a backfill job for a specific stream object. + "object": { # A specific stream object (e.g a specific DB table). # The stream object resource the backfill job was stopped for. + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object resource's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. + }, +}+Datastream API . projects . locations . streams
Instance Methods
++
+objects()
+Returns the objects Resource.
+Close httplib2 connections.
diff --git a/docs/dyn/datastream_v1alpha1.projects.locations.streams.objects.html b/docs/dyn/datastream_v1alpha1.projects.locations.streams.objects.html new file mode 100644 index 00000000000..0aecaf999e5 --- /dev/null +++ b/docs/dyn/datastream_v1alpha1.projects.locations.streams.objects.html @@ -0,0 +1,366 @@ + + + +Datastream API . projects . locations . streams . objects
+Instance Methods
++
+close()
Close httplib2 connections.
+ +Use this method to get details about a stream object.
++
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Use this method to list the objects of a specific stream.
++
+list_next(previous_request, previous_response)
Retrieves the next page of results.
++
+startBackfillJob(object, x__xgafv=None)
Starts backfill job for the specified stream object.
++
+stopBackfillJob(object, x__xgafv=None)
Stops the backfill job for the specified stream object.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +get(name, x__xgafv=None)
+Use this method to get details about a stream object. + +Args: + name: string, Required. The name of the stream object resource to get. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A specific stream object (e.g a specific DB table). + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. +}+++ +list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Use this method to list the objects of a specific stream. + +Args: + parent: string, Required. The parent stream that owns the collection of objects. (required) + pageSize: integer, Maximum number of objects to return. Default is 50. The maximum value is 1000; values above 1000 will be coerced to 1000. + pageToken: string, Page token received from a previous `ListStreamObjectsRequest` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListStreamObjectsRequest` must match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response containing the objects for a stream. + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. + "streamObjects": [ # List of stream objects. + { # A specific stream object (e.g a specific DB table). + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. + }, + ], +}+++ +list_next(previous_request, previous_response)
+Retrieves the next page of results. + +Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + +Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ +startBackfillJob(object, x__xgafv=None)
+Starts backfill job for the specified stream object. + +Args: + object: string, Required. The name of the stream object resource to start a backfill job for. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response for manually initiating a backfill job for a specific stream object. + "object": { # A specific stream object (e.g a specific DB table). # The stream object resource a backfill job was started for. + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. + }, +}+++ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2.projects.conversations.participants.html b/docs/dyn/dialogflow_v2.projects.conversations.participants.html index 838fbc240cc..9b473b8b442 100644 --- a/docs/dyn/dialogflow_v2.projects.conversations.participants.html +++ b/docs/dyn/dialogflow_v2.projects.conversations.participants.html @@ -111,6 +111,11 @@stopBackfillJob(object, x__xgafv=None)
+Stops the backfill job for the specified stream object. + +Args: + object: string, Required. The name of the stream object resource to stop the backfill job for. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response for manually stop a backfill job for a specific stream object. + "object": { # A specific stream object (e.g a specific DB table). # The stream object resource the backfill job was stopped for. + "backfillJob": { # Represents a backfill job on a specific stream object. # The latest backfill job that was initiated for the stream object. + "errors": [ # Output only. Errors which caused the backfill job to fail. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "lastEndTime": "A String", # Output only. Backfill job's end time. + "lastStartTime": "A String", # Output only. Backfill job's start time. + "state": "A String", # Backfill job state. + "trigger": "A String", # Backfill job's triggering reason. + }, + "createTime": "A String", # Output only. The creation time of the object. + "displayName": "A String", # Required. Display name. + "errors": [ # Output only. Active errors on the object. + { # Represent a user-facing Error. + "details": { # Additional information about the error. + "a_key": "A String", + }, + "errorTime": "A String", # The time when the error occurred. + "errorUuid": "A String", # A unique identifier for this specific error, allowing it to be traced throughout the system in logs and API responses. + "message": "A String", # A message containing more information about the error that occurred. + "reason": "A String", # A title that explains the reason for the error. + }, + ], + "name": "A String", # Output only. The object's name. + "sourceObject": { # Represents an identifier of an object in the data source. # The object identifier in the data source. + "mysqlIdentifier": { # Mysql data source object identifier. # Mysql data source object identifier. + "database": "A String", # The database name. + "table": "A String", # The table name. + }, + "oracleIdentifier": { # Oracle data source object identifier. # Oracle data source object identifier. + "schema": "A String", # The schema name. + "table": "A String", # The table name. + }, + }, + "updateTime": "A String", # Output only. The last update time of the object. + }, +}+Method Details
The object takes the form of: { # The request message for Participants.AnalyzeContent. + "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "eventInput": { # Events allow for matching intents by event name instead of the natural language input. For instance, input `` can trigger a personalized welcome response. The parameter `name` may be used by the agent in the response: `"Hello #welcome_event.name! What can I do for you today?"`. # An input event to send to Dialogflow. "languageCode": "A String", # Required. The language of this query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "name": "A String", # Required. The unique identifier of the event. @@ -832,6 +837,9 @@Method Details
The object takes the form of: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -846,6 +854,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -867,6 +878,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -893,6 +907,9 @@Method Details
"nextPageToken": "A String", # Token to retrieve the next page of results or empty if there are no more results in the list. "participants": [ # The list of participants. There is a maximum number of items returned based on the page_size field in the request. { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -925,6 +942,9 @@Method Details
The object takes the form of: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -940,6 +960,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. diff --git a/docs/dyn/dialogflow_v2.projects.conversations.participants.suggestions.html b/docs/dyn/dialogflow_v2.projects.conversations.participants.suggestions.html index c2b95328171..e4d831092ec 100644 --- a/docs/dyn/dialogflow_v2.projects.conversations.participants.suggestions.html +++ b/docs/dyn/dialogflow_v2.projects.conversations.participants.suggestions.html @@ -99,6 +99,11 @@Method Details
The object takes the form of: { # The request message for Participants.SuggestArticles. + "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "contextSize": 42, # Max number of messages prior to and including latest_message to use as context when compiling the suggestion. By default 20 and at most 50. "latestMessage": "A String", # The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`. } @@ -141,6 +146,11 @@Method Details
The object takes the form of: { # The request message for Participants.SuggestFaqAnswers. + "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "contextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 20 and at most 50. "latestMessage": "A String", # The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`. } diff --git a/docs/dyn/dialogflow_v2.projects.locations.conversations.participants.html b/docs/dyn/dialogflow_v2.projects.locations.conversations.participants.html index 4820d1d331c..b5b875579d0 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.conversations.participants.html +++ b/docs/dyn/dialogflow_v2.projects.locations.conversations.participants.html @@ -111,6 +111,11 @@Method Details
The object takes the form of: { # The request message for Participants.AnalyzeContent. + "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "eventInput": { # Events allow for matching intents by event name instead of the natural language input. For instance, input `` can trigger a personalized welcome response. The parameter `name` may be used by the agent in the response: `"Hello #welcome_event.name! What can I do for you today?"`. # An input event to send to Dialogflow. "languageCode": "A String", # Required. The language of this query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "name": "A String", # Required. The unique identifier of the event. @@ -832,6 +837,9 @@Method Details
The object takes the form of: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -846,6 +854,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -867,6 +878,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -893,6 +907,9 @@Method Details
"nextPageToken": "A String", # Token to retrieve the next page of results or empty if there are no more results in the list. "participants": [ # The list of participants. There is a maximum number of items returned based on the page_size field in the request. { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -925,6 +942,9 @@Method Details
The object takes the form of: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. @@ -940,6 +960,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. "sipRecordingMediaLabel": "A String", # Optional. Label applied to streams representing this participant in SIPREC XML metadata and SDP. This is used to assign transcriptions from that media stream to this participant. This field can be updated. diff --git a/docs/dyn/dialogflow_v2.projects.locations.conversations.participants.suggestions.html b/docs/dyn/dialogflow_v2.projects.locations.conversations.participants.suggestions.html index 81c34385083..351a960452f 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.conversations.participants.suggestions.html +++ b/docs/dyn/dialogflow_v2.projects.locations.conversations.participants.suggestions.html @@ -99,6 +99,11 @@Method Details
The object takes the form of: { # The request message for Participants.SuggestArticles. + "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "contextSize": 42, # Max number of messages prior to and including latest_message to use as context when compiling the suggestion. By default 20 and at most 50. "latestMessage": "A String", # The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`. } @@ -141,6 +146,11 @@Method Details
The object takes the form of: { # The request message for Participants.SuggestFaqAnswers. + "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "contextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 20 and at most 50. "latestMessage": "A String", # The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`. } diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html index 8a3755210c6..73209a7be23 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html @@ -111,6 +111,11 @@Method Details
The object takes the form of: { # The request message for Participants.AnalyzeContent. + "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "eventInput": { # Events allow for matching intents by event name instead of the natural language input. For instance, input `` can trigger a personalized welcome response. The parameter `name` may be used by the agent in the response: `"Hello #welcome_event.name! What can I do for you today?"`. # An input event to send to Dialogflow. "languageCode": "A String", # Required. The language of this query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "name": "A String", # Required. The unique identifier of the event. @@ -1783,6 +1788,9 @@Method Details
The object takes the form of: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1797,6 +1805,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1818,6 +1829,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1844,6 +1858,9 @@Method Details
"nextPageToken": "A String", # Token to retrieve the next page of results or empty if there are no more results in the list. "participants": [ # The list of participants. There is a maximum number of items returned based on the page_size field in the request. { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1876,6 +1893,9 @@Method Details
The object takes the form of: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1891,6 +1911,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.suggestions.html index 18d1896b186..b5cd2118ca7 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.suggestions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.suggestions.html @@ -237,6 +237,11 @@Method Details
The object takes the form of: { # The request message for Participants.SuggestArticles. + "assistQueryParams": { # Represents the parameters of human assist query. # Optional. Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "contextSize": 42, # Optional. Max number of messages prior to and including latest_message to use as context when compiling the suggestion. By default 20 and at most 50. "latestMessage": "A String", # Optional. The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`. } @@ -278,6 +283,11 @@Method Details
The object takes the form of: { # The request message for Participants.SuggestFaqAnswers. + "assistQueryParams": { # Represents the parameters of human assist query. # Optional. Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "contextSize": 42, # Optional. Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 20 and at most 50. "latestMessage": "A String", # Optional. The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`. } diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html index fbe74bbd954..3da0ded626a 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html @@ -111,6 +111,11 @@Method Details
The object takes the form of: { # The request message for Participants.AnalyzeContent. + "assistQueryParams": { # Represents the parameters of human assist query. # Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "eventInput": { # Events allow for matching intents by event name instead of the natural language input. For instance, input `` can trigger a personalized welcome response. The parameter `name` may be used by the agent in the response: `"Hello #welcome_event.name! What can I do for you today?"`. # An input event to send to Dialogflow. "languageCode": "A String", # Required. The language of this query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language. "name": "A String", # Required. The unique identifier of the event. @@ -1783,6 +1788,9 @@Method Details
The object takes the form of: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1797,6 +1805,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1818,6 +1829,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1844,6 +1858,9 @@Method Details
"nextPageToken": "A String", # Token to retrieve the next page of results or empty if there are no more results in the list. "participants": [ # The list of participants. There is a maximum number of items returned based on the page_size field in the request. { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1876,6 +1893,9 @@Method Details
The object takes the form of: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. @@ -1891,6 +1911,9 @@Method Details
An object of the form: { # Represents a conversation participant (human agent, virtual agent, end-user). + "documentsMetadataFilters": { # Optional. Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, "name": "A String", # Optional. The unique identifier of this participant. Format: `projects//locations//conversations//participants/`. "obfuscatedExternalUserId": "A String", # Optional. Obfuscated user id that should be associated with the created participant. You can specify a user id as follows: 1. If you set this field in CreateParticipantRequest or UpdateParticipantRequest, Dialogflow adds the obfuscated user id with the participant. 2. If you set this field in AnalyzeContent or StreamingAnalyzeContent, Dialogflow will update Participant.obfuscated_external_user_id. Dialogflow uses this user id for following purposes: 1) Billing and measurement. If user with the same obfuscated_external_user_id is created in a later conversation, dialogflow will know it's the same user. 2) Agent assist suggestion personalization. For example, Dialogflow can use it to provide personalized smart reply suggestions for this user. Note: * Please never pass raw user ids to Dialogflow. Always obfuscate your user id first. * Dialogflow only accepts a UTF-8 encoded string, e.g., a hex digest of a hash function like SHA-512. * The length of the user id must be <= 256 characters. "role": "A String", # Immutable. The role this participant plays in the conversation. This field must be set during participant creation and is then immutable. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.suggestions.html index 3f1553ef3fb..c784acc792e 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.suggestions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.suggestions.html @@ -102,6 +102,11 @@Method Details
The object takes the form of: { # The request message for Participants.SuggestArticles. + "assistQueryParams": { # Represents the parameters of human assist query. # Optional. Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "contextSize": 42, # Optional. Max number of messages prior to and including latest_message to use as context when compiling the suggestion. By default 20 and at most 50. "latestMessage": "A String", # Optional. The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`. } @@ -143,6 +148,11 @@Method Details
The object takes the form of: { # The request message for Participants.SuggestFaqAnswers. + "assistQueryParams": { # Represents the parameters of human assist query. # Optional. Parameters for a human assist query. + "documentsMetadataFilters": { # Key-value filters on the metadata of documents returned by article suggestion. If specified, article suggestion only returns suggested documents that match all filters in their Document.metadata. Multiple values for a metadata key should be concatenated by comma. For example, filters to match all documents that have 'US' or 'CA' in their market metadata values and 'agent' in their user metadata values will be ``` documents_metadata_filters { key: "market" value: "US,CA" } documents_metadata_filters { key: "user" value: "agent" } ``` + "a_key": "A String", + }, + }, "contextSize": 42, # Optional. Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 20 and at most 50. "latestMessage": "A String", # Optional. The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`. } diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.webhooks.html b/docs/dyn/dialogflow_v3.projects.locations.agents.webhooks.html index 8221bb5f12a..78788e4daaf 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.webhooks.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.webhooks.html @@ -114,6 +114,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -124,6 +127,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -148,6 +154,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -158,6 +167,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -208,6 +220,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -218,6 +233,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -254,6 +272,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -264,6 +285,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -306,6 +330,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -316,6 +343,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -341,6 +371,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -351,6 +384,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.webhooks.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.webhooks.html index bb3d4bfbfee..12fe67ee0ca 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.webhooks.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.webhooks.html @@ -114,6 +114,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -124,6 +127,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -148,6 +154,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -158,6 +167,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -208,6 +220,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -218,6 +233,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -254,6 +272,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -264,6 +285,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -306,6 +330,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -316,6 +343,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -341,6 +371,9 @@Method Details
"disabled": True or False, # Indicates whether the webhook is disabled. "displayName": "A String", # Required. The human-readable name of the webhook, unique within the agent. "genericWebService": { # Represents configuration for a generic web service. # Configuration for a generic web service. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", @@ -351,6 +384,9 @@Method Details
"name": "A String", # The unique identifier of the webhook. Required for the Webhooks.UpdateWebhook method. Webhooks.CreateWebhook populates the name automatically. Format: `projects//locations//agents//webhooks/`. "serviceDirectory": { # Represents configuration for a [Service Directory](https://cloud.google.com/service-directory) service. # Configuration for a [Service Directory](https://cloud.google.com/service-directory) service. "genericWebService": { # Represents configuration for a generic web service. # Generic Service configuration of this webhook. + "allowedCaCerts": [ # Optional. Specifies a list of allowed custom CA certificates (in DER format) for HTTPS verification. This overrides the default SSL trust store. If this is empty or unspecified, Dialogflow will use Google's default trust store to verify certificates. N.B. Make sure the HTTPS server certificates are signed with "subject alt name". For instance a certificate can be self-signed using the following command, openssl x509 -req -days 200 -in example.com.csr \ -signkey example.com.key \ -out example.com.crt \ -extfile <(printf "\nsubjectAltName='DNS:www.example.com'") + "A String", + ], "password": "A String", # The password for HTTP Basic authentication. "requestHeaders": { # The HTTP request headers to send together with webhook requests. "a_key": "A String", diff --git a/docs/dyn/dlp_v2.organizations.deidentifyTemplates.html b/docs/dyn/dlp_v2.organizations.deidentifyTemplates.html index e63d88e3760..ed19b6db6fe 100644 --- a/docs/dyn/dlp_v2.organizations.deidentifyTemplates.html +++ b/docs/dyn/dlp_v2.organizations.deidentifyTemplates.html @@ -340,6 +340,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -617,6 +624,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -843,6 +857,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1143,6 +1164,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1420,6 +1448,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1646,6 +1681,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1968,6 +2010,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2245,6 +2294,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2471,6 +2527,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2781,6 +2844,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3058,6 +3128,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3284,6 +3361,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3601,6 +3685,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3878,6 +3969,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4104,6 +4202,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4403,6 +4508,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4680,6 +4792,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4906,6 +5025,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction diff --git a/docs/dyn/dlp_v2.organizations.locations.deidentifyTemplates.html b/docs/dyn/dlp_v2.organizations.locations.deidentifyTemplates.html index 3b219ee8e6d..3e60f60ed74 100644 --- a/docs/dyn/dlp_v2.organizations.locations.deidentifyTemplates.html +++ b/docs/dyn/dlp_v2.organizations.locations.deidentifyTemplates.html @@ -340,6 +340,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -617,6 +624,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -843,6 +857,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1143,6 +1164,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1420,6 +1448,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1646,6 +1681,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1968,6 +2010,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2245,6 +2294,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2471,6 +2527,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2781,6 +2844,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3058,6 +3128,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3284,6 +3361,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3601,6 +3685,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3878,6 +3969,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4104,6 +4202,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4403,6 +4508,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4680,6 +4792,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4906,6 +5025,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction diff --git a/docs/dyn/dlp_v2.projects.content.html b/docs/dyn/dlp_v2.projects.content.html index 17547151e55..3f44eb76944 100644 --- a/docs/dyn/dlp_v2.projects.content.html +++ b/docs/dyn/dlp_v2.projects.content.html @@ -329,6 +329,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -606,6 +613,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -832,6 +846,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1387,6 +1408,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1613,6 +1641,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1885,6 +1920,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2637,6 +2679,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2914,6 +2963,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3140,6 +3196,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3518,6 +3581,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3744,6 +3814,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4016,6 +4093,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction diff --git a/docs/dyn/dlp_v2.projects.deidentifyTemplates.html b/docs/dyn/dlp_v2.projects.deidentifyTemplates.html index 007f788de22..691c3ffa3d2 100644 --- a/docs/dyn/dlp_v2.projects.deidentifyTemplates.html +++ b/docs/dyn/dlp_v2.projects.deidentifyTemplates.html @@ -340,6 +340,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -617,6 +624,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -843,6 +857,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1143,6 +1164,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1420,6 +1448,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1646,6 +1681,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1968,6 +2010,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2245,6 +2294,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2471,6 +2527,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2781,6 +2844,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3058,6 +3128,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3284,6 +3361,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3601,6 +3685,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3878,6 +3969,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4104,6 +4202,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4403,6 +4508,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4680,6 +4792,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4906,6 +5025,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction diff --git a/docs/dyn/dlp_v2.projects.locations.content.html b/docs/dyn/dlp_v2.projects.locations.content.html index 92882455092..ccf4d0f112d 100644 --- a/docs/dyn/dlp_v2.projects.locations.content.html +++ b/docs/dyn/dlp_v2.projects.locations.content.html @@ -329,6 +329,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -606,6 +613,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -832,6 +846,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1387,6 +1408,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1613,6 +1641,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1885,6 +1920,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2637,6 +2679,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2914,6 +2963,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3140,6 +3196,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3518,6 +3581,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3744,6 +3814,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4016,6 +4093,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction diff --git a/docs/dyn/dlp_v2.projects.locations.deidentifyTemplates.html b/docs/dyn/dlp_v2.projects.locations.deidentifyTemplates.html index 3ad3ff07676..26c974e793e 100644 --- a/docs/dyn/dlp_v2.projects.locations.deidentifyTemplates.html +++ b/docs/dyn/dlp_v2.projects.locations.deidentifyTemplates.html @@ -340,6 +340,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -617,6 +624,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -843,6 +857,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1143,6 +1164,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1420,6 +1448,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1646,6 +1681,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -1968,6 +2010,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2245,6 +2294,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2471,6 +2527,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -2781,6 +2844,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3058,6 +3128,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3284,6 +3361,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3601,6 +3685,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -3878,6 +3969,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4104,6 +4202,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4403,6 +4508,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4680,6 +4792,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction @@ -4906,6 +5025,13 @@Method Details
"timestampValue": "A String", # timestamp }, }, + "replaceDictionaryConfig": { # Replace each input value with a value randomly selected from the dictionary. # Replace with a value randomly drawn (with replacement) from a dictionary. + "wordList": { # Message defining a list of words or phrases to search for in the data. # A list of words to select from for random replacement. The [limits](https://cloud.google.com/dlp/limits) page contains details about the size limits of dictionaries. + "words": [ # Words or phrases defining the dictionary. The dictionary must contain at least one phrase and every phrase must contain at least 2 characters that are letters or digits. [required] + "A String", + ], + }, + }, "replaceWithInfoTypeConfig": { # Replace each matching finding with the name of the info_type. # Replace with infotype }, "timePartConfig": { # For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a portion of the value. # Time extraction diff --git a/docs/dyn/gameservices_v1beta.projects.locations.realms.gameServerClusters.html b/docs/dyn/gameservices_v1beta.projects.locations.realms.gameServerClusters.html index 26e3337893b..baa73c6f8b9 100644 --- a/docs/dyn/gameservices_v1beta.projects.locations.realms.gameServerClusters.html +++ b/docs/dyn/gameservices_v1beta.projects.locations.realms.gameServerClusters.html @@ -120,7 +120,6 @@Method Details
The object takes the form of: { # A game server cluster resource. - "allocationPriority": "A String", # Optional. The allocation priority assigned to the game server cluster. Game server clusters receive new game server allocations based on the relative allocation priorites set for each cluster, if the realm is configured for multicluster allocation. "clusterState": { # The state of the Kubernetes cluster. # Output only. The state of the Kubernetes cluster, this will be available if 'view' is set to `FULL` in the relevant List/Get/Preview request. "agonesVersionInstalled": "A String", # Output only. The version of Agones currently installed in the registered Kubernetes cluster. "agonesVersionTargeted": "A String", # Output only. The version of Agones that is targeted to be installed in the cluster. @@ -133,9 +132,6 @@Method Details
"gkeClusterReference": { # A reference to a GKE cluster. # Reference to the GKE cluster where the game servers are installed. "cluster": "A String", # The full or partial name of a GKE cluster, using one of the following forms: * `projects/{project}/locations/{location}/clusters/{cluster}` * `locations/{location}/clusters/{cluster}` * `{cluster}` If project and location are not specified, the project and location of the GameServerCluster resource are used to generate the full name of the GKE cluster. }, - "gkeHubClusterReference": { # GkeHubClusterReference represents a reference to a Kubernetes cluster registered through GKE Hub. # Reference to a Kubernetes cluster registered through GKE Hub. See https://cloud.google.com/anthos/multicluster-management/ for more information about registering Kubernetes clusters. - "membership": "A String", # The full or partial name of a GKE Hub membership, using one of the following forms: * `https://gkehub.googleapis.com/v1beta1/projects/{project_id}/locations/global/memberships/{membership_id}` * `projects/{project_id}/locations/global/memberships/{membership_id}` * `{membership_id}` If project is not specified, the project of the GameServerCluster resource is used to generate the full name of the GKE Hub membership. - }, "namespace": "A String", # Namespace designated on the game server cluster where the Agones game server instances will be created. Existence of the namespace will be validated during creation. }, "createTime": "A String", # Output only. The creation time. @@ -233,7 +229,6 @@Method Details
An object of the form: { # A game server cluster resource. - "allocationPriority": "A String", # Optional. The allocation priority assigned to the game server cluster. Game server clusters receive new game server allocations based on the relative allocation priorites set for each cluster, if the realm is configured for multicluster allocation. "clusterState": { # The state of the Kubernetes cluster. # Output only. The state of the Kubernetes cluster, this will be available if 'view' is set to `FULL` in the relevant List/Get/Preview request. "agonesVersionInstalled": "A String", # Output only. The version of Agones currently installed in the registered Kubernetes cluster. "agonesVersionTargeted": "A String", # Output only. The version of Agones that is targeted to be installed in the cluster. @@ -246,9 +241,6 @@Method Details
"gkeClusterReference": { # A reference to a GKE cluster. # Reference to the GKE cluster where the game servers are installed. "cluster": "A String", # The full or partial name of a GKE cluster, using one of the following forms: * `projects/{project}/locations/{location}/clusters/{cluster}` * `locations/{location}/clusters/{cluster}` * `{cluster}` If project and location are not specified, the project and location of the GameServerCluster resource are used to generate the full name of the GKE cluster. }, - "gkeHubClusterReference": { # GkeHubClusterReference represents a reference to a Kubernetes cluster registered through GKE Hub. # Reference to a Kubernetes cluster registered through GKE Hub. See https://cloud.google.com/anthos/multicluster-management/ for more information about registering Kubernetes clusters. - "membership": "A String", # The full or partial name of a GKE Hub membership, using one of the following forms: * `https://gkehub.googleapis.com/v1beta1/projects/{project_id}/locations/global/memberships/{membership_id}` * `projects/{project_id}/locations/global/memberships/{membership_id}` * `{membership_id}` If project is not specified, the project of the GameServerCluster resource is used to generate the full name of the GKE Hub membership. - }, "namespace": "A String", # Namespace designated on the game server cluster where the Agones game server instances will be created. Existence of the namespace will be validated during creation. }, "createTime": "A String", # Output only. The creation time. @@ -288,7 +280,6 @@Method Details
{ # Response message for GameServerClustersService.ListGameServerClusters. "gameServerClusters": [ # The list of game server clusters. { # A game server cluster resource. - "allocationPriority": "A String", # Optional. The allocation priority assigned to the game server cluster. Game server clusters receive new game server allocations based on the relative allocation priorites set for each cluster, if the realm is configured for multicluster allocation. "clusterState": { # The state of the Kubernetes cluster. # Output only. The state of the Kubernetes cluster, this will be available if 'view' is set to `FULL` in the relevant List/Get/Preview request. "agonesVersionInstalled": "A String", # Output only. The version of Agones currently installed in the registered Kubernetes cluster. "agonesVersionTargeted": "A String", # Output only. The version of Agones that is targeted to be installed in the cluster. @@ -301,9 +292,6 @@Method Details
"gkeClusterReference": { # A reference to a GKE cluster. # Reference to the GKE cluster where the game servers are installed. "cluster": "A String", # The full or partial name of a GKE cluster, using one of the following forms: * `projects/{project}/locations/{location}/clusters/{cluster}` * `locations/{location}/clusters/{cluster}` * `{cluster}` If project and location are not specified, the project and location of the GameServerCluster resource are used to generate the full name of the GKE cluster. }, - "gkeHubClusterReference": { # GkeHubClusterReference represents a reference to a Kubernetes cluster registered through GKE Hub. # Reference to a Kubernetes cluster registered through GKE Hub. See https://cloud.google.com/anthos/multicluster-management/ for more information about registering Kubernetes clusters. - "membership": "A String", # The full or partial name of a GKE Hub membership, using one of the following forms: * `https://gkehub.googleapis.com/v1beta1/projects/{project_id}/locations/global/memberships/{membership_id}` * `projects/{project_id}/locations/global/memberships/{membership_id}` * `{membership_id}` If project is not specified, the project of the GameServerCluster resource is used to generate the full name of the GKE Hub membership. - }, "namespace": "A String", # Namespace designated on the game server cluster where the Agones game server instances will be created. Existence of the namespace will be validated during creation. }, "createTime": "A String", # Output only. The creation time. @@ -347,7 +335,6 @@Method Details
The object takes the form of: { # A game server cluster resource. - "allocationPriority": "A String", # Optional. The allocation priority assigned to the game server cluster. Game server clusters receive new game server allocations based on the relative allocation priorites set for each cluster, if the realm is configured for multicluster allocation. "clusterState": { # The state of the Kubernetes cluster. # Output only. The state of the Kubernetes cluster, this will be available if 'view' is set to `FULL` in the relevant List/Get/Preview request. "agonesVersionInstalled": "A String", # Output only. The version of Agones currently installed in the registered Kubernetes cluster. "agonesVersionTargeted": "A String", # Output only. The version of Agones that is targeted to be installed in the cluster. @@ -360,9 +347,6 @@Method Details
"gkeClusterReference": { # A reference to a GKE cluster. # Reference to the GKE cluster where the game servers are installed. "cluster": "A String", # The full or partial name of a GKE cluster, using one of the following forms: * `projects/{project}/locations/{location}/clusters/{cluster}` * `locations/{location}/clusters/{cluster}` * `{cluster}` If project and location are not specified, the project and location of the GameServerCluster resource are used to generate the full name of the GKE cluster. }, - "gkeHubClusterReference": { # GkeHubClusterReference represents a reference to a Kubernetes cluster registered through GKE Hub. # Reference to a Kubernetes cluster registered through GKE Hub. See https://cloud.google.com/anthos/multicluster-management/ for more information about registering Kubernetes clusters. - "membership": "A String", # The full or partial name of a GKE Hub membership, using one of the following forms: * `https://gkehub.googleapis.com/v1beta1/projects/{project_id}/locations/global/memberships/{membership_id}` * `projects/{project_id}/locations/global/memberships/{membership_id}` * `{membership_id}` If project is not specified, the project of the GameServerCluster resource is used to generate the full name of the GKE Hub membership. - }, "namespace": "A String", # Namespace designated on the game server cluster where the Agones game server instances will be created. Existence of the namespace will be validated during creation. }, "createTime": "A String", # Output only. The creation time. @@ -415,7 +399,6 @@Method Details
The object takes the form of: { # A game server cluster resource. - "allocationPriority": "A String", # Optional. The allocation priority assigned to the game server cluster. Game server clusters receive new game server allocations based on the relative allocation priorites set for each cluster, if the realm is configured for multicluster allocation. "clusterState": { # The state of the Kubernetes cluster. # Output only. The state of the Kubernetes cluster, this will be available if 'view' is set to `FULL` in the relevant List/Get/Preview request. "agonesVersionInstalled": "A String", # Output only. The version of Agones currently installed in the registered Kubernetes cluster. "agonesVersionTargeted": "A String", # Output only. The version of Agones that is targeted to be installed in the cluster. @@ -428,9 +411,6 @@Method Details
"gkeClusterReference": { # A reference to a GKE cluster. # Reference to the GKE cluster where the game servers are installed. "cluster": "A String", # The full or partial name of a GKE cluster, using one of the following forms: * `projects/{project}/locations/{location}/clusters/{cluster}` * `locations/{location}/clusters/{cluster}` * `{cluster}` If project and location are not specified, the project and location of the GameServerCluster resource are used to generate the full name of the GKE cluster. }, - "gkeHubClusterReference": { # GkeHubClusterReference represents a reference to a Kubernetes cluster registered through GKE Hub. # Reference to a Kubernetes cluster registered through GKE Hub. See https://cloud.google.com/anthos/multicluster-management/ for more information about registering Kubernetes clusters. - "membership": "A String", # The full or partial name of a GKE Hub membership, using one of the following forms: * `https://gkehub.googleapis.com/v1beta1/projects/{project_id}/locations/global/memberships/{membership_id}` * `projects/{project_id}/locations/global/memberships/{membership_id}` * `{membership_id}` If project is not specified, the project of the GameServerCluster resource is used to generate the full name of the GKE Hub membership. - }, "namespace": "A String", # Namespace designated on the game server cluster where the Agones game server instances will be created. Existence of the namespace will be validated during creation. }, "createTime": "A String", # Output only. The creation time. @@ -553,7 +533,6 @@Method Details
The object takes the form of: { # A game server cluster resource. - "allocationPriority": "A String", # Optional. The allocation priority assigned to the game server cluster. Game server clusters receive new game server allocations based on the relative allocation priorites set for each cluster, if the realm is configured for multicluster allocation. "clusterState": { # The state of the Kubernetes cluster. # Output only. The state of the Kubernetes cluster, this will be available if 'view' is set to `FULL` in the relevant List/Get/Preview request. "agonesVersionInstalled": "A String", # Output only. The version of Agones currently installed in the registered Kubernetes cluster. "agonesVersionTargeted": "A String", # Output only. The version of Agones that is targeted to be installed in the cluster. @@ -566,9 +545,6 @@Method Details
"gkeClusterReference": { # A reference to a GKE cluster. # Reference to the GKE cluster where the game servers are installed. "cluster": "A String", # The full or partial name of a GKE cluster, using one of the following forms: * `projects/{project}/locations/{location}/clusters/{cluster}` * `locations/{location}/clusters/{cluster}` * `{cluster}` If project and location are not specified, the project and location of the GameServerCluster resource are used to generate the full name of the GKE cluster. }, - "gkeHubClusterReference": { # GkeHubClusterReference represents a reference to a Kubernetes cluster registered through GKE Hub. # Reference to a Kubernetes cluster registered through GKE Hub. See https://cloud.google.com/anthos/multicluster-management/ for more information about registering Kubernetes clusters. - "membership": "A String", # The full or partial name of a GKE Hub membership, using one of the following forms: * `https://gkehub.googleapis.com/v1beta1/projects/{project_id}/locations/global/memberships/{membership_id}` * `projects/{project_id}/locations/global/memberships/{membership_id}` * `{membership_id}` If project is not specified, the project of the GameServerCluster resource is used to generate the full name of the GKE Hub membership. - }, "namespace": "A String", # Namespace designated on the game server cluster where the Agones game server instances will be created. Existence of the namespace will be validated during creation. }, "createTime": "A String", # Output only. The creation time. diff --git a/docs/dyn/gkehub_v1alpha.projects.locations.fleets.html b/docs/dyn/gkehub_v1alpha.projects.locations.fleets.html new file mode 100644 index 00000000000..9052e9d3f3c --- /dev/null +++ b/docs/dyn/gkehub_v1alpha.projects.locations.fleets.html @@ -0,0 +1,218 @@ + + + +GKE Hub API . projects . locations . fleets
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, x__xgafv=None)
Creates a fleet.
+ +Removes a Fleet. There must be no memberships remaining in the Fleet.
+ +Returns the details of a fleet.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)
Updates a fleet.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, x__xgafv=None)
+Creates a fleet. + +Args: + parent: string, Required. The parent (project and location) where the Fleet will be created. Specified in the format `projects/*/locations/*`. (required) + body: object, The request body. + The object takes the form of: + +{ # Fleet contains the Fleet-wide metadata and configuration. + "createTime": "A String", # Output only. When the Fleet was created. + "deleteTime": "A String", # Output only. When the Fleet was deleted. + "displayName": "A String", # Optional. A user-assigned display name of the Fleet. When present, it must be between 4 to 30 characters. Allowed characters are: lowercase and uppercase letters, numbers, hyphen, single-quote, double-quote, space, and exclamation point. Example: `Production Fleet` + "fleetName": "A String", # The name for the fleet. The name must meet the following constraints: + The name of a fleet should be unique within the organization; + It must consist of lower case alphanumeric characters or `-`; + The length of the name must be less than or equal to 63; + Unicode names must be expressed in Punycode format (rfc3492). Examples: + prod-fleet + xn--wlq33vhyw9jb (Punycode form for "生产环境") + "name": "A String", # Output only. The full, unique resource name of this fleet in the format of `projects/{project}/locations/{location}/fleets/{fleet}`. Each GCP project can have at most one fleet resource, named "default". + "uid": "A String", # Output only. Google-generated UUID for this resource. This is unique across all Fleet resources. If a Fleet resource is deleted and another resource with the same name is created, it gets a different uid. + "updateTime": "A String", # Output only. When the Fleet was last updated. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Fleet contains the Fleet-wide metadata and configuration. + "createTime": "A String", # Output only. When the Fleet was created. + "deleteTime": "A String", # Output only. When the Fleet was deleted. + "displayName": "A String", # Optional. A user-assigned display name of the Fleet. When present, it must be between 4 to 30 characters. Allowed characters are: lowercase and uppercase letters, numbers, hyphen, single-quote, double-quote, space, and exclamation point. Example: `Production Fleet` + "fleetName": "A String", # The name for the fleet. The name must meet the following constraints: + The name of a fleet should be unique within the organization; + It must consist of lower case alphanumeric characters or `-`; + The length of the name must be less than or equal to 63; + Unicode names must be expressed in Punycode format (rfc3492). Examples: + prod-fleet + xn--wlq33vhyw9jb (Punycode form for "生产环境") + "name": "A String", # Output only. The full, unique resource name of this fleet in the format of `projects/{project}/locations/{location}/fleets/{fleet}`. Each GCP project can have at most one fleet resource, named "default". + "uid": "A String", # Output only. Google-generated UUID for this resource. This is unique across all Fleet resources. If a Fleet resource is deleted and another resource with the same name is created, it gets a different uid. + "updateTime": "A String", # Output only. When the Fleet was last updated. +}+++ +delete(name, x__xgafv=None)
+Removes a Fleet. There must be no memberships remaining in the Fleet. + +Args: + name: string, Required. The Fleet resource name in the format `projects/*/locations/*/fleets/*`. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`. +}+++ +get(name, x__xgafv=None)
+Returns the details of a fleet. + +Args: + name: string, Required. The Fleet resource name in the format `projects/*/locations/*/fleets/*`. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Fleet contains the Fleet-wide metadata and configuration. + "createTime": "A String", # Output only. When the Fleet was created. + "deleteTime": "A String", # Output only. When the Fleet was deleted. + "displayName": "A String", # Optional. A user-assigned display name of the Fleet. When present, it must be between 4 to 30 characters. Allowed characters are: lowercase and uppercase letters, numbers, hyphen, single-quote, double-quote, space, and exclamation point. Example: `Production Fleet` + "fleetName": "A String", # The name for the fleet. The name must meet the following constraints: + The name of a fleet should be unique within the organization; + It must consist of lower case alphanumeric characters or `-`; + The length of the name must be less than or equal to 63; + Unicode names must be expressed in Punycode format (rfc3492). Examples: + prod-fleet + xn--wlq33vhyw9jb (Punycode form for "生产环境") + "name": "A String", # Output only. The full, unique resource name of this fleet in the format of `projects/{project}/locations/{location}/fleets/{fleet}`. Each GCP project can have at most one fleet resource, named "default". + "uid": "A String", # Output only. Google-generated UUID for this resource. This is unique across all Fleet resources. If a Fleet resource is deleted and another resource with the same name is created, it gets a different uid. + "updateTime": "A String", # Output only. When the Fleet was last updated. +}+++ + \ No newline at end of file diff --git a/docs/dyn/gkehub_v1alpha.projects.locations.html b/docs/dyn/gkehub_v1alpha.projects.locations.html index 57eabce8112..7961ff53450 100644 --- a/docs/dyn/gkehub_v1alpha.projects.locations.html +++ b/docs/dyn/gkehub_v1alpha.projects.locations.html @@ -79,6 +79,11 @@patch(name, body=None, updateMask=None, x__xgafv=None)
+Updates a fleet. + +Args: + name: string, Output only. The full, unique resource name of this fleet in the format of `projects/{project}/locations/{location}/fleets/{fleet}`. Each GCP project can have at most one fleet resource, named "default". (required) + body: object, The request body. + The object takes the form of: + +{ # Fleet contains the Fleet-wide metadata and configuration. + "createTime": "A String", # Output only. When the Fleet was created. + "deleteTime": "A String", # Output only. When the Fleet was deleted. + "displayName": "A String", # Optional. A user-assigned display name of the Fleet. When present, it must be between 4 to 30 characters. Allowed characters are: lowercase and uppercase letters, numbers, hyphen, single-quote, double-quote, space, and exclamation point. Example: `Production Fleet` + "fleetName": "A String", # The name for the fleet. The name must meet the following constraints: + The name of a fleet should be unique within the organization; + It must consist of lower case alphanumeric characters or `-`; + The length of the name must be less than or equal to 63; + Unicode names must be expressed in Punycode format (rfc3492). Examples: + prod-fleet + xn--wlq33vhyw9jb (Punycode form for "生产环境") + "name": "A String", # Output only. The full, unique resource name of this fleet in the format of `projects/{project}/locations/{location}/fleets/{fleet}`. Each GCP project can have at most one fleet resource, named "default". + "uid": "A String", # Output only. Google-generated UUID for this resource. This is unique across all Fleet resources. If a Fleet resource is deleted and another resource with the same name is created, it gets a different uid. + "updateTime": "A String", # Output only. When the Fleet was last updated. +} + + updateMask: string, Required. The fields to be updated; + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Fleet contains the Fleet-wide metadata and configuration. + "createTime": "A String", # Output only. When the Fleet was created. + "deleteTime": "A String", # Output only. When the Fleet was deleted. + "displayName": "A String", # Optional. A user-assigned display name of the Fleet. When present, it must be between 4 to 30 characters. Allowed characters are: lowercase and uppercase letters, numbers, hyphen, single-quote, double-quote, space, and exclamation point. Example: `Production Fleet` + "fleetName": "A String", # The name for the fleet. The name must meet the following constraints: + The name of a fleet should be unique within the organization; + It must consist of lower case alphanumeric characters or `-`; + The length of the name must be less than or equal to 63; + Unicode names must be expressed in Punycode format (rfc3492). Examples: + prod-fleet + xn--wlq33vhyw9jb (Punycode form for "生产环境") + "name": "A String", # Output only. The full, unique resource name of this fleet in the format of `projects/{project}/locations/{location}/fleets/{fleet}`. Each GCP project can have at most one fleet resource, named "default". + "uid": "A String", # Output only. Google-generated UUID for this resource. This is unique across all Fleet resources. If a Fleet resource is deleted and another resource with the same name is created, it gets a different uid. + "updateTime": "A String", # Output only. When the Fleet was last updated. +}+Instance Methods
Returns the features Resource.
++
+fleets()
+Returns the fleets Resource.
+ diff --git a/docs/dyn/index.md b/docs/dyn/index.md index cab18eb10d6..0938dccc54e 100644 --- a/docs/dyn/index.md +++ b/docs/dyn/index.md @@ -217,6 +217,10 @@ * [v2](http://googleapis.github.io/google-api-python-client/docs/dyn/clouddebugger_v2.html) +## clouddeploy +* [v1](http://googleapis.github.io/google-api-python-client/docs/dyn/clouddeploy_v1.html) + + ## clouderrorreporting * [v1beta1](http://googleapis.github.io/google-api-python-client/docs/dyn/clouderrorreporting_v1beta1.html) diff --git a/docs/dyn/logging_v2.billingAccounts.buckets.html b/docs/dyn/logging_v2.billingAccounts.buckets.html index 195d0cc8401..367338fd0b8 100644 --- a/docs/dyn/logging_v2.billingAccounts.buckets.html +++ b/docs/dyn/logging_v2.billingAccounts.buckets.html @@ -106,6 +106,11 @@Method Details
An object of the form: { # Describes a repository in which log entries are stored. + "cmekSettings": { # Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Log Router can currently only be configured for Google Cloud organizations. Once configured, it applies to all projects and folders in the Google Cloud organization.See Enabling CMEK for Log Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. # The CMEK settings of the log bucket. If present, new log entries written to this log bucket are encrypted using the CMEK key provided in this configuration. If a log bucket has CMEK settings, the CMEK settings cannot be disabled later by updating the log bucket. Changing the KMS key is allowed. + "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/[PROJECT_ID]/locations/[LOCATION]/keyRings/[KEYRING]/cryptoKeys/[KEY]" For example:"projects/my-project/locations/us-central1/keyRings/my-ring/cryptoKeys/my-key"To enable CMEK for the Log Router, set this field to a valid kms_key_name for which the associated service account has the required cloudkms.cryptoKeyEncrypterDecrypter roles assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name or disabled by setting the key name to an empty string. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Log Router, set this field to an empty string.See Enabling CMEK for Log Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. + "name": "A String", # Output only. The resource name of the CMEK settings. + "serviceAccountId": "A String", # Output only. The service account that will be used by the Log Router to access your Cloud KMS key.Before enabling CMEK for Log Router, you must first assign the cloudkms.cryptoKeyEncrypterDecrypter role to the service account that the Log Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. + }, "createTime": "A String", # Output only. The creation timestamp of the bucket. This is not set for any of the default buckets. "description": "A String", # Describes this bucket. "lifecycleState": "A String", # Output only. The bucket lifecycle state. diff --git a/docs/dyn/logging_v2.billingAccounts.html b/docs/dyn/logging_v2.billingAccounts.html index 0bd568dcb61..c2bee3b6202 100644 --- a/docs/dyn/logging_v2.billingAccounts.html +++ b/docs/dyn/logging_v2.billingAccounts.html @@ -107,10 +107,34 @@Instance Methods
Close httplib2 connections.
++
+getCmekSettings(name, x__xgafv=None)
Gets the Logging CMEK settings for the given resource.Note: CMEK for the Log Router can be configured for Google Cloud projects, folders, organizations and billing accounts. Once configured for an organization, it applies to all projects and folders in the Google Cloud organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information.
Method Details
+close()
Close httplib2 connections.++getCmekSettings(name, x__xgafv=None)
+Gets the Logging CMEK settings for the given resource.Note: CMEK for the Log Router can be configured for Google Cloud projects, folders, organizations and billing accounts. Once configured for an organization, it applies to all projects and folders in the Google Cloud organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. + +Args: + name: string, Required. The resource for which to retrieve CMEK settings. "projects/[PROJECT_ID]/cmekSettings" "organizations/[ORGANIZATION_ID]/cmekSettings" "billingAccounts/[BILLING_ACCOUNT_ID]/cmekSettings" "folders/[FOLDER_ID]/cmekSettings" For example:"organizations/12345/cmekSettings"Note: CMEK for the Log Router can be configured for Google Cloud projects, folders, organizations and billing accounts. Once configured for an organization, it applies to all projects and folders in the Google Cloud organization. (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Log Router can currently only be configured for Google Cloud organizations. Once configured, it applies to all projects and folders in the Google Cloud organization.See Enabling CMEK for Log Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. + "kmsKeyName": "A String", # The resource name for the configured Cloud KMS key.KMS key name format: "projects/[PROJECT_ID]/locations/[LOCATION]/keyRings/[KEYRING]/cryptoKeys/[KEY]" For example:"projects/my-project/locations/us-central1/keyRings/my-ring/cryptoKeys/my-key"To enable CMEK for the Log Router, set this field to a valid kms_key_name for which the associated service account has the required cloudkms.cryptoKeyEncrypterDecrypter roles assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name or disabled by setting the key name to an empty string. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Log Router, set this field to an empty string.See Enabling CMEK for Log Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. + "name": "A String", # Output only. The resource name of the CMEK settings. + "serviceAccountId": "A String", # Output only. The service account that will be used by the Log Router to access your Cloud KMS key.Before enabling CMEK for Log Router, you must first assign the cloudkms.cryptoKeyEncrypterDecrypter role to the service account that the Log Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. +}+