Skip to content

Commit

Permalink
feat: add enable_private_service_connect field to Endpoint feat: add …
Browse files Browse the repository at this point in the history
…id field to DeployedModel feat: add service_attachment field to PrivateEndpoints feat: add endpoint_id to CreateEndpointRequest and method signature to CreateEndpoint feat: add method... (#878)

* feat: add enable_private_service_connect field to Endpoint feat: add id field to DeployedModel feat: add service_attachment field to PrivateEndpoints feat: add endpoint_id to CreateEndpointRequest and method signature to CreateEndpoint feat: add method signature to CreateFeatureStore, CreateEntityType, CreateFeature feat: add network and enable_private_service_connect to IndexEndpoint feat: add service_attachment to IndexPrivateEndpoints feat: add stratified_split field to training_pipeline InputDataConfig fix: remove invalid resource annotations in LineageSubgraph

PiperOrigin-RevId: 413686247

Source-Link: googleapis/googleapis@244a89d

Source-Link: googleapis/googleapis-gen@c485e44
Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiYzQ4NWU0NGExYjJmZWY1MTZlOWJjYTM2NTE0ZDUwY2ViZDVlYTUxZiJ9

* 🦉 Updates from OwlBot

See https://github.com/googleapis/repo-automation-bots/blob/main/packages/owl-bot/README.md

Co-authored-by: Owl Bot <gcf-owl-bot[bot]@users.noreply.github.com>
  • Loading branch information
gcf-owl-bot[bot] and gcf-owl-bot[bot] committed Dec 2, 2021
1 parent bffbd9d commit ca813be
Show file tree
Hide file tree
Showing 59 changed files with 1,355 additions and 178 deletions.
22 changes: 11 additions & 11 deletions google/cloud/aiplatform_v1/services/migration_service/client.py
Expand Up @@ -183,19 +183,16 @@ def parse_annotated_dataset_path(path: str) -> Dict[str, str]:
return m.groupdict() if m else {}

@staticmethod
def dataset_path(project: str, location: str, dataset: str,) -> str:
def dataset_path(project: str, dataset: str,) -> str:
"""Returns a fully-qualified dataset string."""
return "projects/{project}/locations/{location}/datasets/{dataset}".format(
project=project, location=location, dataset=dataset,
return "projects/{project}/datasets/{dataset}".format(
project=project, dataset=dataset,
)

@staticmethod
def parse_dataset_path(path: str) -> Dict[str, str]:
"""Parses a dataset path into its component segments."""
m = re.match(
r"^projects/(?P<project>.+?)/locations/(?P<location>.+?)/datasets/(?P<dataset>.+?)$",
path,
)
m = re.match(r"^projects/(?P<project>.+?)/datasets/(?P<dataset>.+?)$", path)
return m.groupdict() if m else {}

@staticmethod
Expand All @@ -215,16 +212,19 @@ def parse_dataset_path(path: str) -> Dict[str, str]:
return m.groupdict() if m else {}

@staticmethod
def dataset_path(project: str, dataset: str,) -> str:
def dataset_path(project: str, location: str, dataset: str,) -> str:
"""Returns a fully-qualified dataset string."""
return "projects/{project}/datasets/{dataset}".format(
project=project, dataset=dataset,
return "projects/{project}/locations/{location}/datasets/{dataset}".format(
project=project, location=location, dataset=dataset,
)

@staticmethod
def parse_dataset_path(path: str) -> Dict[str, str]:
"""Parses a dataset path into its component segments."""
m = re.match(r"^projects/(?P<project>.+?)/datasets/(?P<dataset>.+?)$", path)
m = re.match(
r"^projects/(?P<project>.+?)/locations/(?P<location>.+?)/datasets/(?P<dataset>.+?)$",
path,
)
return m.groupdict() if m else {}

@staticmethod
Expand Down
12 changes: 12 additions & 0 deletions google/cloud/aiplatform_v1beta1/__init__.py
Expand Up @@ -112,6 +112,7 @@
from .types.event import Event
from .types.execution import Execution
from .types.explanation import Attribution
from .types.explanation import BlurBaselineConfig
from .types.explanation import Explanation
from .types.explanation import ExplanationMetadataOverride
from .types.explanation import ExplanationParameters
Expand Down Expand Up @@ -189,6 +190,9 @@
from .types.index_endpoint_service import GetIndexEndpointRequest
from .types.index_endpoint_service import ListIndexEndpointsRequest
from .types.index_endpoint_service import ListIndexEndpointsResponse
from .types.index_endpoint_service import MutateDeployedIndexOperationMetadata
from .types.index_endpoint_service import MutateDeployedIndexRequest
from .types.index_endpoint_service import MutateDeployedIndexResponse
from .types.index_endpoint_service import UndeployIndexOperationMetadata
from .types.index_endpoint_service import UndeployIndexRequest
from .types.index_endpoint_service import UndeployIndexResponse
Expand Down Expand Up @@ -447,12 +451,14 @@
from .types.training_pipeline import FractionSplit
from .types.training_pipeline import InputDataConfig
from .types.training_pipeline import PredefinedSplit
from .types.training_pipeline import StratifiedSplit
from .types.training_pipeline import TimestampSplit
from .types.training_pipeline import TrainingPipeline
from .types.types import BoolArray
from .types.types import DoubleArray
from .types.types import Int64Array
from .types.types import StringArray
from .types.unmanaged_container_model import UnmanagedContainerModel
from .types.user_action_reference import UserActionReference
from .types.value import Value
from .types.vizier_service import AddTrialMeasurementRequest
Expand Down Expand Up @@ -529,6 +535,7 @@
"BatchReadTensorboardTimeSeriesDataResponse",
"BigQueryDestination",
"BigQuerySource",
"BlurBaselineConfig",
"BoolArray",
"CancelBatchPredictionJobRequest",
"CancelCustomJobRequest",
Expand Down Expand Up @@ -813,6 +820,9 @@
"ModelMonitoringObjectiveConfig",
"ModelMonitoringStatsAnomalies",
"ModelServiceClient",
"MutateDeployedIndexOperationMetadata",
"MutateDeployedIndexRequest",
"MutateDeployedIndexResponse",
"NearestNeighborSearchOperationMetadata",
"PauseModelDeploymentMonitoringJobRequest",
"PipelineJob",
Expand Down Expand Up @@ -866,6 +876,7 @@
"SpecialistPool",
"SpecialistPoolServiceClient",
"StopTrialRequest",
"StratifiedSplit",
"StreamingReadFeatureValuesRequest",
"StringArray",
"Study",
Expand Down Expand Up @@ -895,6 +906,7 @@
"UndeployModelOperationMetadata",
"UndeployModelRequest",
"UndeployModelResponse",
"UnmanagedContainerModel",
"UpdateArtifactRequest",
"UpdateContextRequest",
"UpdateDatasetRequest",
Expand Down
10 changes: 10 additions & 0 deletions google/cloud/aiplatform_v1beta1/gapic_metadata.json
Expand Up @@ -481,6 +481,11 @@
"list_index_endpoints"
]
},
"MutateDeployedIndex": {
"methods": [
"mutate_deployed_index"
]
},
"UndeployIndex": {
"methods": [
"undeploy_index"
Expand Down Expand Up @@ -521,6 +526,11 @@
"list_index_endpoints"
]
},
"MutateDeployedIndex": {
"methods": [
"mutate_deployed_index"
]
},
"UndeployIndex": {
"methods": [
"undeploy_index"
Expand Down
Expand Up @@ -191,6 +191,7 @@ async def create_endpoint(
*,
parent: str = None,
endpoint: gca_endpoint.Endpoint = None,
endpoint_id: str = None,
retry: OptionalRetry = gapic_v1.method.DEFAULT,
timeout: float = None,
metadata: Sequence[Tuple[str, str]] = (),
Expand All @@ -214,6 +215,21 @@ async def create_endpoint(
This corresponds to the ``endpoint`` field
on the ``request`` instance; if ``request`` is provided, this
should not be set.
endpoint_id (:class:`str`):
Immutable. The ID to use for endpoint, which will become
the final component of the endpoint resource name. If
not provided, Vertex AI will generate a value for this
ID.
This value should be 1-10 characters, and valid
characters are /[0-9]/. When using HTTP/JSON, this field
is populated based on a query string argument, such as
``?endpoint_id=12345``. This is the fallback for fields
that are not included in either the URI or the body.
This corresponds to the ``endpoint_id`` field
on the ``request`` instance; if ``request`` is provided, this
should not be set.
retry (google.api_core.retry.Retry): Designation of what errors, if any,
should be retried.
timeout (float): The timeout for this request.
Expand All @@ -231,7 +247,7 @@ async def create_endpoint(
# Create or coerce a protobuf request object.
# Sanity check: If we got a request object, we should *not* have
# gotten any keyword arguments that map to the request.
has_flattened_params = any([parent, endpoint])
has_flattened_params = any([parent, endpoint, endpoint_id])
if request is not None and has_flattened_params:
raise ValueError(
"If the `request` argument is set, then none of "
Expand All @@ -246,6 +262,8 @@ async def create_endpoint(
request.parent = parent
if endpoint is not None:
request.endpoint = endpoint
if endpoint_id is not None:
request.endpoint_id = endpoint_id

# Wrap the RPC method; this adds retry and timeout information,
# and friendly error handling.
Expand Down
Expand Up @@ -424,6 +424,7 @@ def create_endpoint(
*,
parent: str = None,
endpoint: gca_endpoint.Endpoint = None,
endpoint_id: str = None,
retry: OptionalRetry = gapic_v1.method.DEFAULT,
timeout: float = None,
metadata: Sequence[Tuple[str, str]] = (),
Expand All @@ -447,6 +448,21 @@ def create_endpoint(
This corresponds to the ``endpoint`` field
on the ``request`` instance; if ``request`` is provided, this
should not be set.
endpoint_id (str):
Immutable. The ID to use for endpoint, which will become
the final component of the endpoint resource name. If
not provided, Vertex AI will generate a value for this
ID.
This value should be 1-10 characters, and valid
characters are /[0-9]/. When using HTTP/JSON, this field
is populated based on a query string argument, such as
``?endpoint_id=12345``. This is the fallback for fields
that are not included in either the URI or the body.
This corresponds to the ``endpoint_id`` field
on the ``request`` instance; if ``request`` is provided, this
should not be set.
retry (google.api_core.retry.Retry): Designation of what errors, if any,
should be retried.
timeout (float): The timeout for this request.
Expand All @@ -464,7 +480,7 @@ def create_endpoint(
# Create or coerce a protobuf request object.
# Sanity check: If we got a request object, we should *not* have
# gotten any keyword arguments that map to the request.
has_flattened_params = any([parent, endpoint])
has_flattened_params = any([parent, endpoint, endpoint_id])
if request is not None and has_flattened_params:
raise ValueError(
"If the `request` argument is set, then none of "
Expand All @@ -483,6 +499,8 @@ def create_endpoint(
request.parent = parent
if endpoint is not None:
request.endpoint = endpoint
if endpoint_id is not None:
request.endpoint_id = endpoint_id

# Wrap the RPC method; this adds retry and timeout information,
# and friendly error handling.
Expand Down

0 comments on commit ca813be

Please sign in to comment.