diff --git a/google/cloud/videointelligence_v1/gapic/enums.py b/google/cloud/videointelligence_v1/gapic/enums.py index 09c442d2..c8757091 100644 --- a/google/cloud/videointelligence_v1/gapic/enums.py +++ b/google/cloud/videointelligence_v1/gapic/enums.py @@ -32,6 +32,7 @@ class Feature(enum.IntEnum): SPEECH_TRANSCRIPTION (int): Speech transcription. TEXT_DETECTION (int): OCR text detection and tracking. OBJECT_TRACKING (int): Object detection and tracking. + LOGO_RECOGNITION (int): Logo detection, tracking, and recognition. """ FEATURE_UNSPECIFIED = 0 @@ -42,6 +43,7 @@ class Feature(enum.IntEnum): SPEECH_TRANSCRIPTION = 6 TEXT_DETECTION = 7 OBJECT_TRACKING = 9 + LOGO_RECOGNITION = 12 class LabelDetectionMode(enum.IntEnum): diff --git a/google/cloud/videointelligence_v1/gapic/video_intelligence_service_client.py b/google/cloud/videointelligence_v1/gapic/video_intelligence_service_client.py index 8b6cb9da..3b92add2 100644 --- a/google/cloud/videointelligence_v1/gapic/video_intelligence_service_client.py +++ b/google/cloud/videointelligence_v1/gapic/video_intelligence_service_client.py @@ -237,11 +237,12 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. A video - URI may include wildcards in ``object-id``, and thus identify multiple - videos. Supported wildcards: '\*' to match 0 or more characters; '?' to - match 1 character. If unset, the input video should be embedded in the - request as ``input_content``. If set, ``input_content`` should be unset. + URIs `__. A + video URI may include wildcards in ``object-id``, and thus identify + multiple videos. Supported wildcards: '\*' to match 0 or more + characters; '?' to match 1 character. If unset, the input video should + be embedded in the request as ``input_content``. If set, + ``input_content`` should be unset. input_content (bytes): The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be unset. features (list[~google.cloud.videointelligence_v1.types.Feature]): Required. Requested video annotation features. @@ -255,7 +256,7 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. + URIs `__. location_id (str): Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe-west1``, ``asia-east1``. If no region is specified, a region will be determined diff --git a/google/cloud/videointelligence_v1/proto/video_intelligence.proto b/google/cloud/videointelligence_v1/proto/video_intelligence.proto index 6611e9eb..9dcfda55 100644 --- a/google/cloud/videointelligence_v1/proto/video_intelligence.proto +++ b/google/cloud/videointelligence_v1/proto/video_intelligence.proto @@ -1,4 +1,4 @@ -// Copyright 2019 Google LLC. +// Copyright 2020 Google LLC // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -11,7 +11,6 @@ // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. -// syntax = "proto3"; @@ -62,7 +61,7 @@ message AnnotateVideoRequest { // supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see - // [Request URIs](/storage/docs/reference-uris). + // [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). // A video URI may include wildcards in `object-id`, and thus identify // multiple videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded @@ -85,7 +84,7 @@ message AnnotateVideoRequest { // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see - // [Request URIs](/storage/docs/reference-uris). + // [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). string output_uri = 4 [(google.api.field_behavior) = OPTIONAL]; // Optional. Cloud region where annotation should take place. Supported cloud @@ -148,6 +147,9 @@ enum Feature { // Object detection and tracking. OBJECT_TRACKING = 9; + + // Logo detection, tracking, and recognition. + LOGO_RECOGNITION = 12; } // Label detection mode. @@ -397,6 +399,67 @@ message FaceAnnotation { repeated FaceFrame frames = 3; } +// For tracking related features. +// An object at time_offset with attributes, and located with +// normalized_bounding_box. +message TimestampedObject { + // Normalized Bounding box in a frame, where the object is located. + NormalizedBoundingBox normalized_bounding_box = 1; + + // Time-offset, relative to the beginning of the video, + // corresponding to the video frame for this object. + google.protobuf.Duration time_offset = 2; + + // Optional. The attributes of the object in the bounding box. + repeated DetectedAttribute attributes = 3 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. The detected landmarks. + repeated DetectedLandmark landmarks = 4 [(google.api.field_behavior) = OPTIONAL]; +} + +// A track of an object instance. +message Track { + // Video segment of a track. + VideoSegment segment = 1; + + // The object with timestamp and attributes per frame in the track. + repeated TimestampedObject timestamped_objects = 2; + + // Optional. Attributes in the track level. + repeated DetectedAttribute attributes = 3 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. The confidence score of the tracked object. + float confidence = 4 [(google.api.field_behavior) = OPTIONAL]; +} + +// A generic detected attribute represented by name in string format. +message DetectedAttribute { + // The name of the attribute, i.e. glasses, dark_glasses, mouth_open etc. + // A full list of supported type names will be provided in the document. + string name = 1; + + // Detected attribute confidence. Range [0, 1]. + float confidence = 2; + + // Text value of the detection result. For example, the value for "HairColor" + // can be "black", "blonde", etc. + string value = 3; +} + +// A generic detected landmark represented by name in string format and a 2D +// location. +message DetectedLandmark { + // The name of this landmark, i.e. left_hand, right_shoulder. + string name = 1; + + // The 2D point of the detected landmark using the normalized image + // coordindate system. The normalized coordinates have the range from 0 to 1. + NormalizedVertex point = 2; + + // The confidence score of the detected landmark. Range [0, 1]. + float confidence = 3; +} + // Annotation results for a single video. message VideoAnnotationResults { // Video file location in @@ -453,6 +516,9 @@ message VideoAnnotationResults { // Annotations for list of objects detected and tracked in video. repeated ObjectTrackingAnnotation object_annotations = 14; + // Annotations for list of logos detected, tracked and recognized in video. + repeated LogoRecognitionAnnotation logo_recognition_annotations = 19; + // If set, indicates an error. Note that for a single `AnnotateVideoRequest` // some videos may succeed and some may fail. google.rpc.Status error = 9; @@ -743,3 +809,18 @@ message ObjectTrackingAnnotation { // Streaming mode: it can only be one ObjectTrackingFrame message in frames. repeated ObjectTrackingFrame frames = 2; } + +// Annotation corresponding to one detected, tracked and recognized logo class. +message LogoRecognitionAnnotation { + // Entity category information to specify the logo class that all the logo + // tracks within this LogoRecognitionAnnotation are recognized as. + Entity entity = 1; + + // All logo tracks where the recognized logo appears. Each track corresponds + // to one logo instance appearing in consecutive frames. + repeated Track tracks = 2; + + // All video segments where the recognized logo appears. There might be + // multiple instances of the same logo class appearing in one VideoSegment. + repeated VideoSegment segments = 3; +} diff --git a/google/cloud/videointelligence_v1/proto/video_intelligence_pb2.py b/google/cloud/videointelligence_v1/proto/video_intelligence_pb2.py index 8542ffb8..ceea9d8a 100644 --- a/google/cloud/videointelligence_v1/proto/video_intelligence_pb2.py +++ b/google/cloud/videointelligence_v1/proto/video_intelligence_pb2.py @@ -35,7 +35,7 @@ "\n%com.google.cloud.videointelligence.v1B\035VideoIntelligenceServiceProtoP\001ZRgoogle.golang.org/genproto/googleapis/cloud/videointelligence/v1;videointelligence\252\002!Google.Cloud.VideoIntelligence.V1\312\002!Google\\Cloud\\VideoIntelligence\\V1\352\002$Google::Cloud::VideoIntelligence::V1" ), serialized_pb=_b( - '\n@google/cloud/videointelligence_v1/proto/video_intelligence.proto\x12!google.cloud.videointelligence.v1\x1a\x1cgoogle/api/annotations.proto\x1a\x17google/api/client.proto\x1a\x1fgoogle/api/field_behavior.proto\x1a#google/longrunning/operations.proto\x1a\x1egoogle/protobuf/duration.proto\x1a\x1fgoogle/protobuf/timestamp.proto\x1a\x17google/rpc/status.proto"\xfe\x01\n\x14\x41nnotateVideoRequest\x12\x11\n\tinput_uri\x18\x01 \x01(\t\x12\x15\n\rinput_content\x18\x06 \x01(\x0c\x12\x41\n\x08\x66\x65\x61tures\x18\x02 \x03(\x0e\x32*.google.cloud.videointelligence.v1.FeatureB\x03\xe0\x41\x02\x12\x46\n\rvideo_context\x18\x03 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoContext\x12\x17\n\noutput_uri\x18\x04 \x01(\tB\x03\xe0\x41\x01\x12\x18\n\x0blocation_id\x18\x05 \x01(\tB\x03\xe0\x41\x01"\xe6\x05\n\x0cVideoContext\x12\x41\n\x08segments\x18\x01 \x03(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12W\n\x16label_detection_config\x18\x02 \x01(\x0b\x32\x37.google.cloud.videointelligence.v1.LabelDetectionConfig\x12\x62\n\x1cshot_change_detection_config\x18\x03 \x01(\x0b\x32<.google.cloud.videointelligence.v1.ShotChangeDetectionConfig\x12l\n!explicit_content_detection_config\x18\x04 \x01(\x0b\x32\x41.google.cloud.videointelligence.v1.ExplicitContentDetectionConfig\x12U\n\x15\x66\x61\x63\x65_detection_config\x18\x05 \x01(\x0b\x32\x36.google.cloud.videointelligence.v1.FaceDetectionConfig\x12\x61\n\x1bspeech_transcription_config\x18\x06 \x01(\x0b\x32<.google.cloud.videointelligence.v1.SpeechTranscriptionConfig\x12U\n\x15text_detection_config\x18\x08 \x01(\x0b\x32\x36.google.cloud.videointelligence.v1.TextDetectionConfig\x12W\n\x16object_tracking_config\x18\r \x01(\x0b\x32\x37.google.cloud.videointelligence.v1.ObjectTrackingConfig"\xdd\x01\n\x14LabelDetectionConfig\x12S\n\x14label_detection_mode\x18\x01 \x01(\x0e\x32\x35.google.cloud.videointelligence.v1.LabelDetectionMode\x12\x19\n\x11stationary_camera\x18\x02 \x01(\x08\x12\r\n\x05model\x18\x03 \x01(\t\x12"\n\x1a\x66rame_confidence_threshold\x18\x04 \x01(\x02\x12"\n\x1avideo_confidence_threshold\x18\x05 \x01(\x02"*\n\x19ShotChangeDetectionConfig\x12\r\n\x05model\x18\x01 \x01(\t"%\n\x14ObjectTrackingConfig\x12\r\n\x05model\x18\x01 \x01(\t"D\n\x13\x46\x61\x63\x65\x44\x65tectionConfig\x12\r\n\x05model\x18\x01 \x01(\t\x12\x1e\n\x16include_bounding_boxes\x18\x02 \x01(\x08"/\n\x1e\x45xplicitContentDetectionConfig\x12\r\n\x05model\x18\x01 \x01(\t"<\n\x13TextDetectionConfig\x12\x16\n\x0elanguage_hints\x18\x01 \x03(\t\x12\r\n\x05model\x18\x02 \x01(\t"x\n\x0cVideoSegment\x12\x34\n\x11start_time_offset\x18\x01 \x01(\x0b\x32\x19.google.protobuf.Duration\x12\x32\n\x0f\x65nd_time_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration"d\n\x0cLabelSegment\x12@\n\x07segment\x18\x01 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12\x12\n\nconfidence\x18\x02 \x01(\x02"P\n\nLabelFrame\x12.\n\x0btime_offset\x18\x01 \x01(\x0b\x32\x19.google.protobuf.Duration\x12\x12\n\nconfidence\x18\x02 \x01(\x02"G\n\x06\x45ntity\x12\x11\n\tentity_id\x18\x01 \x01(\t\x12\x13\n\x0b\x64\x65scription\x18\x02 \x01(\t\x12\x15\n\rlanguage_code\x18\x03 \x01(\t"\x94\x02\n\x0fLabelAnnotation\x12\x39\n\x06\x65ntity\x18\x01 \x01(\x0b\x32).google.cloud.videointelligence.v1.Entity\x12\x44\n\x11\x63\x61tegory_entities\x18\x02 \x03(\x0b\x32).google.cloud.videointelligence.v1.Entity\x12\x41\n\x08segments\x18\x03 \x03(\x0b\x32/.google.cloud.videointelligence.v1.LabelSegment\x12=\n\x06\x66rames\x18\x04 \x03(\x0b\x32-.google.cloud.videointelligence.v1.LabelFrame"\x95\x01\n\x14\x45xplicitContentFrame\x12.\n\x0btime_offset\x18\x01 \x01(\x0b\x32\x19.google.protobuf.Duration\x12M\n\x16pornography_likelihood\x18\x02 \x01(\x0e\x32-.google.cloud.videointelligence.v1.Likelihood"d\n\x19\x45xplicitContentAnnotation\x12G\n\x06\x66rames\x18\x01 \x03(\x0b\x32\x37.google.cloud.videointelligence.v1.ExplicitContentFrame"Q\n\x15NormalizedBoundingBox\x12\x0c\n\x04left\x18\x01 \x01(\x02\x12\x0b\n\x03top\x18\x02 \x01(\x02\x12\r\n\x05right\x18\x03 \x01(\x02\x12\x0e\n\x06\x62ottom\x18\x04 \x01(\x02"O\n\x0b\x46\x61\x63\x65Segment\x12@\n\x07segment\x18\x01 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment"\x98\x01\n\tFaceFrame\x12[\n\x19normalized_bounding_boxes\x18\x01 \x03(\x0b\x32\x38.google.cloud.videointelligence.v1.NormalizedBoundingBox\x12.\n\x0btime_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration"\xa3\x01\n\x0e\x46\x61\x63\x65\x41nnotation\x12\x11\n\tthumbnail\x18\x01 \x01(\x0c\x12@\n\x08segments\x18\x02 \x03(\x0b\x32..google.cloud.videointelligence.v1.FaceSegment\x12<\n\x06\x66rames\x18\x03 \x03(\x0b\x32,.google.cloud.videointelligence.v1.FaceFrame"\xbd\x08\n\x16VideoAnnotationResults\x12\x11\n\tinput_uri\x18\x01 \x01(\t\x12@\n\x07segment\x18\n \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12U\n\x19segment_label_annotations\x18\x02 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12^\n"segment_presence_label_annotations\x18\x17 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12R\n\x16shot_label_annotations\x18\x03 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12[\n\x1fshot_presence_label_annotations\x18\x18 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12S\n\x17\x66rame_label_annotations\x18\x04 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12K\n\x10\x66\x61\x63\x65_annotations\x18\x05 \x03(\x0b\x32\x31.google.cloud.videointelligence.v1.FaceAnnotation\x12I\n\x10shot_annotations\x18\x06 \x03(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12Y\n\x13\x65xplicit_annotation\x18\x07 \x01(\x0b\x32<.google.cloud.videointelligence.v1.ExplicitContentAnnotation\x12U\n\x15speech_transcriptions\x18\x0b \x03(\x0b\x32\x36.google.cloud.videointelligence.v1.SpeechTranscription\x12K\n\x10text_annotations\x18\x0c \x03(\x0b\x32\x31.google.cloud.videointelligence.v1.TextAnnotation\x12W\n\x12object_annotations\x18\x0e \x03(\x0b\x32;.google.cloud.videointelligence.v1.ObjectTrackingAnnotation\x12!\n\x05\x65rror\x18\t \x01(\x0b\x32\x12.google.rpc.Status"n\n\x15\x41nnotateVideoResponse\x12U\n\x12\x61nnotation_results\x18\x01 \x03(\x0b\x32\x39.google.cloud.videointelligence.v1.VideoAnnotationResults"\xa6\x02\n\x17VideoAnnotationProgress\x12\x11\n\tinput_uri\x18\x01 \x01(\t\x12\x18\n\x10progress_percent\x18\x02 \x01(\x05\x12.\n\nstart_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12/\n\x0bupdate_time\x18\x04 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12;\n\x07\x66\x65\x61ture\x18\x05 \x01(\x0e\x32*.google.cloud.videointelligence.v1.Feature\x12@\n\x07segment\x18\x06 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment"p\n\x15\x41nnotateVideoProgress\x12W\n\x13\x61nnotation_progress\x18\x01 \x03(\x0b\x32:.google.cloud.videointelligence.v1.VideoAnnotationProgress"\x81\x03\n\x19SpeechTranscriptionConfig\x12\x1a\n\rlanguage_code\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12\x1d\n\x10max_alternatives\x18\x02 \x01(\x05\x42\x03\xe0\x41\x01\x12\x1d\n\x10\x66ilter_profanity\x18\x03 \x01(\x08\x42\x03\xe0\x41\x01\x12N\n\x0fspeech_contexts\x18\x04 \x03(\x0b\x32\x30.google.cloud.videointelligence.v1.SpeechContextB\x03\xe0\x41\x01\x12)\n\x1c\x65nable_automatic_punctuation\x18\x05 \x01(\x08\x42\x03\xe0\x41\x01\x12\x19\n\x0c\x61udio_tracks\x18\x06 \x03(\x05\x42\x03\xe0\x41\x01\x12\'\n\x1a\x65nable_speaker_diarization\x18\x07 \x01(\x08\x42\x03\xe0\x41\x01\x12&\n\x19\x64iarization_speaker_count\x18\x08 \x01(\x05\x42\x03\xe0\x41\x01\x12#\n\x16\x65nable_word_confidence\x18\t \x01(\x08\x42\x03\xe0\x41\x01"%\n\rSpeechContext\x12\x14\n\x07phrases\x18\x01 \x03(\tB\x03\xe0\x41\x01"\x88\x01\n\x13SpeechTranscription\x12U\n\x0c\x61lternatives\x18\x01 \x03(\x0b\x32?.google.cloud.videointelligence.v1.SpeechRecognitionAlternative\x12\x1a\n\rlanguage_code\x18\x02 \x01(\tB\x03\xe0\x41\x03"\x8c\x01\n\x1cSpeechRecognitionAlternative\x12\x12\n\ntranscript\x18\x01 \x01(\t\x12\x17\n\nconfidence\x18\x02 \x01(\x02\x42\x03\xe0\x41\x03\x12?\n\x05words\x18\x03 \x03(\x0b\x32+.google.cloud.videointelligence.v1.WordInfoB\x03\xe0\x41\x03"\xa7\x01\n\x08WordInfo\x12-\n\nstart_time\x18\x01 \x01(\x0b\x32\x19.google.protobuf.Duration\x12+\n\x08\x65nd_time\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration\x12\x0c\n\x04word\x18\x03 \x01(\t\x12\x17\n\nconfidence\x18\x04 \x01(\x02\x42\x03\xe0\x41\x03\x12\x18\n\x0bspeaker_tag\x18\x05 \x01(\x05\x42\x03\xe0\x41\x03"(\n\x10NormalizedVertex\x12\t\n\x01x\x18\x01 \x01(\x02\x12\t\n\x01y\x18\x02 \x01(\x02"_\n\x16NormalizedBoundingPoly\x12\x45\n\x08vertices\x18\x01 \x03(\x0b\x32\x33.google.cloud.videointelligence.v1.NormalizedVertex"\xa1\x01\n\x0bTextSegment\x12@\n\x07segment\x18\x01 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12\x12\n\nconfidence\x18\x02 \x01(\x02\x12<\n\x06\x66rames\x18\x03 \x03(\x0b\x32,.google.cloud.videointelligence.v1.TextFrame"\x94\x01\n\tTextFrame\x12W\n\x14rotated_bounding_box\x18\x01 \x01(\x0b\x32\x39.google.cloud.videointelligence.v1.NormalizedBoundingPoly\x12.\n\x0btime_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration"`\n\x0eTextAnnotation\x12\x0c\n\x04text\x18\x01 \x01(\t\x12@\n\x08segments\x18\x02 \x03(\x0b\x32..google.cloud.videointelligence.v1.TextSegment"\xa0\x01\n\x13ObjectTrackingFrame\x12Y\n\x17normalized_bounding_box\x18\x01 \x01(\x0b\x32\x38.google.cloud.videointelligence.v1.NormalizedBoundingBox\x12.\n\x0btime_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration"\x97\x02\n\x18ObjectTrackingAnnotation\x12\x42\n\x07segment\x18\x03 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegmentH\x00\x12\x12\n\x08track_id\x18\x05 \x01(\x03H\x00\x12\x39\n\x06\x65ntity\x18\x01 \x01(\x0b\x32).google.cloud.videointelligence.v1.Entity\x12\x12\n\nconfidence\x18\x04 \x01(\x02\x12\x46\n\x06\x66rames\x18\x02 \x03(\x0b\x32\x36.google.cloud.videointelligence.v1.ObjectTrackingFrameB\x0c\n\ntrack_info*\xc9\x01\n\x07\x46\x65\x61ture\x12\x17\n\x13\x46\x45\x41TURE_UNSPECIFIED\x10\x00\x12\x13\n\x0fLABEL_DETECTION\x10\x01\x12\x19\n\x15SHOT_CHANGE_DETECTION\x10\x02\x12\x1e\n\x1a\x45XPLICIT_CONTENT_DETECTION\x10\x03\x12\x12\n\x0e\x46\x41\x43\x45_DETECTION\x10\x04\x12\x18\n\x14SPEECH_TRANSCRIPTION\x10\x06\x12\x12\n\x0eTEXT_DETECTION\x10\x07\x12\x13\n\x0fOBJECT_TRACKING\x10\t*r\n\x12LabelDetectionMode\x12$\n LABEL_DETECTION_MODE_UNSPECIFIED\x10\x00\x12\r\n\tSHOT_MODE\x10\x01\x12\x0e\n\nFRAME_MODE\x10\x02\x12\x17\n\x13SHOT_AND_FRAME_MODE\x10\x03*t\n\nLikelihood\x12\x1a\n\x16LIKELIHOOD_UNSPECIFIED\x10\x00\x12\x11\n\rVERY_UNLIKELY\x10\x01\x12\x0c\n\x08UNLIKELY\x10\x02\x12\x0c\n\x08POSSIBLE\x10\x03\x12\n\n\x06LIKELY\x10\x04\x12\x0f\n\x0bVERY_LIKELY\x10\x05\x32\xc0\x02\n\x18VideoIntelligenceService\x12\xcd\x01\n\rAnnotateVideo\x12\x37.google.cloud.videointelligence.v1.AnnotateVideoRequest\x1a\x1d.google.longrunning.Operation"d\x82\xd3\xe4\x93\x02\x18"\x13/v1/videos:annotate:\x01*\xda\x41\x12input_uri,features\xca\x41.\n\x15\x41nnotateVideoResponse\x12\x15\x41nnotateVideoProgress\x1aT\xca\x41 videointelligence.googleapis.com\xd2\x41.https://www.googleapis.com/auth/cloud-platformB\x8b\x02\n%com.google.cloud.videointelligence.v1B\x1dVideoIntelligenceServiceProtoP\x01ZRgoogle.golang.org/genproto/googleapis/cloud/videointelligence/v1;videointelligence\xaa\x02!Google.Cloud.VideoIntelligence.V1\xca\x02!Google\\Cloud\\VideoIntelligence\\V1\xea\x02$Google::Cloud::VideoIntelligence::V1b\x06proto3' + '\n@google/cloud/videointelligence_v1/proto/video_intelligence.proto\x12!google.cloud.videointelligence.v1\x1a\x1cgoogle/api/annotations.proto\x1a\x17google/api/client.proto\x1a\x1fgoogle/api/field_behavior.proto\x1a#google/longrunning/operations.proto\x1a\x1egoogle/protobuf/duration.proto\x1a\x1fgoogle/protobuf/timestamp.proto\x1a\x17google/rpc/status.proto"\xfe\x01\n\x14\x41nnotateVideoRequest\x12\x11\n\tinput_uri\x18\x01 \x01(\t\x12\x15\n\rinput_content\x18\x06 \x01(\x0c\x12\x41\n\x08\x66\x65\x61tures\x18\x02 \x03(\x0e\x32*.google.cloud.videointelligence.v1.FeatureB\x03\xe0\x41\x02\x12\x46\n\rvideo_context\x18\x03 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoContext\x12\x17\n\noutput_uri\x18\x04 \x01(\tB\x03\xe0\x41\x01\x12\x18\n\x0blocation_id\x18\x05 \x01(\tB\x03\xe0\x41\x01"\xe6\x05\n\x0cVideoContext\x12\x41\n\x08segments\x18\x01 \x03(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12W\n\x16label_detection_config\x18\x02 \x01(\x0b\x32\x37.google.cloud.videointelligence.v1.LabelDetectionConfig\x12\x62\n\x1cshot_change_detection_config\x18\x03 \x01(\x0b\x32<.google.cloud.videointelligence.v1.ShotChangeDetectionConfig\x12l\n!explicit_content_detection_config\x18\x04 \x01(\x0b\x32\x41.google.cloud.videointelligence.v1.ExplicitContentDetectionConfig\x12U\n\x15\x66\x61\x63\x65_detection_config\x18\x05 \x01(\x0b\x32\x36.google.cloud.videointelligence.v1.FaceDetectionConfig\x12\x61\n\x1bspeech_transcription_config\x18\x06 \x01(\x0b\x32<.google.cloud.videointelligence.v1.SpeechTranscriptionConfig\x12U\n\x15text_detection_config\x18\x08 \x01(\x0b\x32\x36.google.cloud.videointelligence.v1.TextDetectionConfig\x12W\n\x16object_tracking_config\x18\r \x01(\x0b\x32\x37.google.cloud.videointelligence.v1.ObjectTrackingConfig"\xdd\x01\n\x14LabelDetectionConfig\x12S\n\x14label_detection_mode\x18\x01 \x01(\x0e\x32\x35.google.cloud.videointelligence.v1.LabelDetectionMode\x12\x19\n\x11stationary_camera\x18\x02 \x01(\x08\x12\r\n\x05model\x18\x03 \x01(\t\x12"\n\x1a\x66rame_confidence_threshold\x18\x04 \x01(\x02\x12"\n\x1avideo_confidence_threshold\x18\x05 \x01(\x02"*\n\x19ShotChangeDetectionConfig\x12\r\n\x05model\x18\x01 \x01(\t"%\n\x14ObjectTrackingConfig\x12\r\n\x05model\x18\x01 \x01(\t"D\n\x13\x46\x61\x63\x65\x44\x65tectionConfig\x12\r\n\x05model\x18\x01 \x01(\t\x12\x1e\n\x16include_bounding_boxes\x18\x02 \x01(\x08"/\n\x1e\x45xplicitContentDetectionConfig\x12\r\n\x05model\x18\x01 \x01(\t"<\n\x13TextDetectionConfig\x12\x16\n\x0elanguage_hints\x18\x01 \x03(\t\x12\r\n\x05model\x18\x02 \x01(\t"x\n\x0cVideoSegment\x12\x34\n\x11start_time_offset\x18\x01 \x01(\x0b\x32\x19.google.protobuf.Duration\x12\x32\n\x0f\x65nd_time_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration"d\n\x0cLabelSegment\x12@\n\x07segment\x18\x01 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12\x12\n\nconfidence\x18\x02 \x01(\x02"P\n\nLabelFrame\x12.\n\x0btime_offset\x18\x01 \x01(\x0b\x32\x19.google.protobuf.Duration\x12\x12\n\nconfidence\x18\x02 \x01(\x02"G\n\x06\x45ntity\x12\x11\n\tentity_id\x18\x01 \x01(\t\x12\x13\n\x0b\x64\x65scription\x18\x02 \x01(\t\x12\x15\n\rlanguage_code\x18\x03 \x01(\t"\x94\x02\n\x0fLabelAnnotation\x12\x39\n\x06\x65ntity\x18\x01 \x01(\x0b\x32).google.cloud.videointelligence.v1.Entity\x12\x44\n\x11\x63\x61tegory_entities\x18\x02 \x03(\x0b\x32).google.cloud.videointelligence.v1.Entity\x12\x41\n\x08segments\x18\x03 \x03(\x0b\x32/.google.cloud.videointelligence.v1.LabelSegment\x12=\n\x06\x66rames\x18\x04 \x03(\x0b\x32-.google.cloud.videointelligence.v1.LabelFrame"\x95\x01\n\x14\x45xplicitContentFrame\x12.\n\x0btime_offset\x18\x01 \x01(\x0b\x32\x19.google.protobuf.Duration\x12M\n\x16pornography_likelihood\x18\x02 \x01(\x0e\x32-.google.cloud.videointelligence.v1.Likelihood"d\n\x19\x45xplicitContentAnnotation\x12G\n\x06\x66rames\x18\x01 \x03(\x0b\x32\x37.google.cloud.videointelligence.v1.ExplicitContentFrame"Q\n\x15NormalizedBoundingBox\x12\x0c\n\x04left\x18\x01 \x01(\x02\x12\x0b\n\x03top\x18\x02 \x01(\x02\x12\r\n\x05right\x18\x03 \x01(\x02\x12\x0e\n\x06\x62ottom\x18\x04 \x01(\x02"O\n\x0b\x46\x61\x63\x65Segment\x12@\n\x07segment\x18\x01 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment"\x98\x01\n\tFaceFrame\x12[\n\x19normalized_bounding_boxes\x18\x01 \x03(\x0b\x32\x38.google.cloud.videointelligence.v1.NormalizedBoundingBox\x12.\n\x0btime_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration"\xa3\x01\n\x0e\x46\x61\x63\x65\x41nnotation\x12\x11\n\tthumbnail\x18\x01 \x01(\x0c\x12@\n\x08segments\x18\x02 \x03(\x0b\x32..google.cloud.videointelligence.v1.FaceSegment\x12<\n\x06\x66rames\x18\x03 \x03(\x0b\x32,.google.cloud.videointelligence.v1.FaceFrame"\xba\x02\n\x11TimestampedObject\x12Y\n\x17normalized_bounding_box\x18\x01 \x01(\x0b\x32\x38.google.cloud.videointelligence.v1.NormalizedBoundingBox\x12.\n\x0btime_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration\x12M\n\nattributes\x18\x03 \x03(\x0b\x32\x34.google.cloud.videointelligence.v1.DetectedAttributeB\x03\xe0\x41\x01\x12K\n\tlandmarks\x18\x04 \x03(\x0b\x32\x33.google.cloud.videointelligence.v1.DetectedLandmarkB\x03\xe0\x41\x01"\x84\x02\n\x05Track\x12@\n\x07segment\x18\x01 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12Q\n\x13timestamped_objects\x18\x02 \x03(\x0b\x32\x34.google.cloud.videointelligence.v1.TimestampedObject\x12M\n\nattributes\x18\x03 \x03(\x0b\x32\x34.google.cloud.videointelligence.v1.DetectedAttributeB\x03\xe0\x41\x01\x12\x17\n\nconfidence\x18\x04 \x01(\x02\x42\x03\xe0\x41\x01"D\n\x11\x44\x65tectedAttribute\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x12\n\nconfidence\x18\x02 \x01(\x02\x12\r\n\x05value\x18\x03 \x01(\t"x\n\x10\x44\x65tectedLandmark\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x42\n\x05point\x18\x02 \x01(\x0b\x32\x33.google.cloud.videointelligence.v1.NormalizedVertex\x12\x12\n\nconfidence\x18\x03 \x01(\x02"\xa1\t\n\x16VideoAnnotationResults\x12\x11\n\tinput_uri\x18\x01 \x01(\t\x12@\n\x07segment\x18\n \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12U\n\x19segment_label_annotations\x18\x02 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12^\n"segment_presence_label_annotations\x18\x17 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12R\n\x16shot_label_annotations\x18\x03 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12[\n\x1fshot_presence_label_annotations\x18\x18 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12S\n\x17\x66rame_label_annotations\x18\x04 \x03(\x0b\x32\x32.google.cloud.videointelligence.v1.LabelAnnotation\x12K\n\x10\x66\x61\x63\x65_annotations\x18\x05 \x03(\x0b\x32\x31.google.cloud.videointelligence.v1.FaceAnnotation\x12I\n\x10shot_annotations\x18\x06 \x03(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12Y\n\x13\x65xplicit_annotation\x18\x07 \x01(\x0b\x32<.google.cloud.videointelligence.v1.ExplicitContentAnnotation\x12U\n\x15speech_transcriptions\x18\x0b \x03(\x0b\x32\x36.google.cloud.videointelligence.v1.SpeechTranscription\x12K\n\x10text_annotations\x18\x0c \x03(\x0b\x32\x31.google.cloud.videointelligence.v1.TextAnnotation\x12W\n\x12object_annotations\x18\x0e \x03(\x0b\x32;.google.cloud.videointelligence.v1.ObjectTrackingAnnotation\x12\x62\n\x1clogo_recognition_annotations\x18\x13 \x03(\x0b\x32<.google.cloud.videointelligence.v1.LogoRecognitionAnnotation\x12!\n\x05\x65rror\x18\t \x01(\x0b\x32\x12.google.rpc.Status"n\n\x15\x41nnotateVideoResponse\x12U\n\x12\x61nnotation_results\x18\x01 \x03(\x0b\x32\x39.google.cloud.videointelligence.v1.VideoAnnotationResults"\xa6\x02\n\x17VideoAnnotationProgress\x12\x11\n\tinput_uri\x18\x01 \x01(\t\x12\x18\n\x10progress_percent\x18\x02 \x01(\x05\x12.\n\nstart_time\x18\x03 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12/\n\x0bupdate_time\x18\x04 \x01(\x0b\x32\x1a.google.protobuf.Timestamp\x12;\n\x07\x66\x65\x61ture\x18\x05 \x01(\x0e\x32*.google.cloud.videointelligence.v1.Feature\x12@\n\x07segment\x18\x06 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment"p\n\x15\x41nnotateVideoProgress\x12W\n\x13\x61nnotation_progress\x18\x01 \x03(\x0b\x32:.google.cloud.videointelligence.v1.VideoAnnotationProgress"\x81\x03\n\x19SpeechTranscriptionConfig\x12\x1a\n\rlanguage_code\x18\x01 \x01(\tB\x03\xe0\x41\x02\x12\x1d\n\x10max_alternatives\x18\x02 \x01(\x05\x42\x03\xe0\x41\x01\x12\x1d\n\x10\x66ilter_profanity\x18\x03 \x01(\x08\x42\x03\xe0\x41\x01\x12N\n\x0fspeech_contexts\x18\x04 \x03(\x0b\x32\x30.google.cloud.videointelligence.v1.SpeechContextB\x03\xe0\x41\x01\x12)\n\x1c\x65nable_automatic_punctuation\x18\x05 \x01(\x08\x42\x03\xe0\x41\x01\x12\x19\n\x0c\x61udio_tracks\x18\x06 \x03(\x05\x42\x03\xe0\x41\x01\x12\'\n\x1a\x65nable_speaker_diarization\x18\x07 \x01(\x08\x42\x03\xe0\x41\x01\x12&\n\x19\x64iarization_speaker_count\x18\x08 \x01(\x05\x42\x03\xe0\x41\x01\x12#\n\x16\x65nable_word_confidence\x18\t \x01(\x08\x42\x03\xe0\x41\x01"%\n\rSpeechContext\x12\x14\n\x07phrases\x18\x01 \x03(\tB\x03\xe0\x41\x01"\x88\x01\n\x13SpeechTranscription\x12U\n\x0c\x61lternatives\x18\x01 \x03(\x0b\x32?.google.cloud.videointelligence.v1.SpeechRecognitionAlternative\x12\x1a\n\rlanguage_code\x18\x02 \x01(\tB\x03\xe0\x41\x03"\x8c\x01\n\x1cSpeechRecognitionAlternative\x12\x12\n\ntranscript\x18\x01 \x01(\t\x12\x17\n\nconfidence\x18\x02 \x01(\x02\x42\x03\xe0\x41\x03\x12?\n\x05words\x18\x03 \x03(\x0b\x32+.google.cloud.videointelligence.v1.WordInfoB\x03\xe0\x41\x03"\xa7\x01\n\x08WordInfo\x12-\n\nstart_time\x18\x01 \x01(\x0b\x32\x19.google.protobuf.Duration\x12+\n\x08\x65nd_time\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration\x12\x0c\n\x04word\x18\x03 \x01(\t\x12\x17\n\nconfidence\x18\x04 \x01(\x02\x42\x03\xe0\x41\x03\x12\x18\n\x0bspeaker_tag\x18\x05 \x01(\x05\x42\x03\xe0\x41\x03"(\n\x10NormalizedVertex\x12\t\n\x01x\x18\x01 \x01(\x02\x12\t\n\x01y\x18\x02 \x01(\x02"_\n\x16NormalizedBoundingPoly\x12\x45\n\x08vertices\x18\x01 \x03(\x0b\x32\x33.google.cloud.videointelligence.v1.NormalizedVertex"\xa1\x01\n\x0bTextSegment\x12@\n\x07segment\x18\x01 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment\x12\x12\n\nconfidence\x18\x02 \x01(\x02\x12<\n\x06\x66rames\x18\x03 \x03(\x0b\x32,.google.cloud.videointelligence.v1.TextFrame"\x94\x01\n\tTextFrame\x12W\n\x14rotated_bounding_box\x18\x01 \x01(\x0b\x32\x39.google.cloud.videointelligence.v1.NormalizedBoundingPoly\x12.\n\x0btime_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration"`\n\x0eTextAnnotation\x12\x0c\n\x04text\x18\x01 \x01(\t\x12@\n\x08segments\x18\x02 \x03(\x0b\x32..google.cloud.videointelligence.v1.TextSegment"\xa0\x01\n\x13ObjectTrackingFrame\x12Y\n\x17normalized_bounding_box\x18\x01 \x01(\x0b\x32\x38.google.cloud.videointelligence.v1.NormalizedBoundingBox\x12.\n\x0btime_offset\x18\x02 \x01(\x0b\x32\x19.google.protobuf.Duration"\x97\x02\n\x18ObjectTrackingAnnotation\x12\x42\n\x07segment\x18\x03 \x01(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegmentH\x00\x12\x12\n\x08track_id\x18\x05 \x01(\x03H\x00\x12\x39\n\x06\x65ntity\x18\x01 \x01(\x0b\x32).google.cloud.videointelligence.v1.Entity\x12\x12\n\nconfidence\x18\x04 \x01(\x02\x12\x46\n\x06\x66rames\x18\x02 \x03(\x0b\x32\x36.google.cloud.videointelligence.v1.ObjectTrackingFrameB\x0c\n\ntrack_info"\xd3\x01\n\x19LogoRecognitionAnnotation\x12\x39\n\x06\x65ntity\x18\x01 \x01(\x0b\x32).google.cloud.videointelligence.v1.Entity\x12\x38\n\x06tracks\x18\x02 \x03(\x0b\x32(.google.cloud.videointelligence.v1.Track\x12\x41\n\x08segments\x18\x03 \x03(\x0b\x32/.google.cloud.videointelligence.v1.VideoSegment*\xdf\x01\n\x07\x46\x65\x61ture\x12\x17\n\x13\x46\x45\x41TURE_UNSPECIFIED\x10\x00\x12\x13\n\x0fLABEL_DETECTION\x10\x01\x12\x19\n\x15SHOT_CHANGE_DETECTION\x10\x02\x12\x1e\n\x1a\x45XPLICIT_CONTENT_DETECTION\x10\x03\x12\x12\n\x0e\x46\x41\x43\x45_DETECTION\x10\x04\x12\x18\n\x14SPEECH_TRANSCRIPTION\x10\x06\x12\x12\n\x0eTEXT_DETECTION\x10\x07\x12\x13\n\x0fOBJECT_TRACKING\x10\t\x12\x14\n\x10LOGO_RECOGNITION\x10\x0c*r\n\x12LabelDetectionMode\x12$\n LABEL_DETECTION_MODE_UNSPECIFIED\x10\x00\x12\r\n\tSHOT_MODE\x10\x01\x12\x0e\n\nFRAME_MODE\x10\x02\x12\x17\n\x13SHOT_AND_FRAME_MODE\x10\x03*t\n\nLikelihood\x12\x1a\n\x16LIKELIHOOD_UNSPECIFIED\x10\x00\x12\x11\n\rVERY_UNLIKELY\x10\x01\x12\x0c\n\x08UNLIKELY\x10\x02\x12\x0c\n\x08POSSIBLE\x10\x03\x12\n\n\x06LIKELY\x10\x04\x12\x0f\n\x0bVERY_LIKELY\x10\x05\x32\xc0\x02\n\x18VideoIntelligenceService\x12\xcd\x01\n\rAnnotateVideo\x12\x37.google.cloud.videointelligence.v1.AnnotateVideoRequest\x1a\x1d.google.longrunning.Operation"d\x82\xd3\xe4\x93\x02\x18"\x13/v1/videos:annotate:\x01*\xda\x41\x12input_uri,features\xca\x41.\n\x15\x41nnotateVideoResponse\x12\x15\x41nnotateVideoProgress\x1aT\xca\x41 videointelligence.googleapis.com\xd2\x41.https://www.googleapis.com/auth/cloud-platformB\x8b\x02\n%com.google.cloud.videointelligence.v1B\x1dVideoIntelligenceServiceProtoP\x01ZRgoogle.golang.org/genproto/googleapis/cloud/videointelligence/v1;videointelligence\xaa\x02!Google.Cloud.VideoIntelligence.V1\xca\x02!Google\\Cloud\\VideoIntelligence\\V1\xea\x02$Google::Cloud::VideoIntelligence::V1b\x06proto3' ), dependencies=[ google_dot_api_dot_annotations__pb2.DESCRIPTOR, @@ -102,11 +102,18 @@ serialized_options=None, type=None, ), + _descriptor.EnumValueDescriptor( + name="LOGO_RECOGNITION", + index=8, + number=12, + serialized_options=None, + type=None, + ), ], containing_type=None, serialized_options=None, - serialized_start=6693, - serialized_end=6894, + serialized_start=7779, + serialized_end=8002, ) _sym_db.RegisterEnumDescriptor(_FEATURE) @@ -140,8 +147,8 @@ ], containing_type=None, serialized_options=None, - serialized_start=6896, - serialized_end=7010, + serialized_start=8004, + serialized_end=8118, ) _sym_db.RegisterEnumDescriptor(_LABELDETECTIONMODE) @@ -177,8 +184,8 @@ ], containing_type=None, serialized_options=None, - serialized_start=7012, - serialized_end=7128, + serialized_start=8120, + serialized_end=8236, ) _sym_db.RegisterEnumDescriptor(_LIKELIHOOD) @@ -191,6 +198,7 @@ SPEECH_TRANSCRIPTION = 6 TEXT_DETECTION = 7 OBJECT_TRACKING = 9 +LOGO_RECOGNITION = 12 LABEL_DETECTION_MODE_UNSPECIFIED = 0 SHOT_MODE = 1 FRAME_MODE = 2 @@ -1538,6 +1546,342 @@ ) +_TIMESTAMPEDOBJECT = _descriptor.Descriptor( + name="TimestampedObject", + full_name="google.cloud.videointelligence.v1.TimestampedObject", + filename=None, + file=DESCRIPTOR, + containing_type=None, + fields=[ + _descriptor.FieldDescriptor( + name="normalized_bounding_box", + full_name="google.cloud.videointelligence.v1.TimestampedObject.normalized_bounding_box", + index=0, + number=1, + type=11, + cpp_type=10, + label=1, + has_default_value=False, + default_value=None, + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="time_offset", + full_name="google.cloud.videointelligence.v1.TimestampedObject.time_offset", + index=1, + number=2, + type=11, + cpp_type=10, + label=1, + has_default_value=False, + default_value=None, + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="attributes", + full_name="google.cloud.videointelligence.v1.TimestampedObject.attributes", + index=2, + number=3, + type=11, + cpp_type=10, + label=3, + has_default_value=False, + default_value=[], + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=_b("\340A\001"), + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="landmarks", + full_name="google.cloud.videointelligence.v1.TimestampedObject.landmarks", + index=3, + number=4, + type=11, + cpp_type=10, + label=3, + has_default_value=False, + default_value=[], + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=_b("\340A\001"), + file=DESCRIPTOR, + ), + ], + extensions=[], + nested_types=[], + enum_types=[], + serialized_options=None, + is_extendable=False, + syntax="proto3", + extension_ranges=[], + oneofs=[], + serialized_start=3206, + serialized_end=3520, +) + + +_TRACK = _descriptor.Descriptor( + name="Track", + full_name="google.cloud.videointelligence.v1.Track", + filename=None, + file=DESCRIPTOR, + containing_type=None, + fields=[ + _descriptor.FieldDescriptor( + name="segment", + full_name="google.cloud.videointelligence.v1.Track.segment", + index=0, + number=1, + type=11, + cpp_type=10, + label=1, + has_default_value=False, + default_value=None, + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="timestamped_objects", + full_name="google.cloud.videointelligence.v1.Track.timestamped_objects", + index=1, + number=2, + type=11, + cpp_type=10, + label=3, + has_default_value=False, + default_value=[], + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="attributes", + full_name="google.cloud.videointelligence.v1.Track.attributes", + index=2, + number=3, + type=11, + cpp_type=10, + label=3, + has_default_value=False, + default_value=[], + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=_b("\340A\001"), + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="confidence", + full_name="google.cloud.videointelligence.v1.Track.confidence", + index=3, + number=4, + type=2, + cpp_type=6, + label=1, + has_default_value=False, + default_value=float(0), + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=_b("\340A\001"), + file=DESCRIPTOR, + ), + ], + extensions=[], + nested_types=[], + enum_types=[], + serialized_options=None, + is_extendable=False, + syntax="proto3", + extension_ranges=[], + oneofs=[], + serialized_start=3523, + serialized_end=3783, +) + + +_DETECTEDATTRIBUTE = _descriptor.Descriptor( + name="DetectedAttribute", + full_name="google.cloud.videointelligence.v1.DetectedAttribute", + filename=None, + file=DESCRIPTOR, + containing_type=None, + fields=[ + _descriptor.FieldDescriptor( + name="name", + full_name="google.cloud.videointelligence.v1.DetectedAttribute.name", + index=0, + number=1, + type=9, + cpp_type=9, + label=1, + has_default_value=False, + default_value=_b("").decode("utf-8"), + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="confidence", + full_name="google.cloud.videointelligence.v1.DetectedAttribute.confidence", + index=1, + number=2, + type=2, + cpp_type=6, + label=1, + has_default_value=False, + default_value=float(0), + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="value", + full_name="google.cloud.videointelligence.v1.DetectedAttribute.value", + index=2, + number=3, + type=9, + cpp_type=9, + label=1, + has_default_value=False, + default_value=_b("").decode("utf-8"), + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + ], + extensions=[], + nested_types=[], + enum_types=[], + serialized_options=None, + is_extendable=False, + syntax="proto3", + extension_ranges=[], + oneofs=[], + serialized_start=3785, + serialized_end=3853, +) + + +_DETECTEDLANDMARK = _descriptor.Descriptor( + name="DetectedLandmark", + full_name="google.cloud.videointelligence.v1.DetectedLandmark", + filename=None, + file=DESCRIPTOR, + containing_type=None, + fields=[ + _descriptor.FieldDescriptor( + name="name", + full_name="google.cloud.videointelligence.v1.DetectedLandmark.name", + index=0, + number=1, + type=9, + cpp_type=9, + label=1, + has_default_value=False, + default_value=_b("").decode("utf-8"), + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="point", + full_name="google.cloud.videointelligence.v1.DetectedLandmark.point", + index=1, + number=2, + type=11, + cpp_type=10, + label=1, + has_default_value=False, + default_value=None, + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="confidence", + full_name="google.cloud.videointelligence.v1.DetectedLandmark.confidence", + index=2, + number=3, + type=2, + cpp_type=6, + label=1, + has_default_value=False, + default_value=float(0), + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + ], + extensions=[], + nested_types=[], + enum_types=[], + serialized_options=None, + is_extendable=False, + syntax="proto3", + extension_ranges=[], + oneofs=[], + serialized_start=3855, + serialized_end=3975, +) + + _VIDEOANNOTATIONRESULTS = _descriptor.Descriptor( name="VideoAnnotationResults", full_name="google.cloud.videointelligence.v1.VideoAnnotationResults", @@ -1779,10 +2123,28 @@ serialized_options=None, file=DESCRIPTOR, ), + _descriptor.FieldDescriptor( + name="logo_recognition_annotations", + full_name="google.cloud.videointelligence.v1.VideoAnnotationResults.logo_recognition_annotations", + index=13, + number=19, + type=11, + cpp_type=10, + label=3, + has_default_value=False, + default_value=[], + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), _descriptor.FieldDescriptor( name="error", full_name="google.cloud.videointelligence.v1.VideoAnnotationResults.error", - index=13, + index=14, number=9, type=11, cpp_type=10, @@ -1806,8 +2168,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=3206, - serialized_end=4291, + serialized_start=3978, + serialized_end=5163, ) @@ -1845,8 +2207,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=4293, - serialized_end=4403, + serialized_start=5165, + serialized_end=5275, ) @@ -1974,8 +2336,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=4406, - serialized_end=4700, + serialized_start=5278, + serialized_end=5572, ) @@ -2013,8 +2375,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=4702, - serialized_end=4814, + serialized_start=5574, + serialized_end=5686, ) @@ -2196,8 +2558,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=4817, - serialized_end=5202, + serialized_start=5689, + serialized_end=6074, ) @@ -2235,8 +2597,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=5204, - serialized_end=5241, + serialized_start=6076, + serialized_end=6113, ) @@ -2292,8 +2654,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=5244, - serialized_end=5380, + serialized_start=6116, + serialized_end=6252, ) @@ -2367,8 +2729,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=5383, - serialized_end=5523, + serialized_start=6255, + serialized_end=6395, ) @@ -2478,8 +2840,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=5526, - serialized_end=5693, + serialized_start=6398, + serialized_end=6565, ) @@ -2535,8 +2897,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=5695, - serialized_end=5735, + serialized_start=6567, + serialized_end=6607, ) @@ -2574,8 +2936,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=5737, - serialized_end=5832, + serialized_start=6609, + serialized_end=6704, ) @@ -2649,8 +3011,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=5835, - serialized_end=5996, + serialized_start=6707, + serialized_end=6868, ) @@ -2706,8 +3068,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=5999, - serialized_end=6147, + serialized_start=6871, + serialized_end=7019, ) @@ -2763,8 +3125,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=6149, - serialized_end=6245, + serialized_start=7021, + serialized_end=7117, ) @@ -2820,8 +3182,8 @@ syntax="proto3", extension_ranges=[], oneofs=[], - serialized_start=6248, - serialized_end=6408, + serialized_start=7120, + serialized_end=7280, ) @@ -2939,8 +3301,83 @@ fields=[], ) ], - serialized_start=6411, - serialized_end=6690, + serialized_start=7283, + serialized_end=7562, +) + + +_LOGORECOGNITIONANNOTATION = _descriptor.Descriptor( + name="LogoRecognitionAnnotation", + full_name="google.cloud.videointelligence.v1.LogoRecognitionAnnotation", + filename=None, + file=DESCRIPTOR, + containing_type=None, + fields=[ + _descriptor.FieldDescriptor( + name="entity", + full_name="google.cloud.videointelligence.v1.LogoRecognitionAnnotation.entity", + index=0, + number=1, + type=11, + cpp_type=10, + label=1, + has_default_value=False, + default_value=None, + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="tracks", + full_name="google.cloud.videointelligence.v1.LogoRecognitionAnnotation.tracks", + index=1, + number=2, + type=11, + cpp_type=10, + label=3, + has_default_value=False, + default_value=[], + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + _descriptor.FieldDescriptor( + name="segments", + full_name="google.cloud.videointelligence.v1.LogoRecognitionAnnotation.segments", + index=2, + number=3, + type=11, + cpp_type=10, + label=3, + has_default_value=False, + default_value=[], + message_type=None, + enum_type=None, + containing_type=None, + is_extension=False, + extension_scope=None, + serialized_options=None, + file=DESCRIPTOR, + ), + ], + extensions=[], + nested_types=[], + enum_types=[], + serialized_options=None, + is_extendable=False, + syntax="proto3", + extension_ranges=[], + oneofs=[], + serialized_start=7565, + serialized_end=7776, ) _ANNOTATEVIDEOREQUEST.fields_by_name["features"].enum_type = _FEATURE @@ -2998,6 +3435,18 @@ ].message_type = google_dot_protobuf_dot_duration__pb2._DURATION _FACEANNOTATION.fields_by_name["segments"].message_type = _FACESEGMENT _FACEANNOTATION.fields_by_name["frames"].message_type = _FACEFRAME +_TIMESTAMPEDOBJECT.fields_by_name[ + "normalized_bounding_box" +].message_type = _NORMALIZEDBOUNDINGBOX +_TIMESTAMPEDOBJECT.fields_by_name[ + "time_offset" +].message_type = google_dot_protobuf_dot_duration__pb2._DURATION +_TIMESTAMPEDOBJECT.fields_by_name["attributes"].message_type = _DETECTEDATTRIBUTE +_TIMESTAMPEDOBJECT.fields_by_name["landmarks"].message_type = _DETECTEDLANDMARK +_TRACK.fields_by_name["segment"].message_type = _VIDEOSEGMENT +_TRACK.fields_by_name["timestamped_objects"].message_type = _TIMESTAMPEDOBJECT +_TRACK.fields_by_name["attributes"].message_type = _DETECTEDATTRIBUTE +_DETECTEDLANDMARK.fields_by_name["point"].message_type = _NORMALIZEDVERTEX _VIDEOANNOTATIONRESULTS.fields_by_name["segment"].message_type = _VIDEOSEGMENT _VIDEOANNOTATIONRESULTS.fields_by_name[ "segment_label_annotations" @@ -3030,6 +3479,9 @@ _VIDEOANNOTATIONRESULTS.fields_by_name[ "object_annotations" ].message_type = _OBJECTTRACKINGANNOTATION +_VIDEOANNOTATIONRESULTS.fields_by_name[ + "logo_recognition_annotations" +].message_type = _LOGORECOGNITIONANNOTATION _VIDEOANNOTATIONRESULTS.fields_by_name[ "error" ].message_type = google_dot_rpc_dot_status__pb2._STATUS @@ -3089,6 +3541,9 @@ _OBJECTTRACKINGANNOTATION.fields_by_name[ "track_id" ].containing_oneof = _OBJECTTRACKINGANNOTATION.oneofs_by_name["track_info"] +_LOGORECOGNITIONANNOTATION.fields_by_name["entity"].message_type = _ENTITY +_LOGORECOGNITIONANNOTATION.fields_by_name["tracks"].message_type = _TRACK +_LOGORECOGNITIONANNOTATION.fields_by_name["segments"].message_type = _VIDEOSEGMENT DESCRIPTOR.message_types_by_name["AnnotateVideoRequest"] = _ANNOTATEVIDEOREQUEST DESCRIPTOR.message_types_by_name["VideoContext"] = _VIDEOCONTEXT DESCRIPTOR.message_types_by_name["LabelDetectionConfig"] = _LABELDETECTIONCONFIG @@ -3114,6 +3569,10 @@ DESCRIPTOR.message_types_by_name["FaceSegment"] = _FACESEGMENT DESCRIPTOR.message_types_by_name["FaceFrame"] = _FACEFRAME DESCRIPTOR.message_types_by_name["FaceAnnotation"] = _FACEANNOTATION +DESCRIPTOR.message_types_by_name["TimestampedObject"] = _TIMESTAMPEDOBJECT +DESCRIPTOR.message_types_by_name["Track"] = _TRACK +DESCRIPTOR.message_types_by_name["DetectedAttribute"] = _DETECTEDATTRIBUTE +DESCRIPTOR.message_types_by_name["DetectedLandmark"] = _DETECTEDLANDMARK DESCRIPTOR.message_types_by_name["VideoAnnotationResults"] = _VIDEOANNOTATIONRESULTS DESCRIPTOR.message_types_by_name["AnnotateVideoResponse"] = _ANNOTATEVIDEORESPONSE DESCRIPTOR.message_types_by_name["VideoAnnotationProgress"] = _VIDEOANNOTATIONPROGRESS @@ -3134,6 +3593,9 @@ DESCRIPTOR.message_types_by_name["TextAnnotation"] = _TEXTANNOTATION DESCRIPTOR.message_types_by_name["ObjectTrackingFrame"] = _OBJECTTRACKINGFRAME DESCRIPTOR.message_types_by_name["ObjectTrackingAnnotation"] = _OBJECTTRACKINGANNOTATION +DESCRIPTOR.message_types_by_name[ + "LogoRecognitionAnnotation" +] = _LOGORECOGNITIONANNOTATION DESCRIPTOR.enum_types_by_name["Feature"] = _FEATURE DESCRIPTOR.enum_types_by_name["LabelDetectionMode"] = _LABELDETECTIONMODE DESCRIPTOR.enum_types_by_name["Likelihood"] = _LIKELIHOOD @@ -3156,12 +3618,12 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. A video URI may include - wildcards in ``object-id``, and thus identify multiple videos. - Supported wildcards: '\*' to match 0 or more characters; '?' - to match 1 character. If unset, the input video should be - embedded in the request as ``input_content``. If set, - ``input_content`` should be unset. + `__. + A video URI may include wildcards in ``object-id``, and thus + identify multiple videos. Supported wildcards: '\*' to match 0 + or more characters; '?' to match 1 character. If unset, the + input video should be embedded in the request as + ``input_content``. If set, ``input_content`` should be unset. input_content: The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be @@ -3178,7 +3640,7 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. + `__. location_id: Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe- @@ -3616,6 +4078,109 @@ ) _sym_db.RegisterMessage(FaceAnnotation) +TimestampedObject = _reflection.GeneratedProtocolMessageType( + "TimestampedObject", + (_message.Message,), + dict( + DESCRIPTOR=_TIMESTAMPEDOBJECT, + __module__="google.cloud.videointelligence_v1.proto.video_intelligence_pb2", + __doc__="""For tracking related features. An object at time\_offset + with attributes, and located with normalized\_bounding\_box. + + + Attributes: + normalized_bounding_box: + Normalized Bounding box in a frame, where the object is + located. + time_offset: + Time-offset, relative to the beginning of the video, + corresponding to the video frame for this object. + attributes: + Optional. The attributes of the object in the bounding box. + landmarks: + Optional. The detected landmarks. + """, + # @@protoc_insertion_point(class_scope:google.cloud.videointelligence.v1.TimestampedObject) + ), +) +_sym_db.RegisterMessage(TimestampedObject) + +Track = _reflection.GeneratedProtocolMessageType( + "Track", + (_message.Message,), + dict( + DESCRIPTOR=_TRACK, + __module__="google.cloud.videointelligence_v1.proto.video_intelligence_pb2", + __doc__="""A track of an object instance. + + + Attributes: + segment: + Video segment of a track. + timestamped_objects: + The object with timestamp and attributes per frame in the + track. + attributes: + Optional. Attributes in the track level. + confidence: + Optional. The confidence score of the tracked object. + """, + # @@protoc_insertion_point(class_scope:google.cloud.videointelligence.v1.Track) + ), +) +_sym_db.RegisterMessage(Track) + +DetectedAttribute = _reflection.GeneratedProtocolMessageType( + "DetectedAttribute", + (_message.Message,), + dict( + DESCRIPTOR=_DETECTEDATTRIBUTE, + __module__="google.cloud.videointelligence_v1.proto.video_intelligence_pb2", + __doc__="""A generic detected attribute represented by name in string + format. + + + Attributes: + name: + The name of the attribute, i.e. glasses, dark\_glasses, + mouth\_open etc. A full list of supported type names will be + provided in the document. + confidence: + Detected attribute confidence. Range [0, 1]. + value: + Text value of the detection result. For example, the value for + "HairColor" can be "black", "blonde", etc. + """, + # @@protoc_insertion_point(class_scope:google.cloud.videointelligence.v1.DetectedAttribute) + ), +) +_sym_db.RegisterMessage(DetectedAttribute) + +DetectedLandmark = _reflection.GeneratedProtocolMessageType( + "DetectedLandmark", + (_message.Message,), + dict( + DESCRIPTOR=_DETECTEDLANDMARK, + __module__="google.cloud.videointelligence_v1.proto.video_intelligence_pb2", + __doc__="""A generic detected landmark represented by name in string + format and a 2D location. + + + Attributes: + name: + The name of this landmark, i.e. left\_hand, right\_shoulder. + point: + The 2D point of the detected landmark using the normalized + image coordindate system. The normalized coordinates have the + range from 0 to 1. + confidence: + The confidence score of the detected landmark. Range [0, 1]. + """, + # @@protoc_insertion_point(class_scope:google.cloud.videointelligence.v1.DetectedLandmark) + ), +) +_sym_db.RegisterMessage(DetectedLandmark) + VideoAnnotationResults = _reflection.GeneratedProtocolMessageType( "VideoAnnotationResults", (_message.Message,), @@ -3673,6 +4238,9 @@ information associated with it. object_annotations: Annotations for list of objects detected and tracked in video. + logo_recognition_annotations: + Annotations for list of logos detected, tracked and recognized + in video. error: If set, indicates an error. Note that for a single ``AnnotateVideoRequest`` some videos may succeed and some may @@ -3833,8 +4401,8 @@ dict( DESCRIPTOR=_SPEECHCONTEXT, __module__="google.cloud.videointelligence_v1.proto.video_intelligence_pb2", - __doc__="""Provides "hints" to the speech recognizer to favor - specific words and phrases in the results. + __doc__="""Provides "hints" to the speech recognizer to favor specific words and + phrases in the results. Attributes: @@ -4136,11 +4704,44 @@ ) _sym_db.RegisterMessage(ObjectTrackingAnnotation) +LogoRecognitionAnnotation = _reflection.GeneratedProtocolMessageType( + "LogoRecognitionAnnotation", + (_message.Message,), + dict( + DESCRIPTOR=_LOGORECOGNITIONANNOTATION, + __module__="google.cloud.videointelligence_v1.proto.video_intelligence_pb2", + __doc__="""Annotation corresponding to one detected, tracked and + recognized logo class. + + + Attributes: + entity: + Entity category information to specify the logo class that all + the logo tracks within this LogoRecognitionAnnotation are + recognized as. + tracks: + All logo tracks where the recognized logo appears. Each track + corresponds to one logo instance appearing in consecutive + frames. + segments: + All video segments where the recognized logo appears. There + might be multiple instances of the same logo class appearing + in one VideoSegment. + """, + # @@protoc_insertion_point(class_scope:google.cloud.videointelligence.v1.LogoRecognitionAnnotation) + ), +) +_sym_db.RegisterMessage(LogoRecognitionAnnotation) + DESCRIPTOR._options = None _ANNOTATEVIDEOREQUEST.fields_by_name["features"]._options = None _ANNOTATEVIDEOREQUEST.fields_by_name["output_uri"]._options = None _ANNOTATEVIDEOREQUEST.fields_by_name["location_id"]._options = None +_TIMESTAMPEDOBJECT.fields_by_name["attributes"]._options = None +_TIMESTAMPEDOBJECT.fields_by_name["landmarks"]._options = None +_TRACK.fields_by_name["attributes"]._options = None +_TRACK.fields_by_name["confidence"]._options = None _SPEECHTRANSCRIPTIONCONFIG.fields_by_name["language_code"]._options = None _SPEECHTRANSCRIPTIONCONFIG.fields_by_name["max_alternatives"]._options = None _SPEECHTRANSCRIPTIONCONFIG.fields_by_name["filter_profanity"]._options = None @@ -4167,8 +4768,8 @@ serialized_options=_b( "\312A videointelligence.googleapis.com\322A.https://www.googleapis.com/auth/cloud-platform" ), - serialized_start=7131, - serialized_end=7451, + serialized_start=8239, + serialized_end=8559, methods=[ _descriptor.MethodDescriptor( name="AnnotateVideo", diff --git a/google/cloud/videointelligence_v1beta2/gapic/video_intelligence_service_client.py b/google/cloud/videointelligence_v1beta2/gapic/video_intelligence_service_client.py index 7b7023ea..72fd9a1f 100644 --- a/google/cloud/videointelligence_v1beta2/gapic/video_intelligence_service_client.py +++ b/google/cloud/videointelligence_v1beta2/gapic/video_intelligence_service_client.py @@ -234,11 +234,12 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. A video - URI may include wildcards in ``object-id``, and thus identify multiple - videos. Supported wildcards: '\*' to match 0 or more characters; '?' to - match 1 character. If unset, the input video should be embedded in the - request as ``input_content``. If set, ``input_content`` should be unset. + URIs `__. A + video URI may include wildcards in ``object-id``, and thus identify + multiple videos. Supported wildcards: '\*' to match 0 or more + characters; '?' to match 1 character. If unset, the input video should + be embedded in the request as ``input_content``. If set, + ``input_content`` should be unset. input_content (bytes): The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be unset. features (list[~google.cloud.videointelligence_v1beta2.types.Feature]): Required. Requested video annotation features. @@ -252,7 +253,7 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. + URIs `__. location_id (str): Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe-west1``, ``asia-east1``. If no region is specified, a region will be determined diff --git a/google/cloud/videointelligence_v1beta2/proto/video_intelligence.proto b/google/cloud/videointelligence_v1beta2/proto/video_intelligence.proto index 8e80640e..69009975 100644 --- a/google/cloud/videointelligence_v1beta2/proto/video_intelligence.proto +++ b/google/cloud/videointelligence_v1beta2/proto/video_intelligence.proto @@ -64,7 +64,7 @@ message AnnotateVideoRequest { // supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For - // more information, see [Request URIs](/storage/docs/reference-uris). A video + // more information, see [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). A video // URI may include wildcards in `object-id`, and thus identify multiple // videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded @@ -87,7 +87,7 @@ message AnnotateVideoRequest { // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For - // more information, see [Request URIs](/storage/docs/reference-uris). + // more information, see [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). string output_uri = 4 [(google.api.field_behavior) = OPTIONAL]; // Optional. Cloud region where annotation should take place. Supported cloud diff --git a/google/cloud/videointelligence_v1beta2/proto/video_intelligence_pb2.py b/google/cloud/videointelligence_v1beta2/proto/video_intelligence_pb2.py index 16648add..d72ea1e2 100644 --- a/google/cloud/videointelligence_v1beta2/proto/video_intelligence_pb2.py +++ b/google/cloud/videointelligence_v1beta2/proto/video_intelligence_pb2.py @@ -1793,12 +1793,12 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. A video URI may include - wildcards in ``object-id``, and thus identify multiple videos. - Supported wildcards: '\*' to match 0 or more characters; '?' - to match 1 character. If unset, the input video should be - embedded in the request as ``input_content``. If set, - ``input_content`` should be unset. + `__. + A video URI may include wildcards in ``object-id``, and thus + identify multiple videos. Supported wildcards: '\*' to match 0 + or more characters; '?' to match 1 character. If unset, the + input video should be embedded in the request as + ``input_content``. If set, ``input_content`` should be unset. input_content: The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be @@ -1815,7 +1815,7 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. + `__. location_id: Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe- @@ -1971,7 +1971,8 @@ dict( DESCRIPTOR=_LABELSEGMENT, __module__="google.cloud.videointelligence_v1beta2.proto.video_intelligence_pb2", - __doc__="""Video segment level annotation results for label detection. + __doc__="""Video segment level annotation results for label + detection. Attributes: diff --git a/google/cloud/videointelligence_v1p1beta1/gapic/video_intelligence_service_client.py b/google/cloud/videointelligence_v1p1beta1/gapic/video_intelligence_service_client.py index 51d50777..c46c50f8 100644 --- a/google/cloud/videointelligence_v1p1beta1/gapic/video_intelligence_service_client.py +++ b/google/cloud/videointelligence_v1p1beta1/gapic/video_intelligence_service_client.py @@ -236,11 +236,12 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. A video - URI may include wildcards in ``object-id``, and thus identify multiple - videos. Supported wildcards: '\*' to match 0 or more characters; '?' to - match 1 character. If unset, the input video should be embedded in the - request as ``input_content``. If set, ``input_content`` should be unset. + URIs `__. A + video URI may include wildcards in ``object-id``, and thus identify + multiple videos. Supported wildcards: '\*' to match 0 or more + characters; '?' to match 1 character. If unset, the input video should + be embedded in the request as ``input_content``. If set, + ``input_content`` should be unset. input_content (bytes): The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be unset. features (list[~google.cloud.videointelligence_v1p1beta1.types.Feature]): Required. Requested video annotation features. @@ -254,7 +255,7 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. + URIs `__. location_id (str): Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe-west1``, ``asia-east1``. If no region is specified, a region will be determined diff --git a/google/cloud/videointelligence_v1p1beta1/proto/video_intelligence.proto b/google/cloud/videointelligence_v1p1beta1/proto/video_intelligence.proto index 44d3ca64..a54bddd0 100644 --- a/google/cloud/videointelligence_v1p1beta1/proto/video_intelligence.proto +++ b/google/cloud/videointelligence_v1p1beta1/proto/video_intelligence.proto @@ -62,7 +62,7 @@ message AnnotateVideoRequest { // supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see - // [Request URIs](/storage/docs/reference-uris). + // [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). // A video URI may include wildcards in `object-id`, and thus identify // multiple videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded @@ -85,7 +85,7 @@ message AnnotateVideoRequest { // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see - // [Request URIs](/storage/docs/reference-uris). + // [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). string output_uri = 4 [(google.api.field_behavior) = OPTIONAL]; // Optional. Cloud region where annotation should take place. Supported cloud diff --git a/google/cloud/videointelligence_v1p1beta1/proto/video_intelligence_pb2.py b/google/cloud/videointelligence_v1p1beta1/proto/video_intelligence_pb2.py index 650598ba..939c5950 100644 --- a/google/cloud/videointelligence_v1p1beta1/proto/video_intelligence_pb2.py +++ b/google/cloud/videointelligence_v1p1beta1/proto/video_intelligence_pb2.py @@ -1841,12 +1841,12 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. A video URI may include - wildcards in ``object-id``, and thus identify multiple videos. - Supported wildcards: '\*' to match 0 or more characters; '?' - to match 1 character. If unset, the input video should be - embedded in the request as ``input_content``. If set, - ``input_content`` should be unset. + `__. + A video URI may include wildcards in ``object-id``, and thus + identify multiple videos. Supported wildcards: '\*' to match 0 + or more characters; '?' to match 1 character. If unset, the + input video should be embedded in the request as + ``input_content``. If set, ``input_content`` should be unset. input_content: The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be @@ -1863,7 +1863,7 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. + `__. location_id: Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe- diff --git a/google/cloud/videointelligence_v1p2beta1/gapic/video_intelligence_service_client.py b/google/cloud/videointelligence_v1p2beta1/gapic/video_intelligence_service_client.py index acc1fd42..56f54b69 100644 --- a/google/cloud/videointelligence_v1p2beta1/gapic/video_intelligence_service_client.py +++ b/google/cloud/videointelligence_v1p2beta1/gapic/video_intelligence_service_client.py @@ -236,11 +236,12 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. A video - URI may include wildcards in ``object-id``, and thus identify multiple - videos. Supported wildcards: '\*' to match 0 or more characters; '?' to - match 1 character. If unset, the input video should be embedded in the - request as ``input_content``. If set, ``input_content`` should be unset. + URIs `__. A + video URI may include wildcards in ``object-id``, and thus identify + multiple videos. Supported wildcards: '\*' to match 0 or more + characters; '?' to match 1 character. If unset, the input video should + be embedded in the request as ``input_content``. If set, + ``input_content`` should be unset. input_content (bytes): The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be unset. features (list[~google.cloud.videointelligence_v1p2beta1.types.Feature]): Required. Requested video annotation features. @@ -254,7 +255,7 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. + URIs `__. location_id (str): Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe-west1``, ``asia-east1``. If no region is specified, a region will be determined diff --git a/google/cloud/videointelligence_v1p2beta1/proto/video_intelligence.proto b/google/cloud/videointelligence_v1p2beta1/proto/video_intelligence.proto index 044233b0..3b1d51cd 100644 --- a/google/cloud/videointelligence_v1p2beta1/proto/video_intelligence.proto +++ b/google/cloud/videointelligence_v1p2beta1/proto/video_intelligence.proto @@ -62,7 +62,7 @@ message AnnotateVideoRequest { // supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see - // [Request URIs](/storage/docs/reference-uris). + // [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). // A video URI may include wildcards in `object-id`, and thus identify // multiple videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded @@ -85,7 +85,7 @@ message AnnotateVideoRequest { // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see - // [Request URIs](/storage/docs/reference-uris). + // [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). string output_uri = 4 [(google.api.field_behavior) = OPTIONAL]; // Optional. Cloud region where annotation should take place. Supported cloud diff --git a/google/cloud/videointelligence_v1p2beta1/proto/video_intelligence_pb2.py b/google/cloud/videointelligence_v1p2beta1/proto/video_intelligence_pb2.py index 3923c260..b55b39d9 100644 --- a/google/cloud/videointelligence_v1p2beta1/proto/video_intelligence_pb2.py +++ b/google/cloud/videointelligence_v1p2beta1/proto/video_intelligence_pb2.py @@ -2080,12 +2080,12 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. A video URI may include - wildcards in ``object-id``, and thus identify multiple videos. - Supported wildcards: '\*' to match 0 or more characters; '?' - to match 1 character. If unset, the input video should be - embedded in the request as ``input_content``. If set, - ``input_content`` should be unset. + `__. + A video URI may include wildcards in ``object-id``, and thus + identify multiple videos. Supported wildcards: '\*' to match 0 + or more characters; '?' to match 1 character. If unset, the + input video should be embedded in the request as + ``input_content``. If set, ``input_content`` should be unset. input_content: The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be @@ -2102,7 +2102,7 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. + `__. location_id: Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe- @@ -2459,8 +2459,8 @@ dict( DESCRIPTOR=_ANNOTATEVIDEORESPONSE, __module__="google.cloud.videointelligence_v1p2beta1.proto.video_intelligence_pb2", - __doc__="""Video annotation response. Included in the ``response`` field of the - ``Operation`` returned by the ``GetOperation`` call of the + __doc__="""Video annotation response. Included in the ``response`` + field of the ``Operation`` returned by the ``GetOperation`` call of the ``google::longrunning::Operations`` service. diff --git a/google/cloud/videointelligence_v1p3beta1/gapic/streaming_video_intelligence_service_client.py b/google/cloud/videointelligence_v1p3beta1/gapic/streaming_video_intelligence_service_client.py index e89daf5a..27fd9a98 100644 --- a/google/cloud/videointelligence_v1p3beta1/gapic/streaming_video_intelligence_service_client.py +++ b/google/cloud/videointelligence_v1p3beta1/gapic/streaming_video_intelligence_service_client.py @@ -202,8 +202,6 @@ def streaming_annotate_video( while sending video/audio bytes. This method is only available via the gRPC API (not REST). - EXPERIMENTAL: This method interface might change in the future. - Example: >>> from google.cloud import videointelligence_v1p3beta1 >>> diff --git a/google/cloud/videointelligence_v1p3beta1/gapic/video_intelligence_service_client.py b/google/cloud/videointelligence_v1p3beta1/gapic/video_intelligence_service_client.py index d94df648..f9c553f4 100644 --- a/google/cloud/videointelligence_v1p3beta1/gapic/video_intelligence_service_client.py +++ b/google/cloud/videointelligence_v1p3beta1/gapic/video_intelligence_service_client.py @@ -236,11 +236,12 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. A video - URI may include wildcards in ``object-id``, and thus identify multiple - videos. Supported wildcards: '\*' to match 0 or more characters; '?' to - match 1 character. If unset, the input video should be embedded in the - request as ``input_content``. If set, ``input_content`` should be unset. + URIs `__. A + video URI may include wildcards in ``object-id``, and thus identify + multiple videos. Supported wildcards: '\*' to match 0 or more + characters; '?' to match 1 character. If unset, the input video should + be embedded in the request as ``input_content``. If set, + ``input_content`` should be unset. input_content (bytes): The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be unset. features (list[~google.cloud.videointelligence_v1p3beta1.types.Feature]): Required. Requested video annotation features. @@ -254,7 +255,7 @@ def annotate_video( must be specified in the following format: ``gs://bucket-id/object-id`` (other URI formats return ``google.rpc.Code.INVALID_ARGUMENT``). For more information, see `Request - URIs `__. + URIs `__. location_id (str): Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe-west1``, ``asia-east1``. If no region is specified, a region will be determined diff --git a/google/cloud/videointelligence_v1p3beta1/proto/video_intelligence.proto b/google/cloud/videointelligence_v1p3beta1/proto/video_intelligence.proto index 942f63be..3d418e2f 100644 --- a/google/cloud/videointelligence_v1p3beta1/proto/video_intelligence.proto +++ b/google/cloud/videointelligence_v1p3beta1/proto/video_intelligence.proto @@ -76,7 +76,7 @@ message AnnotateVideoRequest { // supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For - // more information, see [Request URIs](/storage/docs/reference-uris). A video + // more information, see [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). A video // URI may include wildcards in `object-id`, and thus identify multiple // videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded @@ -99,7 +99,7 @@ message AnnotateVideoRequest { // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For - // more information, see [Request URIs](/storage/docs/reference-uris). + // more information, see [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). string output_uri = 4 [(google.api.field_behavior) = OPTIONAL]; // Optional. Cloud region where annotation should take place. Supported cloud diff --git a/google/cloud/videointelligence_v1p3beta1/proto/video_intelligence_pb2.py b/google/cloud/videointelligence_v1p3beta1/proto/video_intelligence_pb2.py index c0e2bee2..dd9b41b9 100644 --- a/google/cloud/videointelligence_v1p3beta1/proto/video_intelligence_pb2.py +++ b/google/cloud/videointelligence_v1p3beta1/proto/video_intelligence_pb2.py @@ -4785,12 +4785,12 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. A video URI may include - wildcards in ``object-id``, and thus identify multiple videos. - Supported wildcards: '\*' to match 0 or more characters; '?' - to match 1 character. If unset, the input video should be - embedded in the request as ``input_content``. If set, - ``input_content`` should be unset. + `__. + A video URI may include wildcards in ``object-id``, and thus + identify multiple videos. Supported wildcards: '\*' to match 0 + or more characters; '?' to match 1 character. If unset, the + input video should be embedded in the request as + ``input_content``. If set, ``input_content`` should be unset. input_content: The video data bytes. If unset, the input video(s) should be specified via ``input_uri``. If set, ``input_uri`` should be @@ -4807,7 +4807,7 @@ ``gs://bucket-id/object-id`` (other URI formats return [google .rpc.Code.INVALID\_ARGUMENT][google.rpc.Code.INVALID\_ARGUMENT ]). For more information, see `Request URIs - `__. + `__. location_id: Optional. Cloud region where annotation should take place. Supported cloud regions: ``us-east1``, ``us-west1``, ``europe- @@ -5522,8 +5522,8 @@ dict( DESCRIPTOR=_ANNOTATEVIDEORESPONSE, __module__="google.cloud.videointelligence_v1p3beta1.proto.video_intelligence_pb2", - __doc__="""Video annotation response. Included in the ``response`` field of the - ``Operation`` returned by the ``GetOperation`` call of the + __doc__="""Video annotation response. Included in the ``response`` + field of the ``Operation`` returned by the ``GetOperation`` call of the ``google::longrunning::Operations`` service. diff --git a/synth.metadata b/synth.metadata index 1d4ea0f0..8e1a9b4b 100644 --- a/synth.metadata +++ b/synth.metadata @@ -1,39 +1,25 @@ { - "updateTime": "2020-02-20T23:21:11.311856Z", + "updateTime": "2020-03-12T12:16:45.021680Z", "sources": [ { "generator": { "name": "artman", - "version": "0.45.0", - "dockerImage": "googleapis/artman@sha256:6aec9c34db0e4be221cdaf6faba27bdc07cfea846808b3d3b964dfce3a9a0f9b" - } - }, - { - "git": { - "name": ".", - "remote": "https://github.com/googleapis/python-videointelligence.git", - "sha": "1385b87c4bf649266373c8b12bdb93abf60a31d3" + "version": "1.1.0", + "dockerImage": "googleapis/artman@sha256:f54b7644a1d2e7a37b23f5c0dfe9bba473e41c675002a507a244389e27487ca9" } }, { "git": { "name": "googleapis", "remote": "https://github.com/googleapis/googleapis.git", - "sha": "3eaaaf8626ce5b0c0bc7eee05e143beffa373b01", - "internalRef": "296274723", - "log": "3eaaaf8626ce5b0c0bc7eee05e143beffa373b01\nAdd BUILD.bazel for v1 secretmanager.googleapis.com\n\nPiperOrigin-RevId: 296274723\n\ne76149c3d992337f85eeb45643106aacae7ede82\nMove securitycenter v1 to use generate from annotations.\n\nPiperOrigin-RevId: 296266862\n\n203740c78ac69ee07c3bf6be7408048751f618f8\nAdd StackdriverLoggingConfig field to Cloud Tasks v2 API.\n\nPiperOrigin-RevId: 296256388\n\ne4117d5e9ed8bbca28da4a60a94947ca51cb2083\nCreate a Bazel BUILD file for the google.actions.type export.\n\nPiperOrigin-RevId: 296212567\n\na9639a0a9854fd6e1be08bba1ac3897f4f16cb2f\nAdd secretmanager.googleapis.com v1 protos\n\nPiperOrigin-RevId: 295983266\n\n" - } - }, - { - "git": { - "name": "synthtool", - "remote": "rpc://devrel/cloud/libraries/tools/autosynth", - "sha": "706a38c26db42299845396cdae55db635c38794a" + "sha": "34a5450c591b6be3d6566f25ac31caa5211b2f3f", + "internalRef": "300474272", + "log": "34a5450c591b6be3d6566f25ac31caa5211b2f3f\nIncreases the default timeout from 20s to 30s for MetricService\n\nPiperOrigin-RevId: 300474272\n\n5d8bffe87cd01ba390c32f1714230e5a95d5991d\nfeat: use the latest gapic-generator in WORKSPACE for bazel build.\n\nPiperOrigin-RevId: 300461878\n\nd631c651e3bcfac5d371e8560c27648f7b3e2364\nUpdated the GAPIC configs to include parameters for Backups APIs.\n\nPiperOrigin-RevId: 300443402\n\n678afc7055c1adea9b7b54519f3bdb228013f918\nAdding Game Servers v1beta API.\n\nPiperOrigin-RevId: 300433218\n\n80d2bd2c652a5e213302041b0620aff423132589\nEnable proto annotation and gapic v2 for talent API.\n\nPiperOrigin-RevId: 300393997\n\n85e454be7a353f7fe1bf2b0affb753305785b872\ndocs(google/maps/roads): remove mention of nonexported api\n\nPiperOrigin-RevId: 300367734\n\nbf839ae632e0f263a729569e44be4b38b1c85f9c\nAdding protocol buffer annotations and updated config info for v1 and v2.\n\nPiperOrigin-RevId: 300276913\n\n309b899ca18a4c604bce63882a161d44854da549\nPublish `Backup` APIs and protos.\n\nPiperOrigin-RevId: 300246038\n\neced64c3f122421350b4aca68a28e89121d20db8\nadd PHP client libraries\n\nPiperOrigin-RevId: 300193634\n\n7727af0e39df1ae9ad715895c8576d7b65cf6c6d\nfeat: use the latest gapic-generator and protoc-java-resource-name-plugin in googleapis/WORKSPACE.\n\nPiperOrigin-RevId: 300188410\n\n2a25aa351dd5b5fe14895266aff5824d90ce757b\nBreaking change: remove the ProjectOrTenant resource and its references.\n\nPiperOrigin-RevId: 300182152\n\na499dbb28546379415f51803505cfb6123477e71\nUpdate web risk v1 gapic config and BUILD file.\n\nPiperOrigin-RevId: 300152177\n\n52701da10fec2a5f9796e8d12518c0fe574488fe\nFix: apply appropriate namespace/package options for C#, PHP and Ruby.\n\nPiperOrigin-RevId: 300123508\n\n365c029b8cdb63f7751b92ab490f1976e616105c\nAdd CC targets to the kms protos.\n\nThese are needed by go/tink.\n\nPiperOrigin-RevId: 300038469\n\n4ba9aa8a4a1413b88dca5a8fa931824ee9c284e6\nExpose logo recognition API proto for GA.\n\nPiperOrigin-RevId: 299971671\n\n1c9fc2c9e03dadf15f16b1c4f570955bdcebe00e\nAdding ruby_package option to accessapproval.proto for the Ruby client libraries generation.\n\nPiperOrigin-RevId: 299955924\n\n1cc6f0a7bfb147e6f2ede911d9b01e7a9923b719\nbuild(google/maps/routes): generate api clients\n\nPiperOrigin-RevId: 299955905\n\n29a47c965aac79e3fe8e3314482ca0b5967680f0\nIncrease timeout to 1hr for method `dropRange` in bigtable/admin/v2, which is\nsynced with the timeout setting in gapic_yaml.\n\nPiperOrigin-RevId: 299917154\n\n8f631c4c70a60a9c7da3749511ee4ad432b62898\nbuild(google/maps/roads/v1op): move go to monorepo pattern\n\nPiperOrigin-RevId: 299885195\n\nd66816518844ebbf63504c9e8dfc7133921dd2cd\nbuild(google/maps/roads/v1op): Add bazel build files to generate clients.\n\nPiperOrigin-RevId: 299851148\n\naf7dff701fabe029672168649c62356cf1bb43d0\nAdd LogPlayerReports and LogImpressions to Playable Locations service\n\nPiperOrigin-RevId: 299724050\n\nb6927fca808f38df32a642c560082f5bf6538ced\nUpdate BigQuery Connection API v1beta1 proto: added credential to CloudSqlProperties.\n\nPiperOrigin-RevId: 299503150\n\n91e1fb5ef9829c0c7a64bfa5bde330e6ed594378\nchore: update protobuf (protoc) version to 3.11.2\n\nPiperOrigin-RevId: 299404145\n\n30e36b4bee6749c4799f4fc1a51cc8f058ba167d\nUpdate cloud asset api v1p4beta1.\n\nPiperOrigin-RevId: 299399890\n\nffbb493674099f265693872ae250711b2238090c\nfeat: cloudbuild/v1 add new fields and annotate OUTPUT_OUT fields.\n\nPiperOrigin-RevId: 299397780\n\nbc973a15818e00c19e121959832676e9b7607456\nbazel: Fix broken common dependency\n\nPiperOrigin-RevId: 299397431\n\n71094a343e3b962e744aa49eb9338219537474e4\nchore: bigtable/admin/v2 publish retry config\n\nPiperOrigin-RevId: 299391875\n\n8f488efd7bda33885cb674ddd023b3678c40bd82\nfeat: Migrate logging to GAPIC v2; release new features.\n\nIMPORTANT: This is a breaking change for client libraries\nin all languages.\n\nCommitter: @lukesneeringer, @jskeet\nPiperOrigin-RevId: 299370279\n\n007605bf9ad3a1fd775014ebefbf7f1e6b31ee71\nUpdate API for bigqueryreservation v1beta1.\n- Adds flex capacity commitment plan to CapacityCommitment.\n- Adds methods for getting and updating BiReservations.\n- Adds methods for updating/splitting/merging CapacityCommitments.\n\nPiperOrigin-RevId: 299368059\n\nf0b581b5bdf803e45201ecdb3688b60e381628a8\nfix: recommendationengine/v1beta1 update some comments\n\nPiperOrigin-RevId: 299181282\n\n10e9a0a833dc85ff8f05b2c67ebe5ac785fe04ff\nbuild: add generated BUILD file for Routes Preferred API\n\nPiperOrigin-RevId: 299164808\n\n86738c956a8238d7c77f729be78b0ed887a6c913\npublish v1p1beta1: update with absolute address in comments\n\nPiperOrigin-RevId: 299152383\n\n73d9f2ad4591de45c2e1f352bc99d70cbd2a6d95\npublish v1: update with absolute address in comments\n\nPiperOrigin-RevId: 299147194\n\nd2158f24cb77b0b0ccfe68af784c6a628705e3c6\npublish v1beta2: update with absolute address in comments\n\nPiperOrigin-RevId: 299147086\n\n7fca61292c11b4cd5b352cee1a50bf88819dd63b\npublish v1p2beta1: update with absolute address in comments\n\nPiperOrigin-RevId: 299146903\n\n583b7321624736e2c490e328f4b1957335779295\npublish v1p3beta1: update with absolute address in comments\n\nPiperOrigin-RevId: 299146674\n\n638253bf86d1ce1c314108a089b7351440c2f0bf\nfix: add java_multiple_files option for automl text_sentiment.proto\n\nPiperOrigin-RevId: 298971070\n\n373d655703bf914fb8b0b1cc4071d772bac0e0d1\nUpdate Recs AI Beta public bazel file\n\nPiperOrigin-RevId: 298961623\n\ndcc5d00fc8a8d8b56f16194d7c682027b2c66a3b\nfix: add java_multiple_files option for automl classification.proto\n\nPiperOrigin-RevId: 298953301\n\na3f791827266f3496a6a5201d58adc4bb265c2a3\nchore: automl/v1 publish annotations and retry config\n\nPiperOrigin-RevId: 298942178\n\n01c681586d8d6dbd60155289b587aee678530bd9\nMark return_immediately in PullRequest deprecated.\n\nPiperOrigin-RevId: 298893281\n\nc9f5e9c4bfed54bbd09227e990e7bded5f90f31c\nRemove out of date documentation for predicate support on the Storage API\n\nPiperOrigin-RevId: 298883309\n\nfd5b3b8238d783b04692a113ffe07c0363f5de0f\ngenerate webrisk v1 proto\n\nPiperOrigin-RevId: 298847934\n\n541b1ded4abadcc38e8178680b0677f65594ea6f\nUpdate cloud asset api v1p4beta1.\n\nPiperOrigin-RevId: 298686266\n\nc0d171acecb4f5b0bfd2c4ca34fc54716574e300\n Updated to include the Notification v1 API.\n\nPiperOrigin-RevId: 298652775\n\n2346a9186c0bff2c9cc439f2459d558068637e05\nAdd Service Directory v1beta1 protos and configs\n\nPiperOrigin-RevId: 298625638\n\na78ed801b82a5c6d9c5368e24b1412212e541bb7\nPublishing v3 protos and configs.\n\nPiperOrigin-RevId: 298607357\n\n4a180bfff8a21645b3a935c2756e8d6ab18a74e0\nautoml/v1beta1 publish proto updates\n\nPiperOrigin-RevId: 298484782\n\n6de6e938b7df1cd62396563a067334abeedb9676\nchore: use the latest gapic-generator and protoc-java-resource-name-plugin in Bazel workspace.\n\nPiperOrigin-RevId: 298474513\n\n244ab2b83a82076a1fa7be63b7e0671af73f5c02\nAdds service config definition for bigqueryreservation v1\n\nPiperOrigin-RevId: 298455048\n\n83c6f84035ee0f80eaa44d8b688a010461cc4080\nUpdate google/api/auth.proto to make AuthProvider to have JwtLocation\n\nPiperOrigin-RevId: 297918498\n\ne9e90a787703ec5d388902e2cb796aaed3a385b4\nDialogflow weekly v2/v2beta1 library update:\n - adding get validation result\n - adding field mask override control for output audio config\nImportant updates are also posted at:\nhttps://cloud.google.com/dialogflow/docs/release-notes\n\nPiperOrigin-RevId: 297671458\n\n1a2b05cc3541a5f7714529c665aecc3ea042c646\nAdding .yaml and .json config files.\n\nPiperOrigin-RevId: 297570622\n\ndfe1cf7be44dee31d78f78e485d8c95430981d6e\nPublish `QueryOptions` proto.\n\nIntroduced a `query_options` input in `ExecuteSqlRequest`.\n\nPiperOrigin-RevId: 297497710\n\ndafc905f71e5d46f500b41ed715aad585be062c3\npubsub: revert pull init_rpc_timeout & max_rpc_timeout back to 25 seconds and reset multiplier to 1.0\n\nPiperOrigin-RevId: 297486523\n\nf077632ba7fee588922d9e8717ee272039be126d\nfirestore: add update_transform\n\nPiperOrigin-RevId: 297405063\n\n0aba1900ffef672ec5f0da677cf590ee5686e13b\ncluster: use square brace for cross-reference\n\nPiperOrigin-RevId: 297204568\n\n5dac2da18f6325cbaed54603c43f0667ecd50247\nRestore retry params in gapic config because securitycenter has non-standard default retry params.\nRestore a few retry codes for some idempotent methods.\n\nPiperOrigin-RevId: 297196720\n\n1eb61455530252bba8b2c8d4bc9832960e5a56f6\npubsub: v1 replace IAM HTTP rules\n\nPiperOrigin-RevId: 297188590\n\n80b2d25f8d43d9d47024ff06ead7f7166548a7ba\nDialogflow weekly v2/v2beta1 library update:\n - updates to mega agent api\n - adding field mask override control for output audio config\nImportant updates are also posted at:\nhttps://cloud.google.com/dialogflow/docs/release-notes\n\nPiperOrigin-RevId: 297187629\n\n0b1876b35e98f560f9c9ca9797955f020238a092\nUse an older version of protoc-docs-plugin that is compatible with the specified gapic-generator and protobuf versions.\n\nprotoc-docs-plugin >=0.4.0 (see commit https://github.com/googleapis/protoc-docs-plugin/commit/979f03ede6678c487337f3d7e88bae58df5207af) is incompatible with protobuf 3.9.1.\n\nPiperOrigin-RevId: 296986742\n\n1e47e676cddbbd8d93f19ba0665af15b5532417e\nFix: Restore a method signature for UpdateCluster\n\nPiperOrigin-RevId: 296901854\n\n7f910bcc4fc4704947ccfd3ceed015d16b9e00c2\nUpdate Dataproc v1beta2 client.\n\nPiperOrigin-RevId: 296451205\n\nde287524405a3dce124d301634731584fc0432d7\nFix: Reinstate method signatures that had been missed off some RPCs\nFix: Correct resource types for two fields\n\nPiperOrigin-RevId: 296435091\n\ne5bc9566ae057fb4c92f8b7e047f1c8958235b53\nDeprecate the endpoint_uris field, as it is unused.\n\nPiperOrigin-RevId: 296357191\n\n8c12e2b4dca94e12bff9f538bdac29524ff7ef7a\nUpdate Dataproc v1 client.\n\nPiperOrigin-RevId: 296336662\n\n17567c4a1ef0a9b50faa87024d66f8acbb561089\nRemoving erroneous comment, a la https://github.com/googleapis/java-speech/pull/103\n\nPiperOrigin-RevId: 296332968\n\n" } }, { "template": { - "name": "python_split_library", + "name": "python_library", "origin": "synthtool.gcp", "version": "2020.2.4" }