Members
(static, constant) Feature :number
Video annotation feature.
Properties:
Name | Type | Description |
---|---|---|
FEATURE_UNSPECIFIED |
number |
Unspecified. |
LABEL_DETECTION |
number |
Label detection. Detect objects, such as dog or flower. |
FACE_DETECTION |
number |
Human face detection and tracking. |
SHOT_CHANGE_DETECTION |
number |
Shot change detection. |
SAFE_SEARCH_DETECTION |
number |
Safe search detection. |
(static, constant) LabelDetectionMode :number
Label detection mode.
Properties:
Name | Type | Description |
---|---|---|
LABEL_DETECTION_MODE_UNSPECIFIED |
number |
Unspecified. |
SHOT_MODE |
number |
Detect shot-level labels. |
FRAME_MODE |
number |
Detect frame-level labels. |
SHOT_AND_FRAME_MODE |
number |
Detect both shot-level and frame-level labels. |
(static, constant) LabelLevel :number
Label level (scope).
Properties:
Name | Type | Description |
---|---|---|
LABEL_LEVEL_UNSPECIFIED |
number |
Unspecified. |
VIDEO_LEVEL |
number |
Video-level. Corresponds to the whole video. |
SEGMENT_LEVEL |
number |
Segment-level. Corresponds to one of |
SHOT_LEVEL |
number |
Shot-level. Corresponds to a single shot (i.e. a series of frames without a major camera position or background change). |
FRAME_LEVEL |
number |
Frame-level. Corresponds to a single video frame. |
(static, constant) Likelihood :number
Bucketized representation of likelihood.
Properties:
Name | Type | Description |
---|---|---|
UNKNOWN |
number |
Unknown likelihood. |
VERY_UNLIKELY |
number |
Very unlikely. |
UNLIKELY |
number |
Unlikely. |
POSSIBLE |
number |
Possible. |
LIKELY |
number |
Likely. |
VERY_LIKELY |
number |
Very likely. |
Type Definitions
AnnotateVideoProgress
Video annotation progress. Included in the metadata
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
Properties:
Name | Type | Description |
---|---|---|
annotationProgress |
Array.<Object> |
Progress metadata for all videos specified in This object should have the same structure as VideoAnnotationProgress |
- Source:
- See:
AnnotateVideoRequest
Video annotation request.
Properties:
Name | Type | Description |
---|---|---|
inputUri |
string |
Input video location. Currently, only
Google Cloud Storage URIs are
supported, which must be specified in the following format:
|
inputContent |
string |
The video data bytes. Encoding: base64. If unset, the input video(s)
should be specified via |
features |
Array.<number> |
Requested video annotation features. The number should be among the values of Feature |
videoContext |
Object |
Additional video context and/or feature-specific parameters. This object should have the same structure as VideoContext |
outputUri |
string |
Optional location where the output (in JSON format) should be stored.
Currently, only Google Cloud Storage
URIs are supported, which must be specified in the following format:
|
locationId |
string |
Optional cloud region where annotation should take place. Supported cloud
regions: |
- Source:
- See:
AnnotateVideoResponse
Video annotation response. Included in the response
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
Properties:
Name | Type | Description |
---|---|---|
annotationResults |
Array.<Object> |
Annotation results for all videos specified in This object should have the same structure as VideoAnnotationResults |
- Source:
- See:
BoundingBox
Bounding box.
Properties:
Name | Type | Description |
---|---|---|
left |
number |
Left X coordinate. |
right |
number |
Right X coordinate. |
bottom |
number |
Bottom Y coordinate. |
top |
number |
Top Y coordinate. |
- Source:
- See:
FaceAnnotation
Face annotation.
Properties:
Name | Type | Description |
---|---|---|
thumbnail |
string |
Thumbnail of a representative face view (in JPEG format). Encoding: base64. |
segments |
Array.<Object> |
All locations where a face was detected. Faces are detected and tracked on a per-video basis (as opposed to across multiple videos). This object should have the same structure as VideoSegment |
locations |
Array.<Object> |
Face locations at one frame per second. This object should have the same structure as FaceLocation |
- Source:
- See:
FaceLocation
Face location.
Properties:
Name | Type | Description |
---|---|---|
boundingBox |
Object |
Bounding box in a frame. This object should have the same structure as BoundingBox |
timeOffset |
number |
Video time offset in microseconds. |
- Source:
- See:
LabelAnnotation
Label annotation.
Properties:
Name | Type | Description |
---|---|---|
description |
string |
Textual description, e.g. |
languageCode |
string |
Language code for |
locations |
Array.<Object> |
Where the label was detected and with what confidence. This object should have the same structure as LabelLocation |
- Source:
- See:
LabelLocation
Label location.
Properties:
Name | Type | Description |
---|---|---|
segment |
Object |
Video segment. Set to [-1, -1] for video-level labels.
Set to [timestamp, timestamp] for frame-level labels.
Otherwise, corresponds to one of This object should have the same structure as VideoSegment |
confidence |
number |
Confidence that the label is accurate. Range: [0, 1]. |
level |
number |
Label level. The number should be among the values of LabelLevel |
- Source:
- See:
SafeSearchAnnotation
Safe search annotation (based on per-frame visual signals only).
If no unsafe content has been detected in a frame, no annotations
are present for that frame. If only some types of unsafe content
have been detected in a frame, the likelihood is set to UNKNOWN
for all other types of unsafe content.
Properties:
Name | Type | Description |
---|---|---|
adult |
number |
Likelihood of adult content. The number should be among the values of Likelihood |
spoof |
number |
Likelihood that an obvious modification was made to the original version to make it appear funny or offensive. The number should be among the values of Likelihood |
medical |
number |
Likelihood of medical content. The number should be among the values of Likelihood |
violent |
number |
Likelihood of violent content. The number should be among the values of Likelihood |
racy |
number |
Likelihood of racy content. The number should be among the values of Likelihood |
timeOffset |
number |
Video time offset in microseconds. |
- Source:
- See:
VideoAnnotationProgress
Annotation progress for a single video.
Properties:
Name | Type | Description |
---|---|---|
inputUri |
string |
Video file location in Google Cloud Storage. |
progressPercent |
number |
Approximate percentage processed thus far. Guaranteed to be 100 when fully processed. |
startTime |
Object |
Time when the request was received. This object should have the same structure as Timestamp |
updateTime |
Object |
Time of the most recent update. This object should have the same structure as Timestamp |
- Source:
- See:
VideoAnnotationResults
Annotation results for a single video.
Properties:
Name | Type | Description |
---|---|---|
inputUri |
string |
Video file location in Google Cloud Storage. |
labelAnnotations |
Array.<Object> |
Label annotations. There is exactly one element for each unique label. This object should have the same structure as LabelAnnotation |
faceAnnotations |
Array.<Object> |
Face annotations. There is exactly one element for each unique face. This object should have the same structure as FaceAnnotation |
shotAnnotations |
Array.<Object> |
Shot annotations. Each shot is represented as a video segment. This object should have the same structure as VideoSegment |
safeSearchAnnotations |
Array.<Object> |
Safe search annotations. This object should have the same structure as SafeSearchAnnotation |
error |
Object |
If set, indicates an error. Note that for a single This object should have the same structure as Status |
- Source:
- See:
VideoContext
Video context and/or feature-specific parameters.
Properties:
Name | Type | Description |
---|---|---|
segments |
Array.<Object> |
Video segments to annotate. The segments may overlap and are not required to be contiguous or span the whole video. If unspecified, each video is treated as a single segment. This object should have the same structure as VideoSegment |
labelDetectionMode |
number |
If label detection has been requested, what labels should be detected
in addition to video-level labels or segment-level labels. If unspecified,
defaults to The number should be among the values of LabelDetectionMode |
stationaryCamera |
boolean |
Whether the video has been shot from a stationary (i.e. non-moving) camera. When set to true, might improve detection accuracy for moving objects. |
labelDetectionModel |
string |
Model to use for label detection. Supported values: "latest" and "stable" (the default). |
faceDetectionModel |
string |
Model to use for face detection. Supported values: "latest" and "stable" (the default). |
shotChangeDetectionModel |
string |
Model to use for shot change detection. Supported values: "latest" and "stable" (the default). |
safeSearchDetectionModel |
string |
Model to use for safe search detection. Supported values: "latest" and "stable" (the default). |
- Source:
- See:
VideoSegment
Video segment.
Properties:
Name | Type | Description |
---|---|---|
startTimeOffset |
number |
Start offset in microseconds (inclusive). Unset means 0. |
endTimeOffset |
number |
End offset in microseconds (inclusive). Unset means 0. |