google.cloud.videointelligence. v1beta1
Source: index.
Properties
Properties
Feature number
Video annotation feature.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
FEATURE_UNSPECIFIED |
|
|
Unspecified. |
|
LABEL_DETECTION |
|
|
Label detection. Detect objects, such as dog or flower. |
|
FACE_DETECTION |
|
|
Human face detection and tracking. |
|
SHOT_CHANGE_DETECTION |
|
|
Shot change detection. |
|
SAFE_SEARCH_DETECTION |
|
|
Safe search detection. |
LabelDetectionMode number
Label detection mode.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
LABEL_DETECTION_MODE_UNSPECIFIED |
|
|
Unspecified. |
|
SHOT_MODE |
|
|
Detect shot-level labels. |
|
FRAME_MODE |
|
|
Detect frame-level labels. |
|
SHOT_AND_FRAME_MODE |
|
|
Detect both shot-level and frame-level labels. |
LabelLevel number
Label level (scope).
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
LABEL_LEVEL_UNSPECIFIED |
|
|
Unspecified. |
|
VIDEO_LEVEL |
|
|
Video-level. Corresponds to the whole video. |
|
SEGMENT_LEVEL |
|
|
Segment-level. Corresponds to one of |
|
SHOT_LEVEL |
|
|
Shot-level. Corresponds to a single shot (i.e. a series of frames without a major camera position or background change). |
|
FRAME_LEVEL |
|
|
Frame-level. Corresponds to a single video frame. |
Likelihood number
Bucketized representation of likelihood.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
UNKNOWN |
|
|
Unknown likelihood. |
|
VERY_UNLIKELY |
|
|
Very unlikely. |
|
UNLIKELY |
|
|
Unlikely. |
|
POSSIBLE |
|
|
Possible. |
|
LIKELY |
|
|
Likely. |
|
VERY_LIKELY |
|
|
Very likely. |
Abstract types
AnnotateVideoProgress
Video annotation progress. Included in the metadata
field of the Operation returned by the GetOperation
call of the google::longrunning::Operations service.
Property
| Name | Type | Optional | Description |
|---|---|---|---|
|
annotationProgress |
Array of Object |
|
Progress metadata for all videos specified in This object should have the same structure as VideoAnnotationProgress |
AnnotateVideoRequest
Video annotation request.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
inputUri |
string |
|
Input video location. Currently, only
Google Cloud Storage URIs are
supported, which must be specified in the following format:
|
|
inputContent |
string |
|
The video data bytes. Encoding: base64. If unset, the input video(s)
should be specified via |
|
features |
Array of number |
|
Requested video annotation features. The number should be among the values of Feature |
|
videoContext |
Object |
|
Additional video context and/or feature-specific parameters. This object should have the same structure as VideoContext |
|
outputUri |
string |
|
Optional location where the output (in JSON format) should be stored.
Currently, only Google Cloud Storage
URIs are supported, which must be specified in the following format:
|
|
locationId |
string |
|
Optional cloud region where annotation should take place. Supported cloud
regions: |
AnnotateVideoResponse
Video annotation response. Included in the response
field of the Operation returned by the GetOperation
call of the google::longrunning::Operations service.
Property
| Name | Type | Optional | Description |
|---|---|---|---|
|
annotationResults |
Array of Object |
|
Annotation results for all videos specified in This object should have the same structure as VideoAnnotationResults |
BoundingBox
Bounding box.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
left |
number |
|
Left X coordinate. |
|
right |
number |
|
Right X coordinate. |
|
bottom |
number |
|
Bottom Y coordinate. |
|
top |
number |
|
Top Y coordinate. |
FaceAnnotation
Face annotation.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
thumbnail |
string |
|
Thumbnail of a representative face view (in JPEG format). Encoding: base64. |
|
segments |
Array of Object |
|
All locations where a face was detected. Faces are detected and tracked on a per-video basis (as opposed to across multiple videos). This object should have the same structure as VideoSegment |
|
locations |
Array of Object |
|
Face locations at one frame per second. This object should have the same structure as FaceLocation |
FaceLocation
Face location.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
boundingBox |
Object |
|
Bounding box in a frame. This object should have the same structure as BoundingBox |
|
timeOffset |
number |
|
Video time offset in microseconds. |
LabelAnnotation
Label annotation.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
description |
string |
|
Textual description, e.g. |
|
languageCode |
string |
|
Language code for |
|
locations |
Array of Object |
|
Where the label was detected and with what confidence. This object should have the same structure as LabelLocation |
LabelLocation
Label location.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
segment |
Object |
|
Video segment. Set to [-1, -1] for video-level labels.
Set to [timestamp, timestamp] for frame-level labels.
Otherwise, corresponds to one of This object should have the same structure as VideoSegment |
|
confidence |
number |
|
Confidence that the label is accurate. Range: [0, 1]. |
|
level |
number |
|
Label level. The number should be among the values of LabelLevel |
SafeSearchAnnotation
Safe search annotation (based on per-frame visual signals only).
If no unsafe content has been detected in a frame, no annotations
are present for that frame. If only some types of unsafe content
have been detected in a frame, the likelihood is set to UNKNOWN
for all other types of unsafe content.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
adult |
number |
|
Likelihood of adult content. The number should be among the values of Likelihood |
|
spoof |
number |
|
Likelihood that an obvious modification was made to the original version to make it appear funny or offensive. The number should be among the values of Likelihood |
|
medical |
number |
|
Likelihood of medical content. The number should be among the values of Likelihood |
|
violent |
number |
|
Likelihood of violent content. The number should be among the values of Likelihood |
|
racy |
number |
|
Likelihood of racy content. The number should be among the values of Likelihood |
|
timeOffset |
number |
|
Video time offset in microseconds. |
VideoAnnotationProgress
Annotation progress for a single video.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
inputUri |
string |
|
Video file location in Google Cloud Storage. |
|
progressPercent |
number |
|
Approximate percentage processed thus far. Guaranteed to be 100 when fully processed. |
|
startTime |
Object |
|
Time when the request was received. This object should have the same structure as Timestamp |
|
updateTime |
Object |
|
Time of the most recent update. This object should have the same structure as Timestamp |
VideoAnnotationResults
Annotation results for a single video.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
inputUri |
string |
|
Video file location in Google Cloud Storage. |
|
labelAnnotations |
Array of Object |
|
Label annotations. There is exactly one element for each unique label. This object should have the same structure as LabelAnnotation |
|
faceAnnotations |
Array of Object |
|
Face annotations. There is exactly one element for each unique face. This object should have the same structure as FaceAnnotation |
|
shotAnnotations |
Array of Object |
|
Shot annotations. Each shot is represented as a video segment. This object should have the same structure as VideoSegment |
|
safeSearchAnnotations |
Array of Object |
|
Safe search annotations. This object should have the same structure as SafeSearchAnnotation |
|
error |
Object |
|
If set, indicates an error. Note that for a single This object should have the same structure as Status |
VideoContext
Video context and/or feature-specific parameters.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
segments |
Array of Object |
|
Video segments to annotate. The segments may overlap and are not required to be contiguous or span the whole video. If unspecified, each video is treated as a single segment. This object should have the same structure as VideoSegment |
|
labelDetectionMode |
number |
|
If label detection has been requested, what labels should be detected
in addition to video-level labels or segment-level labels. If unspecified,
defaults to The number should be among the values of LabelDetectionMode |
|
stationaryCamera |
boolean |
|
Whether the video has been shot from a stationary (i.e. non-moving) camera. When set to true, might improve detection accuracy for moving objects. |
|
labelDetectionModel |
string |
|
Model to use for label detection. Supported values: "latest" and "stable" (the default). |
|
faceDetectionModel |
string |
|
Model to use for face detection. Supported values: "latest" and "stable" (the default). |
|
shotChangeDetectionModel |
string |
|
Model to use for shot change detection. Supported values: "latest" and "stable" (the default). |
|
safeSearchDetectionModel |
string |
|
Model to use for safe search detection. Supported values: "latest" and "stable" (the default). |
VideoSegment
Video segment.
Properties
| Name | Type | Optional | Description |
|---|---|---|---|
|
startTimeOffset |
number |
|
Start offset in microseconds (inclusive). Unset means 0. |
|
endTimeOffset |
number |
|
End offset in microseconds (inclusive). Unset means 0. |