Retrieve video information
This method retrieves information about the specified video.
Path parameters
The unique identifier of the index to which the video has been uploaded.
The unique identifier of the video to retrieve.
Headers
Query parameters
Specifies which types of embeddings to retrieve. You can include one or more of the following values:
visual-text
: Returns visual embeddings optimized for text search.audio
: Returns audio embeddings.
To retrieve embeddings for a video, it must be indexed using the Marengo video understanding model version 2.7 or later. For details on enabling this model for an index, see the Create an index page.
The platform does not return embeddings if you don’t provide this parameter.
The values you specify in embedding_option
must be included in the model_options
defined when the index was created. For example, if model_options
is set to visual,
you cannot set embedding_option
to audio
or both visual-text
and audio
.
The parameter indicates whether to retrieve a transcription of the spoken words for the indexed video. Note that the official SDKs will support this feature in a future release.
Response
The specified video information has successfully been retrieved.
The unique identifier of the video.
A string indicating the date and time, in the RFC 3339 format (“YYYY-MM-DDTHH:mm:ssZ”), that the video indexing task was created.
A string indicating the date and time, in the RFC 3339 format (“YYYY-MM-DDTHH:mm:ssZ”), that the corresponding video indexing task was last updated. The platform updates this field every time the corresponding video indexing task transitions to a different state.
A string indicating the date and time, in the RFC 3339 format (“YYYY-MM-DDTHH:mm:ssZ”), that the video indexing task has been completed.
System-generated metadata about the video.
User-generated metadata about the video.
The platform returns this object only for the videos that you uploaded with the enable_video_stream
parameter set to true
.
Contains the embedding and the associated information. The platform returns this field when the embedding_option
parameter is specified in the request.
An array of objects that contains the transcription. For each time range for which the platform finds spoken words, it returns an object that contains the fields below. If the platform doesn’t find any spoken words, the data
field is set to null
. Note that the official SDKs will support this feature in a future release.