This page provides an overview of common workflows for interacting with the Twelve Labs Video Understanding Platform using an HTTP client. Each workflow consists of a series of steps, with links to detailed documentation for each step.
All workflows involving uploading video content to the platform require asynchronous processing. You must wait for the video processing to complete before proceeding with the subsequent steps.
Prerequisites
- To use the platform, you need an API key.
Search
Follow the steps in this section to search through your video content and find specific moments, scenes, or information.
Steps:
- Create an index, enabling the Marengo video understanding model.
- Upload videos and monitor the processing.
- Perform a search request, using text or images as queries.
Notes:
- The search scope is an individual index.
- Results support pagination, filtering, sorting, and grouping.
For an interactive implementation using the Python SDK, see the Quickstart Search quickstart notebook.
Generate text from videos
Follow the steps in this section to generate texts based on your videos.
Steps:
- Create an index, enabling the Pegasus video understanding model.
- Upload videos and monitor the processing.
- Depending on your use case, generate one of the following:
For an interactive implementation using the Python SDK, see the Quickstart Generate quickstart notebook.
Create text, image, and audio embeddings
This workflow guides you through creating embeddings for text.
Steps:
Note:
- Creating text, image, and audio embeddings is a synchronous process.
Create video embeddings
This workflow guides you through creating embeddings for videos.
Steps:
Note:
- Creating video embeddings is a synchronous process.
For an interactive implementation using the Python SDK, see the Quickstart Embed quickstart notebook.