Video understanding engines

Twelve Labs' video understanding engines consist of a family of deep neural networks built on our multimodal foundation model for video understanding that you can use for the following downstream tasks:

  • Search using natural language queries
  • Zero-shot classification
  • Generate text from video.

Videos contain multiple types of information, including visuals, sounds, spoken words, and texts. The human brain combines all types of information and their relations with each other to comprehend the overall meaning of a scene. For example, you're watching a video of a person jumping and clapping, both visual cues, but the sound is muted. You might realize they're happy, but you can't understand why they're happy without the sound. However, if the sound is unmuted, you could realize they're cheering for a soccer team that scored a goal.

Thus, an application that analyzes a single type of information can't provide a comprehensive understanding of a video. Twelve Labs' video understanding engines, however, analyze and combine information from all the modalities to accurately interpret the meaning of a video holistically, similar to how humans watch, listen, and read simultaneously to understand videos.

Our video understanding engines have the ability to identify, analyze, and interpret a variety of elements, including but not limited to the following:

ElementModalityExample
People, including famous individualsVisualMichael Jordan, Steve Jobs
ActionsVisualRunning, dancing, kickboxing
ObjectsVisualCars, computers, stadiums
Animals or petsVisualMonkeys, cats, horses
NatureVisualMountains, lakes, forests
Sounds (excluding human speech)VisualChirping (birds), applause, fireworks popping or exploding
Human speechConversation"Good morning. How may I help you?"
Text displayed on the screen (OCR)Text in videoLicense plates, handwritten words, number on a player's jersey
Brand logosLogoNike, Starbucks, Mercedes

Engine Types

Twelve Labs provides two distinct engine types - embedding and generative, each serving unique purposes in multimodal video understanding.

  • Embedding engines (Marengo) : These engines are proficient at performing tasks such as search and classification, enabling enhanced video understanding.
  • Generative engines (Pegasus): These engines generate text based on your videos.

The following engines are available:

NameFeaturesDescription
Marengo2.6Search and classificationThis version of the Marengo video understanding engine provides the following main features:

- Expanded Multimodal Capabilities: Marengo supports any-to-any retrieval tasks, including text-to-video, text-to-image, text-to-audio, audio-to-video, and image-to-video. However, note that the platform currently supports text-to-video search and classification features. Other modalities will be supported in a future release.
- Enhanced Temporal Localization: By introducing a Reranker model, Marengo allows for precise search results through better temporal localization.
Pegasus1.1Video-to-text generation The 1.1 version of the Pegasus video understanding engine provides the following enhancements compared to 1.0:

- Improved model accuracy: Enhanced video description and question-answering capabilities, delivering more precise and relevant results.
- Fine-grained visual understanding and instruction following: Improved ability to analyze and interpret visual content, enabling more detailed insights.
- Streaming support: The ability to generate open-ended texts with real-time access to each token as it's generated allows for faster processing and utilization of generated content. For details, refer to the Streaming responses page.
- Increased maximum prompt length: Expanded from 300 to 1500 characters, allowing for the integration of more complex business logic and detailed examples in prompts.
- Extended video duration: The maximum duration of the videos you can upload has been increased from 20 to 30 minutes, allowing for analysis of longer content.
Marengo-retrieval-2.6EmbeddingsThis version of the Marengo video understanding engine creates embeddings that you can use in various downstream tasks. For details, see the Create embeddings section.

The following engines are no longer supported:

NameFeaturesNotes
Pegasus1.0Video-to-text generationEffective July 8, 2024, Pegasus 1.0 is no longer supported. All existing indexes created with Pegasus 1.0 will be automatically upgraded to Pegasus 1.1. No manual intervention is required for this migration process, and all indexes will utilize Pegasus 1.1 upon completion.