Image embeddings
This guide shows how you can create image embeddings using the Marengo video understanding model. For a list of available versions, complete specifications and input requirements for each version, see the Marengo page.
The Marengo video understanding model generates embeddings for all modalities in the same latent space. This shared space enables any-to-any searches across different types of content.
Prerequisites
-
To use the platform, you need an API key:
-
Depending on the programming language you are using, install the TwelveLabs SDK by entering one of the following commands:
-
Your image files must meet the format requirements.
Complete example
This complete example shows how you can create image embeddings. Ensure you replace the placeholders surrounded by <> with your values.
Step-by-step guide
Python
Node.js
Import the SDK and initialize the client
Create a client instance to interact with the TwelveLabs Video Understanding Platform.
Function call: You call the constructor of the TwelveLabs class.
Parameters:
api_key: The API key to authenticate your requests to the platform.
Return value: An object of type TwelveLabs configured for making API calls.
Create image embeddings
Function call: You call the embed.create function.
Parameters:
model_name: The name of the model you want to use (example: “marengo3.0”).image_urlorimage_file:image_url: The publicly accessible URL of your image file (string)image_file: An opened file object in binary read mode. Useopen(path, 'rb')to open your local file
Return value: The response contains the following fields:
image_embedding: An object that contains the embedding data for your image file. It includes the following fields:segments: An object that contains the following:float_: An array of floats representing the embedding
metadata: An object that contains metadata about the embedding.
model_name: The name of the video understanding model the platform has used to create this embedding.