Media and entertainment
The applications on this page demonstrate the capabilities of the TwelveLabs Video Understanding Platform in transforming how you interact with and consume digital content.
Ghool
Summary: Ghool is a trailer creation application that uses the TwelveLabs Video Understanding Platform to search, retrieve, and concatenate video clips based on specific prompts and quality metrics.
Description: The application uses both the Marengo and Pegasus video understanding models to process video content and create optimized video sequences with minimal user intervention. It retrieves relevant video clips based on user prompts. It then processes these videos by concatenating them and evaluating the transitions between different videos based on specific quality metrics, such as smoothness.
Ghool was created by Fazil Onuralp Adic and Oscar Chen and was one of the winners at the Cerebral Beach Hacks – LA Tech Week 2024 Kickoff Hackathon.
GitHub repo: cerebral_ghool
Integration with TwelveLabs
Ghool integrates with the TwelveLabs Video Understanding Platform to search, retrieve, and analyze video content based on specific prompts. The integration enables automated evaluation of video transitions and quality assessment without manual intervention.
The find_video
function uses the TwelveLabs Python SDK to perform text queries against a specified index. It filters results by confidence level and evaluates each clip’s quality against user-defined criteria.
The get_video
function retrieves the HLS streaming URL for a video by its ID and uses MoviePy to extract the relevant segment.
The get_comment
function utlizies the Generate API to evaluate video clips, producing numeric quality scores for criteria like “scariness” or “smoothness of transitions.”
The compare_transition
function creates and evaluates transitions between video clips by:
- Concatenating pairs of videos
- Uploading the concatenated videos to TwelveLabs
- Analyzing transition quality using the
get_comment
function described above - Ranking the transitions based on smoothness
AI Sports Recap
Summary: AI Sports Recap is a Streamlit-based application that generates video highlights and textual summaries of sports press conferences from YouTube video links. The application utilizes the Pegasus video understanding engine, GPT-4o, and Docker to provide an efficient and user-friendly experience.
Description: The application streamlines the process of extracting essential information from sports press conferences. Users can input a YouTube video link and a specific query, and the application will generate relevant video highlights and a concise textual summary. This application is particularly useful for sports enthusiasts, journalists, and analysts who need to extract and share important information from lengthy press conferences quickly.
The application was developed by Prateek Chhikara, Omkar Masur, Tanmay Rode, and it has won second place at the Multimodal AI Media & Entertainment Hackathon.
GitHub: sports-highlights.
Integration with TwelveLabs
The code below initializes the TwelveLabs Python SDK and creates a new index for which it enables the Marengo and Pegasus video understanding engines:
The upload_video
function uploads a video to YouTube:
The get_transcript
function generates the transcript for a specified video. For each segment, the function returns the text, start times, and end times in separate lists:
ThirteenLabs Smart AI Editor
Summary: ThirteenLabs Smart AI Editor uses artificial intelligence to process videos based on user prompts. It creates highlight reels, generates descriptions, and dubs in multiple languages.
Description: The application utilizes AI models from TwelveLabs, Gemini, and ElevenLabs to deliver high-quality video editing and transcription services. It offers a user-friendly interface built with Gradio.
The application was developed by Dylan Ler and it has won first place at the Multimodal AI Media & Entertainment Hackathon.
Integration with TwelveLabs
The get_transcript
function retrieves the transcript of a video using the TwelveLabs Python SDK. It returns the text, start time, and end time for each segment as a list of dictionaries.
Cactus
Summary: Cactus is a content generation application that uses functionalities of the TwelveLabs Video Understanding Platform to automatically transform long-form YouTube videos into engaging short-form reels. The application democratizes content creation, making it accessible and affordable for all creators, regardless of their resources.
Description: Cactus addresses the challenge of time-consuming video editing by bridging the gap between long-form and short-form content creation. The platform analyzes video content, identifies the most engaging moments, and compiles them into optimized highlight reels tailored for various social media platforms.
Cactus offers several key features and benefits for content creators:
- Saves time by automating the editing process, allowing creators to focus on content creation.
- Reduces costs by minimizing the need for professional editing services.
- Enhances reach by enabling the quick production of more content.
- Content creators can expand their audience across multiple platforms.
- The application was developed by Saurabh Ghanekar, Noah Bergren, Christopher Kinoshita, and Shrutika Nikola.
GitHub: Cactus
Integration with TwelveLabs
The generate_segment_itinerary
function invokes the POST
method of the /generate
endpoint to segment a video and create an itinerary based on activities and locations shown in the video. The function returns the segmented itinerary if successful or logs any errors encountered.
Hello Garfield
Summary: The “Hello Garfield” application provides an immersive virtual reality experience combining traditional movie theaters with cutting-edge technology. It features a personalized AI concierge, themed environments, and interactive elements to enhance the movie-watching experience.
Description: The application transforms how you engage with movies in a virtual space. Upon entering the virtual theater, you are greeted by an AI concierge. This concierge offers personalized movie recommendations based on your preferences and viewing history.
Key features include:
- Video Q&A chatbot: The application uses the Generate API to allow you to ask questions about the movies you’re watching.
- Immersive VR/MR environment: A realistic virtual movie theater with a large screen, created using VR/MR development platforms such as Unity and Unreal Engine.
- AI concierge: A chatbot that provides personalized movie suggestions and enhances the user experience through friendly interaction.
- Enhanced viewing experience: The concierge suggests themed snacks, recipes, and merchandise related to the chosen movie, creating a more immersive and enjoyable viewing experience.
- AR filters: You can “try on” costumes from your favorite films and decorate your virtual spaces using augmented reality technology.
- Community interaction: A shared virtual theater space that allows you to connect with other film enthusiasts, fostering a sense of community.
The application was developed by Lauren Descher, Dulce Baerga, and Catherine Rhee.
GitHub: aila-hack
Integration with TwelveLabs
The code below handles different types of requests:
- It checks the type of request and prepares appropriate data for each.
- For each type of request, it constructs a data object with a specific video ID and other relevant parameters.
- It sends a
POST
request to the /gist or /summarize endpoints.
Sports Recap
Summary: Sports Recap is a NextJS-based application that generates video highlights and summaries from sports press conferences.
Description: The application is designed to transform lengthy sports press conferences into concise, engaging highlight reels. The application utilizes the Generate API to create relevant highlights based on user-specified criteria.
The application was developed by Daniel Jacobs, Yurko Turskiy, Suxu Li, and Melissa Regan.
GitHub: hackathone-challenge-3
Integration with TwelveLabs
The function below extracts data from the incoming request and invokes the POST
method of the /summarize
endpoint, passing the unique identifier of a video and a prompt to generate highlights:
The function below extracts the unique identifier of a video from the incoming request and invokes the POST
method of the /summarize
endpoint. It passes the video ID along with a predefined prompt to generate a summary that identifies the main character and lists key topics discussed: