Author
James Le
Date Published
March 19, 2024
Tags
Generate API
API Tutorial
Database
Partnership
Share
Join our newsletter
You’re now subscribed to the Twelve Labs Newsletter! You'll be getting the latest news and updates in video understanding.
Oh no, something went wrong.
Please try again.
This tutorial is co-authored with the wonderful team at MindsDB: Zoran Pandovski (Senior Full Stack Engineer), Martyna Slawinska (Software Engineer and Technical Writer), and Minura Punchihewa (Data Integrations Engineer)!

‍

Introduction

In recent years, there have been incredible advancements in machine learning models. However, integrating and serving these models into applications can still be challenging for developers. We're excited to share a tutorial showcasing how you can make building powerful ML apps far easier by combining Twelve Labs' state-of-the-art foundation models for video understanding with MindsDB's platform for building custom AI. Video is everywhere on the internet, but unlike text, it's harder to parse, summarize, and understand. In the tutorial below, you can build a tool that lets you do all these things.

Twelve Labs offers access to powerful multimodal foundation models that can understand natural language prompts and generate human-like text from video content. MindsDB, enables access to machine learning models and data sources for building customized AI solutions. Combining these two technologies opens up a world of possibilities for developers.

Whether you want to generate text, classify your content, or search for specific moments in your videos, this integration makes Twelve Labs' capabilities accessible to any developer. We hope you'll find this tutorial informative and helpful in building powerful ML apps.

Throughout this tutorial, you'll learn how to configure the Twelve Labs integration in MindsDB, deploy the Twelve Labs model within MindsDB (it will use the Twelve Labs summarization tasks), and automate the whole flow through a Slack bot that will periodically post the video summarizations as announcements.

By the time you complete this tutorial, you'll have created a Slack bot that leverages the strengths of MindsDB and Twelve Labs to deliver concise video summaries posted as announcements on Slack channels.

Let's get started and explore the limitless possibilities that this integration has to offer.

‍

Prerequisites

‍

Install MindsDB

To run the MindsDB Docker container, execute the following command:

docker run -p 47334:47334 -p 47335:47335 --name mdb mindsdb/mindsdb

While the MindsDB container is running, access its shell by executing:

docker exec -it mdb sh

Inside the Docker container, install the additional dependencies for TwelveLabs and Slack integrations using the following command:

pip install mindsdb[twelve_labs] mindsdb[slack]

After installation, you can access MindsDB's web interface at http://127.0.0.1:47334 using your browser.

‍

Get a Twelve Labs API Key

To retrieve the Twelve Labs API key, log in to the Dashboard page, locate your API key, and select the Copy icon to copy it to your clipboard.

Twelve Labs recommends using environment variables to pass configuration to your application. However, for this tutorial, you can provide your API key as a parameter to MindsDB, as seen below, while creating an ML engine.

‍

Deploy and Use Twelve Labs Models from MindsDB

Configure the Twelve Labs engine

Before diving into the summarization task, you need to establish a connection between MindsDB and Twelve Labs. This involves creating a Machine Learning (ML) Engine that utilizes the Twelve Labs API. To initiate this integration, copy your API key and execute the following SQL command to create the Twelve Labs engine:

CREATE ML_ENGINE twelve_labs_engine
FROM twelve_labs
USING
 twelve_labs_api_key = '<YOUR_API_KEY>';

‍

Deploy the Twelve Labs model

With the ML engine ready, the next phase involves creating a model to handle the summarization task:

CREATE MODEL mindsdb.demo_day_model
PREDICT result_summary
USING
 engine = 'twelve_labs_engine',
 task = "summarization",
 engine_id = 'pegasus1',
 index_name = 'index_1',
 index_options = ['visual', 'conversation'],
 video_urls = ['https://mindsdb.s3.eu-north-1.amazonaws.com/demo_day.mp4'],
 summarization_type = 'summary',
 prompt = 'Explore MindsDBs latest updates! Discover innovative features enhancing your analytics experience. Understand how these improvements bring precision and efficiency to your work, reflecting our commitment to forefront technology. Learn about integrating these new capabilities seamlessly.'

The parameters that we have provided in the USING statement are:

  • engine: The name of the ML Engine to use. In this case, it is the 'twelve_labs_engine' that you created before.
  • task: The name of the task to be performed by the model.
  • engine_id: The ID of the Twelve Labs engine to use. Note you need to provide 'pegasus1' for the summarization task.
  • index_name: The name of the index to use. If the index doesn’t exist it will be automatically created.
  • index_options: A list of the types of information within the video that will be processed. Note the 'pegasus1' engine only supports 'visual' and 'conversation'.
  • video_files: A list of local paths to videos
  • summarization_type: The type of summarization to perform in this case, 'summary'.
  • prompt: Provides context for the summarization task and changes the tone of the summary to be in a style of an announcement.

Note, depending on the video size, it may take some time to create the index and store the video data (video embeddings and metadata). Once the above command is executed, you can run the describe statement to check the model status by executing:

DESCRIBE mindsdb.demo_day_model

If everything works, you should see status as complete in the STATUS column. In case of an error, check the ERROR column, which contains detailed information about the error.

Before we generate a summarization from the video, we need to know the video_id. To get that, run the DESCRIBE statement for models indexed_videos:

DESCRIBE mindsdb.demo_day_model.indexed_videos;

The returned result set should contain the video_id. Copy the id and query the model to get the key points for the video.

‍

Get summaries generated by the Twelve Labs model

With the video ID, you can now extract the summarization:

SELECT * FROM mindsdb.demo_day_model
WHERE video_id = '65e88e6f48db9fa780cb466f'

Now, alongside result_id and video_id, you should be able to see the result_summary column which contains the actual summarization text.

How cool is this? With just a few SQL statements, we were able to get a clear summary of the video. Let’s take this even forward and create a Slack Bot that will post the summaries in the Slack channel.

‍

SlackConnect to Slack

Before connecting MindsDB to your Slack workspace, make sure you generate a token by following up on the MindsDB Slack documentation.

Once you have the token run the following command:

CREATE DATABASE demo_day_bot
WITH
ENGINE = 'slack',
PARAMETERS =  {
    "token": "xoxb-2347-21112-csuae1231" 
};

Then, verify the integration by posting a welcome message to your Slack channel:

INSERT INTO demo_day_bot.channels (channel, text)
VALUES("announcements, "Welcome from MindsDB and Twelve Labs");

Now, on the announcement channel you should be able to see the above message.

If that works, you can use the Twelve Labs model and post the summarizations on the channel:

INSERT INTO demo_day_bot.channels (channel, text)
SELECT "announcements" as channel, result_summary as text 
FROM mindsdb.demo_day_model 
WHERE video_id='65e7396748db9fa780cb4575'

This query will call the demo_day_model, get the summarization text, and use the demo_day_bot to post it on the announcement channel.

‍

Conclusion

Congratulations! By completing this tutorial, you have successfully created a Slack bot that harnesses the power of MindsDB and Twelve Labs to deliver concise video summaries as announcements on Slack channels.

You learned how to:

  1. Install and configure MindsDB
  2. Obtain a Twelve Labs API key
  3. Configure the Twelve Labs engine in MindsDB
  4. Deploy the Twelve Labs summarization model in MindsDB
  5. Generate video summaries using the Twelve Labs model via SQL queries
  6. Connect the summarization workflow to Slack for automated posting

The complete GitHub repository can be found here: https://github.com/mindsdb/mindsdb/tree/staging/mindsdb/integrations/handlers/twelve_labs_handler.

This integration showcases the ease of unlocking cutting-edge foundation models from Twelve Labs within your applications by leveraging MindsDB's AI-SQL capabilities. With just a few lines of SQL, you could access state-of-the-art video understanding and seamlessly incorporate it into a useful Slack bot.

To further explore the potential of MindsDB and Twelve Labs, check out these additional resources:

We encourage you to experiment with other Twelve Labs models and MindsDB integrations to build even more powerful video AI applications. The possibilities are endless!

If you have any questions or feedback, feel free to reach out to the MindsDB and Twelve Labs communities. Happy building!

Generation Examples
No items found.
No items found.
Comparison against existing models
No items found.

Related articles

Semantic Video Search Engine with Twelve Labs and ApertureDB

Learn how to build a semantic video search engine with the powerful integration of Twelve Labs' Embed API with ApertureDB for advanced semantic video search.

James Le
Building a Shade Finder App: Using Twelve Labs' API to Pinpoint Specific Colors in Videos

Whether you're looking to find the perfect berry-toned lipstick or just curious about spotting specific colors in your videos, this guide will help you leverage cutting-edge AI to do so effortlessly.

Meeran Kim
Building Advanced Video Understanding Applications: Integrating Twelve Labs Embed API with LanceDB for Multimodal AI

Leverage Twelve Labs Embed API and LanceDB to create AI applications that can process and analyze video content with unprecedented accuracy and efficiency.

James Le, Manish Maheshwari
Advanced Video Search: Leveraging Twelve Labs and Milvus for Semantic Retrieval

Harness the power of Twelve Labs' advanced multimodal embeddings and Milvus' efficient vector database to create a robust video search solution.

James Le, Manish Maheshwari