Twelve Labs Press kit

Everything you need, all in one place.
Stay in the loop with Twelve Labs.

DESCRIPTIONS

About Twelve Labs

Here are some descriptions about us.

Feel free to use them as reference.
Summary

Twelve Labs develops cutting-edge multimodal foundation models that enable human-like understanding of video content.

Our APIs support features like semantic search, video summarization, and content analysis, empowering developers and enterprises to effectively and securely leverage video data for a wide range of use cases across industries.

Boilerplate

Twelve Labs delivers industry-leading video Al solutions that unlock the full potential of enterprise vast video archives. Our proprietary multimodal foundation models bring human-like understanding to videos, enabling precise semantic search, summarization, analysis and Q&A through easy-to-integrate APls.


This empowers enterprises to effortlessly search, monetize extensive video libraries, extract insights, and repurpose content at scale. Unlike conventional methods that struggle with the complexities of video, Twelve Labs overcomes the limitations of manual tagging and inadequate computer vision techniques, streamlining processes with state-of-the-art, customizable models.

These models make previously inaccessible video assets searchable, seamlessly integrating into existing workflows. Media leaders like sports organizations, studios and creators rely on Twelve Labs to transform their video content.

View more
Product overview

Twelve Labs multimodal foundation models generate powerful vector embeddings that enable a wide range of downstream applications. Our Marengo model natively understands video, identifying and interpreting movements, actions, objects, individuals, sounds, on-screen text, and spoken words with human-like accuracy, facilitating high-precision semantic search.


Pegasus, our state-of-the-art video-to-text generation model supports a variety of use cases across industries. Built by developers, for developers, our APls provide access to these advanced multimodal foundation models, enabling capabilities such as:

  • Powerful semantic search: Find exact moments within any video using natural language queries, without the need for tags or metadata.
  • Video-to-text generation: Generate deep analyses, video specific Q&A, or general highlight generation for any video content.
  • Zero-shot classification: Utilize natural language to create your custom taxonomies, allowing for precise and efficient video classification tailored to your unique use case.
  • Intuitive integration: Embed our video understanding models into your application with just a few API calls.
  • Rapid result retrieval: Obtain results within seconds.
  • Scalability: Our cloud-native distributed infrastructure effortlessly handles thousands of concurrent requests.
View more
STATS

Backed by numbers

0
Twelve Labs was founded

2
Offices.
HQ: San Francisco

APAC: Seoul
0
Employees.
75% are researchers and engineers
0
+
Developers in 92+ countries 
are currently using our product
$
0
M
Capital raised

1M+
Monthly API calls


Fact sheet

Series A

  • Leads: NEA & NVentures
  • Follow-ons: Index Ventures, Radical Ventures, WndrCo, Ethos, KIP
  • Read more

Cloud provider

We’re cloud agnostic: We work with GCP, Azure, AWS, and OCI

Fact sheet

  • Media asset management
    • Archival management
    • Production workflows
  • Content licensing
    • Brand management
    • Product placement
  • Semantic search for evidence management
  • Evidence report generation

Key verticals

  • Media and Entertainment
  • Security and Surveillance
  • Automotive
  • Medical