Skip to main content

SDK for Twelve Labs API

Project description

TwelveLabs Python SDK

PyPI version

This SDK provides a convenient way to interact with the Twelve Labs Video Understanding Platform from an application written in the Python language. The SDK equips you with a set of intuitive classes and methods that streamline the process of interacting with the platform, minimizing the need for boilerplate code.

Prerequisites

Ensure that the following prerequisites are met before using the SDK:

  • Python 3.7 or newer must be installed on your machine.
  • You have an API key. If you don't have an account, please sign up for a free account. Then, to retrieve your API key, go to the Dashboard page, and select the Copy icon to the right of the key to copy it to your clipboard.

Install the SDK

Install the twelvelabs package:

pip install twelvelabs

Initialize the SDK

  1. Import the SDK into your application:

    from twelvelabs import TwelveLabs
    
  2. Instantiate the SDK client with your API key. This example code assumes that your API key is stored in an environment variable named TL_API_KEY:

    client = TwelveLabs(api_key=os.getenv('TL_API_KEY'))
    

Use the SDK

To get started with the SDK, follow these basic steps:

  1. Create an index.
  2. Upload videos.
  3. Perform downstream tasks, such as searching or generating text from video.

Create an index

To create an index, use the example code below, replacing "<YOUR_INDEX_NAME>" with the desired name for your index:

from twelvelabs import APIStatusError

index_obj = None
try:
    index_obj = client.index.create(
        name = "<YOUR_INDEX_NAME>",
        engines =[
            {
                "name": "marengo2.6",
                "options": ["visual", "conversation", "text_in_video"],
            },
            {
                "name": "pegasus1",
                "options": ["visual", "conversation"],
            },
        ],
    )
    print(index_obj)
except APIStatusError as e:
    print('API Status Error, 4xx or 5xx')
    print(e)
except Exception as e:
    print(e)

Note the following about this example:

  • The platform provides two distinct engine types - embedding and generative, each serving unique purposes in multimodal video understanding.
    • Embedding engines (Marengo): These engines are proficient at performing tasks such as search and classification, enabling enhanced video understanding.
    • Generative engines (Pegasus): These engines generate text based on your videos. For your index, both Marengo and Pegasus are enabled.
  • The engines.options fields specify the types of information each video understanding engine will process.
  • The engines and the engine options specified when you create an index apply to all the videos you upload to that index and cannot be changed. For details, see the Engine options page.

The output should look similar to the following:

Index(id='65b1b926560f741da96836d7', created_at='2024-01-25T01:28:06.061Z', updated_at='2024-01-25T01:28:06.061Z', name='test-index-to-researchers1', engines=[Engine(name='marengo2.6', options=['visual', 'conversation', 'text_in_video'], addons=None), Engine(name='pegasus1', options=['visual', 'conversation'], addons=None)], video_count=0, total_duration=0.0, expires_at='2024-04-24T01:28:06.061Z')

Note that the API returns, among other information, a field named id, representing the unique identifier of your new index.

For a description of each field in the request and response, see the Create an index page.

Upload videos

Before you upload a video to the platform, ensure that it meets the following requirements:

  • Video resolution: Must be greater or equal than 360p and less or equal than 4K. For consistent search results, Twelve Labs recommends you upload 360p videos.
  • Video and audio formats: The video files you wish to upload must be encoded in the video and audio formats listed on the FFmpeg Formats Documentation page. For videos in other formats, contact us at support@twelvelabs.io.
  • Duration: For Marengo, it must be between 4 seconds and 2 hours (7,200s). For Pegasus, it must be between 4 seconds and 20 minutes (1200s).
  • File size: Must not exceed 2 GB. If you require different options, send us an email at support@twelvelabs.io.
  • Audio track: If the conversation engine option is selected, the video you're uploading must contain an audio track.

To upload videos, use the example code below, replacing the following:

  • <YOUR_VIDEO_PATH>: with a string representing the path to the directory containing the video files you wish to upload.
  • <YOUR_INDEX_ID>: with a string representing the unique identifier of the index to which you want to upload your video.
from glob import glob
from twelvelabs.models.task import Task

video_files = glob("<YOUR_VIDEO_PATH>") # Example: "/videos/*.mp4
for video_file in video_files:
  print(f"Uploading {video_file}")
  task = client.task.create(index_id="<YOUR_INDEX_ID>", file=video_file, language="en")
  print(f"Task id={task.id}")

  # (Optional) Monitor the video indexing process
  # Utility function to print the status of a video indexing task
  def on_task_update(task: Task):
          print(f"  Status={task.status}")
  task.wait_for_done(callback=on_task_update)
  if task.status != "ready":
      raise RuntimeError(f"Indexing failed with status {task.status}")
  print(f"Uploaded {video_file}. The unique identifer of your video is {task.video_id}.")

Note that once a video has been successfully uploaded and indexed, the response will contain a field named video_id, representing the unique identifier of your video.

For a description of each field in the request and response, see the Create a video indexing task page.

Perform downstream tasks

The sections below show how you can perform the most common downstream tasks. See our documentation for a complete list of all the features the Twelve Labs Understanding Platform provides.

Search

To perform a search request, use the example code below, replacing the following:

  • <YOUR_INDEX_ID>: with a string representing the unique identifier of your index.
  • <YOUR_QUERY>: with a string representing your search query. Note that the API supports full natural language-based search. The following examples are valid queries: "birds flying near a castle," "sun shining on water," and "an officer holding a child's hand."
  • [<YOUR_SEARCH_OPTIONS>]: with an array of strings that specifies the sources of information the platform uses when performing a search. For example, to search based on visual and conversation cues, use ["visual", "conversation"]. Note that the search options you specify must be a subset of the engine options used when you created the index. For more details, see the Search options page.
search_results = client.search.query("<YOUR_INDEX_ID>", "<YOUR_QUERY>", ["<YOUR_SEARCH_OPTIONS>"])

# Utility function to print a specific page
def print_page(page):
  for clip in page:
    print(
        f" video_id={clip.video_id} score={clip.score} start={clip.start} end={clip.end} confidence={clip.confidence}"
    )

print_page(search_results.data)

while True:
    try:
        print_page(next(search_results))
    except StopIteration:
        break

The results are returned one page at a time, with a default limit of 10 results on each page. The next method returns the next page of results. When you've reached the end of the dataset, a StopIteration exception is raised.

 video_id=65ca2bce48db9fa780cb3fa4 score=84.9 start=104.9375 end=111.90625 confidence=high
 video_id=65ca2bce48db9fa780cb3fa4 score=84.82 start=160.46875 end=172.75 confidence=high
 video_id=65ca2bce48db9fa780cb3fa4 score=84.77 start=55.375 end=72.46875 confidence=high

Note that the response contains, among other information, the following fields:

  • video_id: The unique identifier of the video that matched your search terms.
  • score: A quantitative value determined by the AI engine representing the level of confidence that the results match your search terms.
  • start: The start time of the matching video clip, expressed in seconds.
  • end: The end time of the matching video clip, expressed in seconds.
  • confidence: A qualitative indicator based on the value of the score field. This field can take one of the following values:
    • high
    • medium
    • low
    • extremely low

For a description of each field in the request and response, see the Make a search request page.

Generate text from video

The Twelve Labs Video Understanding Platform offers three distinct endpoints tailored to meet various requirements. Each endpoint has been designed with specific levels of flexibility and customization to accommodate different needs.

Note the following about using these endpoints:

  • The Pegasus video understanding engine must be enabled for the index to which your video has been uploaded.
  • Your prompts must be instructive or descriptive, and you should not phrase them as questions.
  • The maximum length of a prompt is 300 characters.

Topics, titles, and hashtags

To generate topics, titles, and hashtags, use the example code below, replacing the following:

  • <YOUR_VIDEO_ID>: with a string representing the unique identifier of your video.
  • [<TYPES>]: with an array of strings representing the type of text the platform should generate. Example: ["title", "topic", "hashtag"].
res = client.generate.gist("<YOUR_VIDEO_ID>", types=["<TYPES>"])
print(f"Title = {res.title}\nTopics = {res.topics}\nHashtags = {res.hashtags}")

For a description of each field in the request and response, see the Titles, topics, or hashtags page.

Summaries, chapters, and highlights

To generate summaries, chapters, and highlights, use the example code below, replacing the following:

  • <YOUR_VIDEO_ID>: with a string representing the unique identifier of your video.
  • <TYPE>: with a string representing the type of text the platform should generate. This parameter can take one of the following values: "summary", "chapter", or "highlight".
  • (Optional) <YOUR_PROMPT>: with a string that provides context for the summarization task, such as the target audience, style, tone of voice, and purpose. Example: "Generate a summary in no more than 5 bullet points."
res = client.generate.summarize("<YOUR_VIDEO_ID>", type="<TYPE>", prompt="<YOUR_PROMPT>")
print(f"{res.summary}")

For a description of each field in the request and response, see the Summaries, chapters, or highlights page.

Open-ended texts

To generate open-ended texts, use the example code below, replacing the following:

  • <YOUR_VIDEO_ID>: with a string representing the unique identifier of your video.
  • <YOUR_PROMPT>: with a string that guides the model on the desired format or content. The maximum length of the prompt is 500 tokens or roughly 350 words. Example: "I want to generate a description for my video with the following format: Title of the video, followed by a summary in 2-3 sentences, highlighting the main topic, key events, and concluding remarks."
res = client.generate.text(video_id="<YOUR_VIDEO_ID>", prompt="<YOUR_PROMPT>")
print(f"{res.data}")

Error Handling

The SDK includes a set of exceptions that are mapped to specific HTTP status codes, as shown in the table below:

Exception HTTP Status Code
BadRequestError 400
AuthenticationError 401
PermissionDeniedError 403
NotFoundError 404
ConflictError 409
UnprocessableEntityError 422
RateLimitError 429
InternalServerError 5xx

The following example shows how you can handle specific HTTP errors in your application:

import os
from twelvelabs import TwelveLabs

client = TwelveLabs(os.getenv("TWELVELABS_API_KEY"))
try:
    engines = client.engines.list()
    print(engines)
except twelvelabs.APIConnectionError as e:
    print("Cannot connect to API server")
except twelvelabs.BadRequestError as e:
    print("Bad request.")
except twelvelabs.APIStatusError as e:
    print(f"Status code {e.status_code} received")
    print(e.response)

License

We use the Developer Certificate of Origin (DCO) in lieu of a Contributor License Agreement for all contributions to Twelve Labs' open-source projects. We request that contributors agree to the terms of the DCO and indicate that agreement by signing all commits made to Twelve Labs' projects by adding a line with your name and email address to every Git commit message contributed, as shown in the example below:

Signed-off-by: Jane Doe <jane.doe@example.com>

You can sign your commit automatically with Git by using git commit -s if you have your user.name and user.email set as part of your Git configuration. We ask that you use your real name (please, no anonymous contributions or pseudonyms). By signing your commitment, you are certifying that you have the right have the right to submit it under the open-source license used by that particular project. You must use your real name (no pseudonyms or anonymous contributions are allowed.) We use the Probot DCO GitHub app to check for DCO signoffs of every commit. If you forget to sign your commits, the DCO bot will remind you and give you detailed instructions for how to amend your commits to add a signature.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

twelvelabs-0.2.4.tar.gz (27.6 kB view hashes)

Uploaded Source

Built Distribution

twelvelabs-0.2.4-py3-none-any.whl (32.4 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page