Skip to main content

Assistant plugin for Pinecone SDK

Project description

Assistant

Interact with Pinecone's Assistant APIs, e.g. create, manage, and chat with assistants (currently in beta). Pinecone Assistant is also available in the console.

Quickstart

The following example highlights how to use an assistant to store and understand documents on a particular topic and chat with the assistant about those documents with the ultimate goal of semantically understanding your data.

from pinecone import Pinecone
from pinecone_plugins.assistant.models.chat import Message

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

# Create an assistant (in this case we'll store documents about planets)
space_assistant = pc.assistant.create_assistant(assistant_name="space")

# Upload information to your assistant
space_assistant.upload_file("./space-fun-facts.pdf")

# Once the upload succeeded, ask the assistant a question
msg = Message(content="How old is the earth?")
resp = space_assistant.chat_completions(messages=[msg])
print(resp)

# {'choices': [{'finish_reason': 'stop',
# 'index': 0,
# 'message': {'content': 'The age of the Earth is estimated to be '
#                         'about 4.54 billion years, based on '
#                         'evidence from radiometric age dating of '
#                         'meteorite material and Earth rocks, as '
#                         'well as lunar samples. This estimate has '
#                         'a margin of error of about 1%.',
#             'role': 'assistant'}}],
# 'id': '00000000000000001a377ceeaabf3c18',

Assistants API

Create Assistant

To create an assistant, see the below example. This API creates a assistant with the specified name, metadata, and optional timeout settings.

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
metadata = {"author": "Jane Doe", "version": "1.0"}

assistant = pc.assistant.create_assistant(
    assistant_name="example_assistant", 
    instructions="Always use British English spelling and vocabulary.",
    metadata=metadata,
    timeout=30
)

Arguments:

  • assistant_name The name to assign to the assistant.
    • type: str
  • instructions Custom instructions for the assistant. These will be applied to all future chat interactions.
    • type: Optional[str] = None
  • metadata: A dictionary containing metadata for the assistant.
    • type: Optional[dict[str, any]] = None
  • timeout: Specify the number of seconds to wait until assistant operation is completed.
    • If None, wait indefinitely until operation completes
    • If >=0, time out after this many seconds
    • If -1, return immediately and do not wait.
    • type: Optional[int] = None

Returns:

  • AssistantModel object with the following properties:
    • name: Contains the name of the assistant.
    • instructions Custom instructions for the assistant.
    • metadata: Contains the provided metadata.
    • created_at: Contains the timestamp of when the assistant was created.
    • updated_at: Contains the timestamp of when the assistant was last updated.
    • status: Contains the status of the assistant. This is one of:
      • 'Initializing'
      • 'Ready'
      • 'Terminating'
      • 'Failed'

Describe Assistant

The example below describes/fetches an assistant with the specified name. Will raise a 404 if no model exists with the specified name. There are two methods for this:

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

assistant = pc.assistant.describe_assistant(
    assistant_name="example_assistant", 
)

# we can also do this
assistant = pc.assistant.Assistant(
    assistant_name="example_assistant", 
)

Arguments:

  • assistant_name: The name of the assistant to fetch.
    • type: str, required

Returns:

Update Assistant

To update an assistant's metadata and/or instructions, see the below example.

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
metadata = {"author": "Jane Doe", "version": "2.0"}

assistant = pc.assistant.update_assistant(
    assistant_name="example_assistant", 
    instructions="Always use Australian English spelling and vocabulary.",
    metadata=metadata,
)

Arguments:

  • assistant_name: The name of the assistant to fetch.
    • type: str, required
  • instructions Custom instructions for the assistant. These will be applied to all future chat interactions.
    • type: Optional[str] = None
  • metadata: A dictionary containing metadata for the assistant. If provided, it will completely replace the existing metadata unless set to None (default).
    • type: Optional[dict[str, any]] = None

Returns:

List Assistants

Lists all assistants created from the current project. Will raise a 404 if no assistant exists with the specified name.

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

assistants = pc.assistant.list_assistants()

Returns:

  • List[AssistantModel] objects

Delete Assistant

Deletes a assistant with the specified name. Will raise a 404 if no assistant exists with the specified name.

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

pc.assistant.delete_assistant(
    assistant_name="example_assistant", 
)

Arguments:

  • assistant_name: The name of the assistant to fetch.
    • type: str, required

Returns:

  • NoneType

Assistants Model API

Upload File to Assistant

Uploads a file from the specified path to this assistant for internal processing.

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

assistant = pc.assistant.Assistant(
    assistant_name="example_assistant", 
)

# upload file
resp = assistant.upload_file(
    file_path="/path/to/file.txt"
    timeout=None
)

Arguments:

  • file_path: The path to the file that needs to be uploaded.

    • type: str, required
  • timeout: Specify the number of seconds to wait until file processing is done.

    • If None, wait indefinitely.
    • If >= 0, time out after this many seconds.
    • If -1, return immediately and do not wait.
    • type: Optional[int] = None

Return

  • FileModel object with the following properties:
    • id: The file id of the uploaded file.
    • name: The name of the uploaded file.
    • created_on: The timestamp of when the file was created.
    • updated_on: The timestamp of the last update to the file.
    • metadata: Metadata associated with the file.
    • status: The status of the file.

Describe File to Assistant

Describes a file with the specified file id from this assistant. Includes information on its status and metadata.

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

assistant = pc.assistant.Assistant(
    assistant_name="example_assistant", 
)

# describe file
file = assistant.describe_file(file_id="070513b3-022f-4966-b583-a9b12e0290ff")

Arguments:

  • file_id: The file ID of the file to be described.
    • type: str, required

Returns:

  • FileModel object with the following properties:
    • id: The UUID of the requested file.
    • name: The name of the requested file.
    • created_on: The timestamp of when the file was created.
    • updated_on: The timestamp of the last update to the file.
    • metadata: Metadata associated with the file.
    • status: The status of the file.

List Files

Lists all uploaded files in this assistant.

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

assistant = pc.assistant.Assistant(
    assistant_name="example_assistant", 
)

files = assistant.list_files()

Arguments: None

Returns:

  • List[FileModel], the list of files in the assistant

Delete file from assistant

Deletes a file with the specified file_id from this assistant.

from pinecone import Pinecone

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

assistant = pc.assistant.Assistant(
    assistant_name="example_assistant", 
)

# delete file
assistant.delete_file(file_id="070513b3-022f-4966-b583-a9b12e0290ff")

Arguments:

  • file_id: The file ID of the file to be described.
    • type: str, required

Returns:

  • NoneType

Chat

Performs a chat request to the following assistant which returns a stream of chat results in our custom format. Use this API if you want to have more control over the format of the citations. If the stream bool is set to true, this function will stream the response in chunks by returning a generator.

from pinecone import Pinecone
from pinecone_plugins.assistant.models.chat import Message

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

space_assistant = pc.assistant.Assistant(assistant_name="space")

msg = Message(content="How old is the earth?")
resp = space_assistant.chat(messages=[msg])

# The stream version
chunks = space_assistant.chat(messages=[msg], stream=True)

for chunk in chunks:
    if chunk:
        print(chunk)

Arguments:

  • messages: The current context for the chat request. The final element in the list represents the user query to be made from this context.

    • type: List[Message] where Message requires the following:
      • role: str, the role of the context ('user' or 'agent')
      • content: str, the content of the context
  • stream: If this flag is turned on, then the return type is an Iterable[StreamingChatResultModel] where data is returned as a generator/stream.

    • type: bool, default false

Return:

  • The default result is a ChatResultModel with the following format:
    • finish_reason: The reason the response finished, e.g., "stop".
    • index: The index of the choice in the list.
    • message: An object with the following properties:
      • content: The content of the message.
      • role: The role of the message sender, e.g., "assistant".
    • id: The unique identifier of the chat completion.
    • model: The model used for the chat completion, e.g., "gpt-3.5-turbo-0613".
    • citations: A list of citations with the following structure:
      • position: The position of the citation in the document.
      • references: A list of references with the following structure:
        • file: A dictionary with the following properties:
          • created_on: The timestamp of when the file was created.
          • id: The file ID.
          • name: The name of the file.
          • status: The status of the file.
          • updated_on: The timestamp of the last update to the file.
        • pages: The list of pages that the citation references.

The default result is a ChatModel with the following format:

{
    "finish_reason": "stop",
    "index": 0,
    "message": {
        "content": "The 2020 World Series was played in Texas at Globe Life Field in Arlington.",
        "role": "assistant"
    },
    "id": "chatcmpl-7QyqpwdfhqwajicIEznoc6Q47XAyW",
    "model": "gpt-3.5-turbo-0613",
    "citations": [
        {
            "position": 3,
            "references": [
                {
                    "file": {
                        "created_on": "2024-06-02T19:48:00Z",
                        "id": "070513b3-022f-4966-b583-a9b12e0290ff",
                        "name": "tiny_file.txt",
                        "status": "Available",
                        "updated_on": "2024-06-02T19:48:00Z"
                    },
                    "pages": [1, 2, 3]
                }
            ],
        }
    ]
}
  • When stream is set to true, the response is an iterable of StreamingChatResultModel objects with the following properties:

    • choices: A list with the following structure:
      • finish_reason: The reason the response finished, which can be null while streaming.
      • index: The index of the choice in the list.
      • delta: An object with the following properties:
        • content: The incremental content of the message.
        • role: The role of the message sender, which can be empty while streaming.
      • logprobs: The log probabilities (if applicable), otherwise null.
    • id: The unique identifier of the chat completion.
    • model: The model used for the chat completion, e.g., "gpt-4o-2024-05-13".
  • However, when stream is set to true, the response is an stream of ChatResultModel's. This can be one of the following types:

    • StreamChatResultModelMessageStart:
      • type: The type of the message, which is "message_start".
      • id: The unique identifier of the message.
      • model: The model used for the chat completion, e.g., "gpt-4o-2024-05-13".
      • role: The role of the message sender, which is "assistant".

    Example:

        {
            "type": "message_start",
            "id": "0000000000000000468323be9d266e55",
            "model": "gpt-4o-2024-05-13",
            "role": "assistant"
        }
    
    • StreamChatResultModelContentDelta
      • type: The type of the message, which is "content_chunk".
      • id: The unique identifier of the message.
      • model: The model used for the chat completion, e.g., "gpt-4o-2024-05-13".
      • delta: An object with the following properties:
        • content: The incremental content of the message.
        {
            "type": "content_chunk",
            "id": "0000000000000000468323be9d266e55",
            "model": "gpt-4o-2024-05-13",
            "delta": {
                "content": "The"
            }
        }
    
    • StreamChatResultModelCitation
      • type: The type of the message, which is "citation".
      • id: The unique identifier of the message.
      • model: The model used for the chat completion, e.g., "gpt-4o-2024-05-13".
      • citation: An object with the following properties:
        • position: The position of the citation in the document.
        • references: A list of references with the following structure:
          • id: The file ID.
          • file: A dictionary with the following properties:
            • status: The status of the file.
            • id: The file ID.
            • name: The name of the file.
            • size: The size of the file.
            • metadata: The metadata of the file.
            • updated_on: The timestamp of the last update to the file.
            • created_on: The timestamp of when the file was created.
            • percent_done: The percentage of the file that has been processed.
            • signed_url: The signed URL of the file.
          • pages: The list of pages that the citation references.
        {
            "type": "citation",
            "id": "0000000000000000116990b44044d21e",
            "model": "gpt-4o-2024-05-13",
            "citation": {
                "position": 247,
                "references": [{
                    "id": "s0",
                    "file": {
                        "status": "Available",
                        "id": "985edb6c-f649-4334-8f14-9a16b7039ab6",
                        "name": "PEPSICO_2022_10K.pdf",
                        "size": 2993516,
                        "metadata": {},
                        "updated_on": "2024-08-08T15:41:58.839846634Z",
                        "created_on": "2024-08-08T15:41:07.427879083Z",
                        "percent_done": 0,
                        "signed_url": "example.com"
                    },
                    "pages": [
                        32
                    ]
                }]
            }
        }
    
    • StreamChatResultModelMessageEnd
      • type: The type of the message, which is "message_end".
      • id: The unique identifier of the message.
      • model: The model used for the chat completion, e.g., "gpt-4o-2024-05-13".
      • finish_reason: The reason the response finished, e.g., "stop".
      • usage: An object with the following properties:
        • prompt_tokens: The number of prompt tokens used.
        • completion_tokens: The number of completion tokens used.
        • total_tokens: The total number of tokens used.
        {
            "type": "message_end",
            "id": "0000000000000000116990b44044d21e",
            "model": "gpt-4o-2024-05-13",
            "finish_reason": "stop",
            "usage": {
                "prompt_tokens": 1,
                "completion_tokens": 1,
                "total_tokens": 2
            }
        }
    

Chat Completions

Performs a chat completion request to the following assistant. If the stream bool is set to true, this function will stream the response in chunks by returning a generator.

from pinecone import Pinecone
from pinecone_plugins.assistant.models.chat import Message

pc = Pinecone(api_key='<<PINECONE_API_KEY>>')

space_assistant = pc.assistant.Assistant(assistant_name="space")

msg = Message(content="How old is the earth?")
resp = space_assistant.chat_completions(messages=[msg])

# The stream version
chunks = space_assistant.chat_completions(messages=[msg], stream=True)

for chunk in chunks:
    if chunk:
        print(chunk)

Arguments:

  • messages: The current context for the chat request. The final element in the list represents the user query to be made from this context.

    • type: List[Message] where Message requires the following:
      • role: str, the role of the context ('user' or 'agent')
      • content: str, the content of the context
  • stream: If this flag is turned on, then the return type is an Iterable[StreamingChatResultModel] where data is returned as a generator/stream.

    • type: bool, default false

Return:

  • The default result is a ChatResultModel with the following format:
    • choices: A list with the following structure:
      • finish_reason: The reason the response finished, e.g., "stop".
      • index: The index of the choice in the list.
      • message: An object with the following properties:
        • content: The content of the message.
        • role: The role of the message sender, e.g., "assistant".
      • logprobs: The log probabilities (if applicable), otherwise null.
    • id: The unique identifier of the chat completion.
    • model: The model used for the chat completion, e.g., "gpt-3.5-turbo-0613".

See the example below

{
    "choices": [
        {
            "finish_reason": "stop",
            "index": 0,
            "message": {
                "content": "The 2020 World Series was played in Texas at Globe Life Field in Arlington.",
                "role": "assistant"
            },
            "logprobs": null
        }
    ],
    "id": "00000000000000005c12d4d71263b642",
    "model": "space"
}
  • When stream is set to true, the response is an iterable of StreamingChatResultModel objects with the following properties:
    • choices: A list with the following structure:
      • finish_reason: The reason the response finished, which can be null while streaming.
      • index: The index of the choice in the list.
      • delta: An object with the following properties:
        • content: The incremental content of the message.
        • role: The role of the message sender, which can be empty while streaming.
      • logprobs: The log probabilities (if applicable), otherwise null.
    • id: The unique identifier of the chat completion.
    • model: The model used for the chat completion, e.g., "gpt-3.5-turbo-0613".

See the example below

    {
        "choices": [
            {
                "finish_reason": null,
                "index": 0,
                "delta": {
                    "content": "The",
                    "role": ""
                },
                "logprobs": null
            }
        ],
        "id": "00000000000000005d487d0ba0cde006",
        "model": "space"
    }

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pinecone_plugin_assistant-0.3.0.tar.gz (136.6 kB view details)

Uploaded Source

Built Distribution

pinecone_plugin_assistant-0.3.0-py3-none-any.whl (219.0 kB view details)

Uploaded Python 3

File details

Details for the file pinecone_plugin_assistant-0.3.0.tar.gz.

File metadata

  • Download URL: pinecone_plugin_assistant-0.3.0.tar.gz
  • Upload date:
  • Size: 136.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.3 CPython/3.12.6 Linux/6.8.0-1014-azure

File hashes

Hashes for pinecone_plugin_assistant-0.3.0.tar.gz
Algorithm Hash digest
SHA256 50fa61a7d4199800a1f99303cc39ee82fa3f43ecc0268982c6deaebe8290bf59
MD5 161ababfdd3690ad893f6e2e81e3c378
BLAKE2b-256 dc32b6f7d4b9cf1edfd9e8d57d554c027bc0989109e122b853b4fa58f93cd777

See more details on using hashes here.

File details

Details for the file pinecone_plugin_assistant-0.3.0-py3-none-any.whl.

File metadata

File hashes

Hashes for pinecone_plugin_assistant-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 77633dd23bfc95cc3333d1df4156d3c1cba37b5c1ea0ff81639153b7f5b31fb8
MD5 3c76a8a751552b3bf580174d48cf02ba
BLAKE2b-256 541b9769f0db11a28860b7d456e3f29afed7c7473c3674533332cb8c16434516

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page