Assistant plugin for Pinecone SDK
Project description
Assistant
Interact with Pinecone's Assistant APIs, e.g. create, manage, and chat with assistants (currently in beta). Pinecone Assistant is also available in the console.
⚠️ Note
Pinecone Assistant is currently in beta and access is limited by a waitlist. If you're interested in trying out Pinecone Assistant, please submit a request via the console.
Quickstart
The following example highlights how to use an assistant to store and understand documents on a particular topic and chat with the assistant about those documents with the ultimate goal of semantically understanding your data.
from pinecone import Pinecone
from pinecone_plugins.assistant.models.chat import Message
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
# Create an assistant (in this case we'll store documents about planets)
space_assistant = pc.assistant.create_assistant(assistant_name="space")
# Upload information to your assistant
space_assistant.upload_file("./space-fun-facts.pdf")
# Once the upload succeeded, ask the assistant a question
msg = Message(content="How old is the earth?")
resp = space_assistant.chat_completions(messages=[msg])
print(resp)
# {'choices': [{'finish_reason': 'stop',
# 'index': 0,
# 'message': {'content': 'The age of the Earth is estimated to be '
# 'about 4.54 billion years, based on '
# 'evidence from radiometric age dating of '
# 'meteorite material and Earth rocks, as '
# 'well as lunar samples. This estimate has '
# 'a margin of error of about 1%.',
# 'role': 'assistant'}}],
# 'id': '00000000000000001a377ceeaabf3c18',
Assistants API
Create Assistant
To create an assistant, see the below example. This API creates a assistant with the specified name, metadata, and optional timeout settings.
from pinecone import Pinecone
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
metadata = {"author": "Jane Doe", "version": "1.0"}
assistant = pc.assistant.create_assistant(
assistant_name="example_assistant",
metadata=metadata,
timeout=30
)
Arguments:
assistant_name
The name to assign to the assistant.- type:
str
- type:
metadata
: A dictionary containing metadata for the assistant.- type:
Optional[dict[str, any]] = None
- type:
timeout
: Specify the number of seconds to wait until assistant operation is completed.- If None, wait indefinitely until operation completes
- If >=0, time out after this many seconds
- If -1, return immediately and do not wait.
- type:
Optional[int] = None
Returns:
AssistantModel
object with the following properties:name
: Contains the name of the assistant.metadata
: Contains the provided metadata.created_at
: Contains the timestamp of when the assistant was created.updated_at
: Contains the timestamp of when the assistant was last updated.status
: Contains the status of the assistant. This is one of:- 'Initializing'
- 'Ready'
- 'Terminating'
- 'Failed'
Describe Assistant
The example below describes/fetches an assistant with the specified name. Will raise a 404 if no model exists with the specified name. There are two methods for this:
from pinecone import Pinecone
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
assistant = pc.assistant.describe_assistant(
assistant_name="example_assistant",
)
# we can also do this
assistant = pc.assistant.Assistant(
assistant_name="example_assistant",
)
Arguments:
assistant_name
: The name of the assistant to fetch.- type:
str
, required
- type:
Returns:
AssistantModel
object with the following properties:name
: Contains the name of the assistant.metadata
: Contains the provided metadata.created_at
: Contains the timestamp of when the assistant was created.updated_at
: Contains the timestamp of when the assistant was last updated.status
: Contains the status of the assistant. This is one of:- 'Initializing'
- 'Ready'
- 'Terminating'
- 'Failed'
List Assistants
Lists all assistants created from the current project. Will raise a 404 if no assistant exists with the specified name.
from pinecone import Pinecone
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
assistants = pc.assistant.list_assistants()
Returns:
List[AssistantModel]
objects
Delete Assistant
Deletes a assistant with the specified name. Will raise a 404 if no assistant exists with the specified name.
from pinecone import Pinecone
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
pc.assistant.delete_assistant(
assistant_name="example_assistant",
)
Arguments:
assistant_name
: The name of the assistant to fetch.- type:
str
, required
- type:
Returns:
- NoneType
Assistants Model API
Upload File to Assistant
Uploads a file from the specified path to this assistant for internal processing.
from pinecone import Pinecone
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
assistant = pc.assistant.Assistant(
assistant_name="example_assistant",
)
# upload file
resp = assistant.upload_file(
file_path="/path/to/file.txt"
timeout=None
)
Arguments:
-
file_path
: The path to the file that needs to be uploaded.- type:
str
, required
- type:
-
timeout
: Specify the number of seconds to wait until file processing is done.- If
None
, wait indefinitely. - If
>= 0
, time out after this many seconds. - If
-1
, return immediately and do not wait. - type:
Optional[int] = None
- If
Return
FileModel
object with the following properties:id
: The file id of the uploaded file.name
: The name of the uploaded file.created_on
: The timestamp of when the file was created.updated_on
: The timestamp of the last update to the file.metadata
: Metadata associated with the file.status
: The status of the file.
Describe File to Assistant
Describes a file with the specified file id
from this assistant. Includes information on its status and metadata.
from pinecone import Pinecone
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
assistant = pc.assistant.Assistant(
assistant_name="example_assistant",
)
# describe file
file = assistant.describe_file(file_id="070513b3-022f-4966-b583-a9b12e0290ff")
Arguments:
file_id
: The file ID of the file to be described.- type:
str
, required
- type:
Returns:
FileModel
object with the following properties:id
: The UUID of the requested file.name
: The name of the requested file.created_on
: The timestamp of when the file was created.updated_on
: The timestamp of the last update to the file.metadata
: Metadata associated with the file.status
: The status of the file.
List Files
Lists all uploaded files in this assistant.
from pinecone import Pinecone
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
assistant = pc.assistant.Assistant(
assistant_name="example_assistant",
)
files = assistant.list_files()
Arguments: None
Returns:
List[FileModel]
, the list of files in the assistant
Delete file from assistant
Deletes a file with the specified file_id from this assistant.
from pinecone import Pinecone
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
assistant = pc.assistant.Assistant(
assistant_name="example_assistant",
)
# delete file
assistant.delete_file(file_id="070513b3-022f-4966-b583-a9b12e0290ff")
Arguments:
file_id
: The file ID of the file to be described.- type:
str
, required
- type:
Returns:
NoneType
Chat Completions
Performs a chat completion request to the following assistant. If the stream bool is set to true
, this function will stream the response in chunks by returning a generator.
from pinecone import Pinecone
from pinecone_plugins.assistant.models.chat import Message
pc = Pinecone(api_key='<<PINECONE_API_KEY>>')
space_assistant = pc.assistant.Assistant(assistant_name="space")
msg = Message(content="How old is the earth?")
resp = space_assistant.chat_completions(messages=[msg])
# The stream version
chunks = space_assistant.chat_completions(messages=[msg], stream=True)
for chunk in chunks:
if chunk:
print(chunk)
Arguments:
-
messages
: The current context for the chat request. The final element in the list represents the user query to be made from this context.- type:
List[Message]
whereMessage
requires the following:role
:str
, the role of the context ('user' or 'agent')content
:str
, the content of the context
- type:
-
stream
: If this flag is turned on, then the return type is anIterable[StreamingChatResultModel]
where data is returned as a generator/stream.- type:
bool
, defaultfalse
- type:
Return:
- The default result is a
ChatResultModel
with the following format:choices
: A list with the following structure:finish_reason
: The reason the response finished, e.g., "stop".index
: The index of the choice in the list.message
: An object with the following properties:content
: The content of the message.role
: The role of the message sender, e.g., "assistant".
logprobs
: The log probabilities (if applicable), otherwisenull
.
id
: The unique identifier of the chat completion.model
: The model used for the chat completion, e.g., "gpt-3.5-turbo-0613".
See the example below
{
"choices": [
{
"finish_reason": "stop",
"index": 0,
"message": {
"content": "The 2020 World Series was played in Texas at Globe Life Field in Arlington.",
"role": "assistant"
},
"logprobs": null
}
],
"id": "00000000000000005c12d4d71263b642",
"model": "space"
}
- When
stream
is set totrue
, the response is an iterable ofStreamingChatResultModel
objects with the following properties:choices
: A list with the following structure:finish_reason
: The reason the response finished, which can benull
while streaming.index
: The index of the choice in the list.delta
: An object with the following properties:content
: The incremental content of the message.role
: The role of the message sender, which can be empty while streaming.
logprobs
: The log probabilities (if applicable), otherwisenull
.
id
: The unique identifier of the chat completion.model
: The model used for the chat completion, e.g., "gpt-3.5-turbo-0613".
See the example below
{
"choices": [
{
"finish_reason": null,
"index": 0,
"delta": {
"content": "The",
"role": ""
},
"logprobs": null
}
],
"id": "00000000000000005d487d0ba0cde006",
"model": "space"
}
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file pinecone_plugin_assistant-0.2.0.tar.gz
.
File metadata
- Download URL: pinecone_plugin_assistant-0.2.0.tar.gz
- Upload date:
- Size: 129.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.3 CPython/3.12.5 Linux/6.5.0-1025-azure
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 03ab610181eb6f782955269a285636abff349c704b810b55bbbe233b560678e6 |
|
MD5 | 7fee60548e340235c794faa32e6c8dbd |
|
BLAKE2b-256 | 9ac0119e9202d7739c099fb48e896d1281bd449a5411bb24c84df3b392bc53d6 |
File details
Details for the file pinecone_plugin_assistant-0.2.0-py3-none-any.whl
.
File metadata
- Download URL: pinecone_plugin_assistant-0.2.0-py3-none-any.whl
- Upload date:
- Size: 196.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.3 CPython/3.12.5 Linux/6.5.0-1025-azure
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ac573a1f4b4970d7dcaba2dd61274953a15f9bcacc4ed5bb6b408fba6745c092 |
|
MD5 | 8f2b4f9898a65d8c933cbbaf284b81b3 |
|
BLAKE2b-256 | 6d9065c123cbca6558d8e1d9f82f80c2d18a9b2d17c8b2625691681eed717078 |