Python client for NautilusDB
Project description
nautilusdb-python-client
Python client for NautilusDB, a fully-managed, cloud-native chat service.
NautilusDB is currently in public alpha. We're actively improving the product and releasing new features, and we'd love to hear your feedback! Please take a moment to fill out this feedback form to help us understand your use-case better.
By default, all collections are subject to permanent deletion after 2 weeks. Please let us know if you need to keep it for longer via the feedback form.
NautilusDB python client provides very simple APIs for you to directly upload files and ask questions.
Quickstart
You can try out NautilusDB in just a few lines of code. We have
prepared a special public collection openai-web
that can answer
questions about the contents of www.openai.com
import nautilusdb as ndb
answer, _ = ndb.collection('openai-web').ask('what is red team?')
print(answer)
"""
Sample answer:
Red team refers to the group of external experts who work with OpenAI to
identify and evaluate potential risks and harmful capabilities in new systems.
The red team's role is to help develop taxonomies of risk and provide input
throughout the model and product development lifecycle.
"""
You can also create your own collections, upload files, then get answers specific to your data assets. The following example walks you through the process of creating a collection and indexing the original transformer paper into that collection.
import nautilusdb as ndb
# Create an API key
my_api_key = ndb.create_api_key()
# Configure ndb to use the newly minted API key
ndb.init(api_key=my_api_key)
# Create a new collection
llm_research = ndb.Collection(name="llm_research")
ndb.create_collection(llm_research)
# Index the original Transformer paper into this collection.
llm_research.upload_document("https://arxiv.org/pdf/1706.03762.pdf")
# Get answers from this paper
llm_research.ask("what is a transformer?")
Installation
Install a released NautilusDB python client from pip.
python3 version >= 3.10
is required
pip3 install nautilusdb-client
Creating an API key
You need an API key to create, update, delete own collections. A collection can only be accessed by the API key that created it.
Account management and related functionalities will be released soon.
import nautilusdb as ndb
# Create a new API key
my_api_key = ndb.create_api_key()
# Please record this API key and keep it a secrete
#
# Collections created with this key can only be accessed
# through this key!
print(my_api_key)
# Use this API key in all subsequent calls
ndb.init(api_key=my_api_key)
Creating a Collection
See this page for a brief overview of NautilusDB data model
You can create a collection that is only accessible with a specific API key.
import nautilusdb as ndb
ndb.init(api_key="<my_api_key>")
# Create a collection called llm_research
collection = ndb.Collection(name='llm_research')
ndb.create_collection(collection)
Listing collection
You can see list of collections you have access to. For example, this list will include all collections that were created using the currently configured API key.
import nautilusdb as ndb
ndb.init(api_key="<my_api_key>")
collections = ndb.list_collections()
Uploading a document
You can upload a local file or a file from a web URL and index it into a collection.
Supported file format
.pdf
PDF files.csv
CSV files.txt
Plain-text files.md
Markdown files.docx
Microsoft word documents
import nautilusdb as ndb
ndb.init(api_key="<my_api_key>")
# llm_research collection was created in the previous step
collection = ndb.collection('llm_research')
# Local file and URLs are both supported.
# URL must contain the full scheme prefix (http:// or https://)
collection.upload_document('/path/to/file.pdf')
collection.upload_document('https://path/to/file.pdf')
Crawl a website
You can pass the website root url. NautilusDB will automatically crawl it.
import nautilusdb as ndb
ndb.init(api_key="<my_api_key>")
# Create the crawl task
collection = ndb.Collection(name='example-web')
ndb.create_collection(collection)
crawl_id = collection.create_crawl('https://www.example.com')
# Query the crawl status
resp = collection.get_crawl(crawl_id)
# check resp.crawl_status becomes CrawlStatus.SUCCEEDED, and index the pages
collection.index_crawl(crawl_id)
Asking a question
You can ask questions within a collection. API key is required for private
collections only. ask()
method returns a plain-text answer to
your question, as well as a list of most relevance references used to derive
the answer.
Available public collections that do not require an API key to access
openai-web
: Contains contents ofwww.openai.com
import nautilusdb as ndb
ndb.init(api_key="<my_api_key>")
# Get a plain text answer, as well as a list of references from the collection
# that are the most relevant to the question.
answer, refs = ndb.collection('openai-web').ask('what is red team?')
answer, refs = ndb.collection('llm_research').ask('what is a transformer?')
Deleting a collection
You can delete a collection using the same API key that was used to create it.
import nautilusdb as ndb
ndb.init(api_key="<my_api_key>")
ndb.delete_collection('llm_research')
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file nautilusdb_client-0.8.0.tar.gz
.
File metadata
- Download URL: nautilusdb_client-0.8.0.tar.gz
- Upload date:
- Size: 17.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | b89416e98c3f43c222c0ef7acae0e8ab247e617d6b6c118662b40a1ac4ef68e2 |
|
MD5 | 143da7100cbd80f5bb8fe1082f10de30 |
|
BLAKE2b-256 | bf753edfd794cc638be1b625c1712913ceb7886f4ed3b3176d6065167c00ea93 |
Provenance
File details
Details for the file nautilusdb_client-0.8.0-py3-none-any.whl
.
File metadata
- Download URL: nautilusdb_client-0.8.0-py3-none-any.whl
- Upload date:
- Size: 21.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 88af4a300b8b79cb39aef2ad7af60447208333d94032316e25cd9016f2797ec1 |
|
MD5 | 4de1a0f564c64f77c559aa98b676b555 |
|
BLAKE2b-256 | 7cc7e6b79707827541755981af7f2b72175196c467800e217615cb87f8ed8259 |