No project description provided
Project description
lango-cli-beta
Installation
Run this command to add the CLI to your environment:
pip install lango-cli-beta
Verify the installation by running:
lango-cli --version
MongoDB Setup
This CLI is build around MongoDB Atlas, and the MongoDB LangChain integration. In order to use this CLI, you'll need to perform the following steps to create an account, cluster, database and search index.
You can create an account here, or login to an existing account.
If you're creating a new account, simply follow their onboarding flow to create a new cluster. If you have an existing account, follow these steps to deploy a new cluster.
Once your cluster has been created, it should prompt you to create a new database user. Record the username and password, as you'll need it to connect to the database.
After creating your cluster, database and collection, you'll need to create a search index.
IMPORTANT Follow the steps on this page to create a new search index, using the JSON configuration object defined below.
{
"fields": [
{
"numDimensions": 1024,
"path": "embedding",
"similarity": "cosine",
"type": "vector"
},
{
"type": "filter",
"path": "_path"
},
{
"type": "filter",
"path": "_index"
}
]
}
Here, you can see two filters for _path
and _index
. These are automatically added when ingesting data using the CLI.
_path
is the path to the file that the data was ingested from._index
is the index of the chunk from the file that the data was ingested from.
If you plan on adding more metadata, or want to filter on additional fields, you should add them to the fields
array like so:
{
"type": "filter",
"path": "metadata_field_name",
}
Once your search index has finished building, you can use the CLI to ingest data and query the database.
Usage
First, ensure you've set the required environment variables:
export MONGODB_URI=connection-string
export MONGODB_VECTORSTORE_NAME=db-name
export MONGODB_COLLECTION_NAME=-collection-name
export MONGODB_VECTORSTORE_INDEX_NAME=search-index-name
Next, set the API keys for the language/embedding models to use. Cohere is required because it's used for embeddings & document re-ranking.
The rest are optional, and only required if you plan to use that provider in your RAG model:
Required
export COHERE_API_KEY=
Optional
export OPENAI_API_KEY=
export ANTHROPIC_API_KEY=
export MISTRAL_API_KEY=
export FIREWORKS_API_KEY=
export GOOGLE_APPLICATION_CREDENTIALS=
init
The first command you should run to setup is:
lango-cli init <project name>
By default this will create a new project with the following structure:
├── <project name>
│ ├── _sample_data
│ │ ├── ...txt files
│ ├── lango_cli.config.json
The lango_cli.config.json
file is where your RAG configuration will live. You will never need to edit this file directly, as the CLI will handle all configuration changes for you.
The _sample_data
directory contains sample data that you can use to test the CLI. These .txt
files are blog posts from Lilian Weng's blog on language models, and other interesting technical topics!
You can replace this data with your own data, or use the CLI to ingest data from a different location. You may choose to not include this directory by passing the --no-sample-data
flag when running init
.
lango-cli init <project name> --no-sample-data
This will create a new directory with thr project name, and populate it with a default configuration file.
Once initialized, navigate into the project directory:
cd <project name>
config
The config
command has two modes: interactive and fast.
Help:
lango-cli config --help
Usage: lango-cli config [OPTIONS]
╭─ Options ──────────────────────────────────────────────────────────────────────────────────────────╮
│ --fast --no-fast Pass this flag to skip the interactive prompts. │
│ [default: no-fast] │
│ --data-path TEXT The path to the data folder or file to ingest. │
│ [default: None] │
│ --rag-type TEXT The type of RAG to create. Options: 'basic', │
│ 'query_construction' │
│ [default: basic] │
│ --llm TEXT The LLM to use. Options: [1]: Anthropic (Claude 3 Opus) │
│ [2]: OpenAI (GPT-4 Turbo) [3]: Cohere (Command R) [4]: │
│ Mistral (Mistral-Large) [5]: Google (Gemini Pro) [6]: │
│ Mixtral-8x7b │
│ [default: 1] │
│ --metadata TEXT The path to the metadata JSON file, or a JSON string. │
│ [default: None] │
│ --chunk-size INTEGER The size of the chunks to split the data into. │
│ [default: 500] │
│ --chunk-overlap INTEGER The overlap between chunks. [default: 75] │
│ --help Show this message and exit. │
╰────────────────────────────────────────────────────────────────────────────────────────────────────╯
To use interactive prompt, run:
lango-cli config
During config
, you'll be prompted to either ingest data now, or later. If you choose now, your MongoDB vectorstore database will be populate with whatever data you provide the CLI. If you choose later, you can run the upsert
command to ingest data at a later time.
Fast mode
lango-cli config --fast --data-path /this/is/required/when/using/fast/mode
upsert
The upsert
command has two modes: interactive and fast.
Help:
lango-cli upsert --help
Usage: lango-cli upsert [OPTIONS]
╭─ Options ──────────────────────────────────────────────────────────────────────────────────────────╮
│ --fast --no-fast Pass this flag to skip the interactive prompts. │
│ [default: no-fast] │
│ --data-path TEXT The path to the data folder or file to ingest. │
│ [default: None] │
│ --metadata TEXT The path to the metadata JSON file, or a JSON string. │
│ [default: None] │
│ --chunk-size INTEGER The size of the chunks to split the data into. │
│ [default: 500] │
│ --chunk-overlap INTEGER The overlap between chunks. [default: 75] │
│ --help Show this message and exit. │
╰────────────────────────────────────────────────────────────────────────────────────────────────────╯
To use interactive prompt, run:
lango-cli upsert
To use fast mode, run:
lango-cli upsert --fast --data-path /this/is/required/when/using/fast/mode
chat
To use chat:
lango-cli chat
start
The Lango CLI also comes with build in support for LangServe where you can chat with your newly created RAG model.
Simply run:
lango-cli start
And visit http://0.0.0.0:8000/chat/playground/
to chat with your RAG model.
export
The lango-cli
comes with an export
command to eaisily deploy your app. The export
command creates a new Dockerfile
which performs the following actions when run:
- Installs the
lango-cli
package. - Copies your
lango_cli.config.json
file into the container. - Starts the LangServe server via
lango-cli start
.
To use the export
command, simply run:
lango-cli export
Then, build your docker container via:
docker build -t lango-cli .
Then, start the container via:
docker run -p 8000:8000 -it --env-file .env lango-cli
After running this command, the first log you'll see will look like this:
Start the Docker container, then navigate to <your ip>:8000/docs to visit the server
Then, if you copy that URL into your browser, you'll be able to see all the available endpoints for your RAG model.
This command automatically loads your .env
file into the container. You can optionally pass the env vars via the -e
flag:
docker run -p 8000:8000 -it -e MONGODB_URI=... lango-cli
NOTE: The
Dockerfile
which is created as a result of runninglango-cli export
must be built inside the generated project directory since it copies a script file, and yourlango_cli.config.json
file into the container.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file lango_cli_beta-0.0.1rc2.tar.gz
.
File metadata
- Download URL: lango_cli_beta-0.0.1rc2.tar.gz
- Upload date:
- Size: 134.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.7.1 CPython/3.11.4 Darwin/23.4.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 84f58840744b61a523e567857a4e5d97a2f82a66c6603a58856887951cec903b |
|
MD5 | db40c30d44591f7c125759619ce56055 |
|
BLAKE2b-256 | 1300b23c901a0008d6544044b97dbacf34e525ec7860f7fd7e6a4e300d7104b4 |
File details
Details for the file lango_cli_beta-0.0.1rc2-py3-none-any.whl
.
File metadata
- Download URL: lango_cli_beta-0.0.1rc2-py3-none-any.whl
- Upload date:
- Size: 149.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.7.1 CPython/3.11.4 Darwin/23.4.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 8ca2092f00e2f9595c8490de30b7dc3c09a2e7bd94acce900a1aeff6d5c6a578 |
|
MD5 | fca40d264c37ff188d01569a0e10696f |
|
BLAKE2b-256 | 49e0afcf76662725b4177cf1af7a82af79e14f5a9903ce6b7ba3853e8b39b75f |