Skip to main content

A CLI utility and Python library for interacting with Large Language Models, including OpenAI, PaLM and local models installed on your own machine.

Project description

LLM

PyPI Documentation Changelog Tests License Discord Homebrew

A CLI utility and Python library for interacting with Large Language Models, both via remote APIs and models that can be installed and run on your own machine.

Run prompts from the command-line, store the results in SQLite, generate embeddings and more.

Consult the LLM plugins directory for plugins that provide access to remote and local models.

Full documentation: llm.datasette.io

Background on this project:

Installation

Install this tool using pip:

pip install llm

Or using Homebrew:

brew install llm

Detailed installation instructions.

Getting started

If you have an OpenAI API key you can get started using the OpenAI models right away.

As an alternative to OpenAI, you can install plugins to access models by other providers, including models that can be installed and run on your own device.

Save your OpenAI API key like this:

llm keys set openai

This will prompt you for your key like so:

Enter key: <paste here>

Now that you've saved a key you can run a prompt like this:

llm "Five cute names for a pet penguin"
1. Waddles
2. Pebbles
3. Bubbles
4. Flappy
5. Chilly

Read the usage instructions for more.

Installing a model that runs on your own machine

LLM plugins can add support for alternative models, including models that run on your own machine.

To download and run Mistral 7B Instruct locally, you can install the llm-gpt4all plugin:

llm install llm-gpt4all

Then run this command to see which models it makes available:

llm models
gpt4all: all-MiniLM-L6-v2-f16 - SBert, 43.76MB download, needs 1GB RAM
gpt4all: orca-mini-3b-gguf2-q4_0 - Mini Orca (Small), 1.84GB download, needs 4GB RAM
gpt4all: mistral-7b-instruct-v0 - Mistral Instruct, 3.83GB download, needs 8GB RAM
...

Each model file will be downloaded once the first time you use it. Try Mistral out like this:

llm -m mistral-7b-instruct-v0 'difference between a pelican and a walrus'

You can also start a chat session with the model using the llm chat command:

llm chat -m mistral-7b-instruct-v0
Chatting with mistral-7b-instruct-v0
Type 'exit' or 'quit' to exit
Type '!multi' to enter multiple lines, then '!end' to finish
> 

Using a system prompt

You can use the -s/--system option to set a system prompt, providing instructions for processing other input to the tool.

To describe how the code in a file works, try this:

cat mycode.py | llm -s "Explain this code"

Help

For help, run:

llm --help

You can also use:

python -m llm --help

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llm-0.19.1.tar.gz (42.9 kB view details)

Uploaded Source

Built Distribution

llm-0.19.1-py3-none-any.whl (44.5 kB view details)

Uploaded Python 3

File details

Details for the file llm-0.19.1.tar.gz.

File metadata

  • Download URL: llm-0.19.1.tar.gz
  • Upload date:
  • Size: 42.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/5.1.1 CPython/3.12.7

File hashes

Hashes for llm-0.19.1.tar.gz
Algorithm Hash digest
SHA256 64f0c9500ec26a7de61a3a07b1f0f1cdd333a753c4a7aba7791b4ed3cd54117f
MD5 45b6db1e9216e1f03377a64d73dbcfa6
BLAKE2b-256 1a057f9f9295e166e71a839eaf961ea21bdc44a5f2f155c656be9943f2ff3d4a

See more details on using hashes here.

Provenance

The following attestation bundles were made for llm-0.19.1.tar.gz:

Publisher: publish.yml on simonw/llm

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file llm-0.19.1-py3-none-any.whl.

File metadata

  • Download URL: llm-0.19.1-py3-none-any.whl
  • Upload date:
  • Size: 44.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/5.1.1 CPython/3.12.7

File hashes

Hashes for llm-0.19.1-py3-none-any.whl
Algorithm Hash digest
SHA256 6450fe6ab7b844365da21a8dfacf03d1e26730109f487aec456b872ffc85ae63
MD5 8de4f6486c44473179b4ef9cbb41aa7b
BLAKE2b-256 b5ef14d336309ceddce6d35670aca73bcbd7aede32ce60d4c39a8a949773b8f0

See more details on using hashes here.

Provenance

The following attestation bundles were made for llm-0.19.1-py3-none-any.whl:

Publisher: publish.yml on simonw/llm

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page