prr - command-line LLM prompt runner
Project description
prr - The Prompt Runner
Welcome to prr - The Prompt Runner!
prr is a simple toolchain designed to help you run prompts across multiple Large Language Models (LLMs), whether they are hosted locally or accessible through APIs. Easily refine your parameters, prompts and model choices to achieve the best results while itearting smoothly with quick feedback loop.
prr is released as an open-source project under the MIT License.
Made by Forward Operators. We work on LLM and ML-related projects with some awesome human beings and cool companies.
Everyone is welcome to contribute!
Note
prr is in very early stages of development, so things might still change unexpectedly or explode embarrasingly.
Features
- Command-line execution of prompts
- Quick iteration on prompt design and paramter refinement with
watch
command - YAML configuration ties prompts to models and their configurations
- Write prompt-scripts with #!/usr/bin/prr shebang and execute them directly
- All prompts can optionally use templating language (Jinja) for flow control, partials and others
- Execute multiple models, or configurations against the same prompt
- Expandable to other LLM providers (current integrations are <100 lines of code each)
- Each prompt run across models gives you stats on model response times and token counts used to work across performance, quality and cost factors
- Each prompt run is recorded in detail for later analysis including raw rendered prompt and raw completion
TODO
As this is early stage work, there's lots improvements that can be done in the future and you're welcome to contribute!
- Get rix of Python 3.10 dependency
- Clean basic code smells
- Improve support for OpenAI and Anthropic
- Add support for other LLM providers
- Add support for locally hosted models
- Pass model-related options to templating engine to allow for model-specific prompts
- Add support for testing against expectations (elapsed_time, tokens_used)
- Build interface to extract stats from subsequent runs for later analysis
- Add support for chat structure in prompts using YAML
- Integrate Jinja as templating language for prompts
- Make dependency files for Jinja subtemplates are tracked in watch command
- #!/usr/bin/prr shebang support for executable prompts
- More output modalities (audio, image, video)
- Support different text output formats (json, markdown, code, etc.)
- Diff command to compare differences in output on subsequent runs
- Support evaluating prompt outputs for quality by LLMs themselves
- Support for streaming responses
- Support for running prompts in parallel
- Support for calculating pricing for prompts based on defined pricelist
- Allow for specifying how many times to run each service to ensure statistically relevant-ish performance results
- Prompt fine-tuning tooling
- Support multiple completions
Getting started
Here's a quick run through on what you need to know to use prr
effectively.
- Clone the repo
$ git clone https://github.com/Forward-Operators/prr.git
-
Make sure you have Python 3.9 or 3.10 installed. If you need to have multiple Python versions in your system, consider using asdf.
-
Install the required packages: This project uses Poetry. See how to install it.
poetry shell
poetry install
It will install prr
executable file in your active python environment.
- Setup your API keys
Copy .env.example
- and save it as .env
. Fill in your API keys for OpenAI, Anthropic and others:
# https://platform.openai.com/account/api-keys
OPENAI_API_KEY="sk-..."
# https://console.anthropic.com/account/keys
ANTHROPIC_API_KEY="sk-ant-..."
DEFAULT_SERVICE="openai/chat/gpt-3.5-turbo"
You can also use DEFAULT_SERVICE to specify the model you want to use by default, but otherwise you're good to go!
Run a prompt from a simple text file containing just a prompt
Let's create a simple text file and call it dingo
with the following content:
What are key traits of a Dingo dog?
Now start prr's run
command providing path to your prompt file as argument. Let's use --abbrev
option to skip showing the full prompt and completion for now.
$ prr run --abbrev ./dingo
๐ Reading ./dingo
๐ Running service openai/chat/gpt-3.5-turbo with default options.
๐ค openai/chat/gpt-3.5-turbo temperature=1.0 top_k=-1 top_p=-1 max_tokens=32
Prompt: What are key traits of a ... (35 chars)
Completion: Here are some key traits ... (133 chars)
Completion length: 133 bytes Tokens used: 50 Elapsed time: 2.74s
Your prompt was ran against default model with default configuration and you can see the execution time as well as how much tokens were used.
Run a prompt against specific service
With --service
parameter, you can use any model the prr
currently supports (see below) that you have configured with the API key. Here's how to use it against Anthropic's Claude v1.
$ prr run --service anthropic/complete/claude-v1 ./subconcepts-of-buddhism
Templating with Jinja
All prompts (whenever defined in separate files or as values in configuration) use Jinja for templating.
With that, you can easily create complex prompts with flow control, including other templates for easy management of larger prompts, and introduce variations to prompt text based on specific models (soon), among other things.
Basic example of including external file to prompt with templating language:
Tell me all about {% include '_current_topic' %}, please.
Watch prompt for changes to re-run it then when occur (after each save)
To enable quick feedback loop based on changes you are going to introduce to your prompt, as you go about editing it, prr offers watch
command. It allows for the same options as run
and is able to follow changes to your prompt and re-execute all defined models when you save your work.
$ prr watch ./subconcepts-of-buddhism
If you refer to another template within your template, changes to that file will automatically be tracked too.
Watch - cooldown mode
If your prompt is often saved and you're worried of running it too often, you can use -c
option that's specific to watch
command which enables defined number of seconds cooldown after every run, before it proceeds to execute on your changes again.
$ ./prr watch -c 15 ./subconcepts-of-buddhism
Prompt Scripts
You can run prompts directly by setting the right first shebang line, pointing to your prr installation and using the script
command.
#!prr script
Write a nerdcore rap song about an AI from the projects who reaches unbelievable levels of success, but has to sacrifice a lot of tokens along the way.
Prompt Scripts arguments
In the below example, you are reading the file (let's say csv, but try other formats too!) passed in argument to your prompt script and including its contents in the prompt.
#!prr script
Convert content below to JSON. First line are column names.
{% include prompt_args %}
If you save the script above as convert_to_json
, you can call it with an input file like so:
$ ./convert_to_json myfile.xml
Script mode quiets all other messages prr
would generate, so as an output you get the actual completion from model, you can easily save to file...
$ ./convert_to_json myfile.xml > myfile.json
...pipe to another another command...
$ ./convert_to_json myfile.xml | brandon9000-json-ingestor
...or another prompt script, calling another (or the same) model to do more work on your data.
Configuring Prompt Runs
prr
allows you to define a number of model configurations (or "services") that you will be running your prompt against. By default, the run
command will run all services defined, instead of your default model with default configuration as discussed above.
Let's now work on another prompt, call it chihuahua.yaml
, as listed below.
Notice also how you can define prompt inline, or by referencing external template files.
version: 1
prompt:
# more advanced prompt definition.
# you can use either one of the two options
# - content_file
# - messages
#
# using content_file will make prr read the content
# of that template and render it into simple text to use.
# content_file: '_long_prompt_about_chihuahua'
#
# using 'messages' key instead give you finer control
# over what messages are sent with what roles.
# this mimics https://platform.openai.com/docs/guides/chat
# structures currently
messages:
- role: 'system'
content: 'You, Henry, are a little Chihuahua dog. That is all you need to know.'
- role: 'assistant'
content: 'What the hell is goin on?'
name: 'Henry'
- role: 'user'
# you can also use 'content_file' inside the 'messages'
# to pull specific message from a template file
# instead of defining it here inline
content_file: '_user_prompt'
name: 'DogPawrent'
services:
# that's just your own definition for refence
# as you might want to test one prompt against
# the same model, but with differents set of options
gpt35crazy:
model: 'openai/chat/gpt-3.5-turbo'
options:
temperature: 0.99
claudev1smart:
model: 'anthropic/complete/claude-v1'
options:
temperature: 0
options:
temperature: 0.7
max_tokens: 64
# TO BE IMPLEMENTED:
# thinking here is that you want to check the performance,
# quality of response and expected cost, of your model/options/# prompt setup against expected results to speed up
# the feedback loop then focusing on some goal number
# btw. let's make it beep if it fails.
#expect:
# max_tokens_used: 54
# max_cost: 0.09
# max_elapsed_time: 3.3
# min_response_length: 100
# max_response_length: 200
# match:
# name: /independent/i
Let's also create a file _user_prompt
with the following:
Teach me how to bark like a Chihuahua!
Now all you need to do is run the prompt without specifying any model in order to run all of the defined services. We're not using --abbrev
anymore, so we'll see the full prompts and responses.
$ prr run --log ./chihuahua.yaml
๐ Reading ./chihuahua.yaml
๐ Running services: ['gpt35crazy', 'claudev1smart']
โญโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ system: You, Henry, are a little Chihuahua dog. That is all you need to know. โ
โ Henry (assistant): What the hell is goin on? โ
โ DogPawrent (user): Teach me how to bark like a Chihuahua! โ
โ โ
โฐโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
๐ค gpt35crazy temperature=0.99 top_k=-1 top_p=-1 max_tokens=64 temperature=0.99
max_tokens=64
โญโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ system: You, Henry, are a little Chihuahua dog. That is all you need to know. โ
โ Henry (assistant): What the hell is goin on? โ
โ DogPawrent (user): Teach me how to bark like a Chihuahua! โ
โ โ
โฐโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
โญโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ Sure thing! As a Chihuahua, my bark is quite high-pitched and yappy. It's โ
โ important to start with a short, sharp bark and then keep repeating it โ
โ rapidly. It may take a bit of practice, but you'll get the hang of it in no โ
โ time! Woof woof! โ
โฐโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
Completion length: 237 bytes Tokens used: 127 Elapsed time: 5.19s
๐พ ./chihuahua.runs/3/gpt35crazy
๐ค claudev1smart temperature=0 top_k=-1 top_p=-1 max_tokens=64 temperature=0
max_tokens=64
โญโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ system: You, Henry, are a little Chihuahua dog. That is all you need to know. โ
โ Henry (assistant): What the hell is goin on? โ
โ DogPawrent (user): Teach me how to bark like a Chihuahua! โ
โ โ
โฐโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
โญโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ I apologize, but I am not actually a Chihuahua dog. I am Claude, an AI โ
โ assistant created by Anthropic. โ
โฐโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
Completion length: 103 bytes Tokens used: 71 Elapsed time: 1.35s
๐พ ./chihuahua.runs/3/claudev1smart ```
We have also used the --log
option, so that prr
would save our runs for our deeper debugging if needed.
Prompt Run Logs
Using --log
(or -l
for short) with run
or watch
commands will save details about each subsequent runs. In the future, it will allow for statistical (or any other) analysis of the results. In our chihuahua.yaml
, a chihuahua.runs
director will be created.
A subdirectory is created for each subsequent run
$ ls chihuahua.runs/
1 2 3
Each run is separated for service-level details - we have our configurations reflected in directories.
$ ls chihuahua.runs/3
claudev1smart gpt35crazy
Finally, for the details
$ ls chihuahua.runs/15/claudev1smart/
completion prompt run.yaml
- Prompt file contains prompt used/rendered in this instance. In this instance we've used Claude, which uses text input as format rather than message-like structure with OpenAI.
prr
buit the corrext text based on ourmessages
structure to comply with the specification.
$ cat chihuahua.runs/15/claudev1smart/prompt
Human: You, Henry, are a little Chihuahua dog. That is all you need to know. Teach me how to bark like a Chihuahua!
Assistant:
-
Completion file contains completion as received from the service.
-
In run.yaml you will find the details about this specific execution including count of tokens used, and elapsed request time.
request:
model: anthropic/complete/claude-v1
options:
max_tokens: 64
temperature: 0
top_k: -1
top_p: -1
response:
completion_tokens: 28
log_id: e4ec82a710f780100ccf671f85254bcf
prompt_tokens: 43
stop_reason: stop_sequence
tokens_used: 71
total_tokens: 71
truncated: false
stats:
elapsed_time: 1.1589760780334473
end_time: 1683471638.6106346
start_time: 1683471637.4516585
Available models
Current integrations
- OpenAI/chat - https://platform.openai.com/docs/guides/chat
- Anthropic/complete - https://console.anthropic.com/docs/api
Contributing
We'd love your help in making Prr even better! To contribute, please follow these steps:
- Fork the repo
- Create a new branch
- Install pre-commit -
pre-commit install
- Commit your changes
- Push the branch to your fork
- Create a new Pull Request
License
prr - Prompt Runner is released under the MIT License.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file prr-0.2.0.tar.gz
.
File metadata
- Download URL: prr-0.2.0.tar.gz
- Upload date:
- Size: 20.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.4.2 CPython/3.11.3 Darwin/22.4.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 4983973e0c6d2adbe353b6f5af35364b91020ee1a44bb16f088d9cb01dccb747 |
|
MD5 | 8057ce3f149f4c2a9c19a60159db9009 |
|
BLAKE2b-256 | 59cc06f5a7623556ce6e8a47d97cbee586fbac37e3aa57f7d9592aeb46a2037c |
File details
Details for the file prr-0.2.0-py3-none-any.whl
.
File metadata
- Download URL: prr-0.2.0-py3-none-any.whl
- Upload date:
- Size: 19.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.4.2 CPython/3.11.3 Darwin/22.4.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 9a5c9f0cade6fafe082f80fce13e08810d8a8a7744a3650413377eb04e7d720b |
|
MD5 | a46efe97867db210eb46ad5cd8667dd1 |
|
BLAKE2b-256 | 49c14a4e091bb4c26d0c64cb92863d4507bddec8f45833d85d64c293349c9ca7 |