Skip to main content

CLI tool to wrap ChatGPT Python API.

Project description

chatgpt-prompt-wrapper

test test coverage

Python CLI implementation for ChatGPT.

Requirements

  • Python >= 3.10

Installation

By Homebrew:

$ brew install rcmdnk/rcmdnkpac/chatgpt-prompt-wrapper

By pip:

$ pip3 install chatgpt-prompt-wrapper

Preparation

For OpenAI API

To get an OpenAI API key, go to Account API Keys - OpenAI API.

Set it as the environment variable OPENAI_API_KEY.

To load the variable when you start the terminal, write it in .bashrc or .zshrc.

export OPENAI_API_KEY="sk-..."

Alternatively, pass the key using the '-k' or '--key' argument in the command, or set KEY in the configuration file.

For other APIs

You can use other APIs by setting the base URL and the key in the configuration file.

To use DeepSeek API, give the base URL and the key by options like:

$ cg --model deepseek-chat --base-url https://api.deepseek.com/v1 --key "<DEEPSEEK_API_KEY>" ...

You can also use ollama local API like:

$ cg --model llama3.2 --base-url http://localhost:11434/v1/ ...

You can set them in the configuration file by MODEL, BASE_URL and KEY keywords.

Usage

Command-line interface Help

usage: cg [-h] [-c CONF] [-k KEY] [-b BASE_URL] [-m MODEL] [-w CONTEXT_WINDOW] [-o MAX_OUTPUT_TOKENS]
          [-O MIN_OUTPUT_TOKENS] [--show] [--hide] [--multiline] [--no_multiline] [--vi] [--emacs] [--show_cost]
          subcommand [message ...]

positional arguments:
  subcommand            Subcommand to run. Use 'commands' subcommand to list up available subcommands.
  message               Message to send to ChatGPT

options:
  -h, --help            show this help message and exit
  -c CONF, --conf CONF  Path to the configuration TOML file.
  -k KEY, --key KEY     API key.
  -b BASE_URL, --base-url BASE_URL
                        API base URL.
  -m MODEL, --model MODEL
                        ChatGPT Model to use.
  -w CONTEXT_WINDOW, --context-window CONTEXT_WINDOW
                        The maximum number of tokens the model can process at once, including both input and output.
                        Set 0 to use the max values for the model.
  -o MAX_OUTPUT_TOKENS, --max-output-tokens MAX_OUTPUT_TOKENS
                        The maximum of output tokens for the completion. Set 0 to use the max values for the model.
  -O MIN_OUTPUT_TOKENS, --min-output-tokens MIN_OUTPUT_TOKENS
                        The minimum of output tokens for the completion. The input tokens must be less than
                        conext_window - min_output_tokens (- a few tokens for the model to process).
  --show                Show prompt for `ask` mode.
  --hide                Hide prompt for `ask` mode.
  --multiline           Use multiline input for `chat` mode.
  --no_multiline        Use single line input for `chat` mode.
  --vi                  Use vi mode at `chat`.
  --emacs               Use emacs mode at `chat`.
  --show_cost           Show cost used.

$ cg commands
Available subcommands:
  Reserved commands:
    ask       : Ask w/o predefined prompt.
    chat      : Start chat w/o predefined prompt.
    discuss   : Start a discussion between GPTs. Give a them as a message.
    init      : Initialize config file with an example command.
    cost      : Show estimated cost used until now.
    commands  : List up subcommands (show this).
    version   : Show version.
    help      : Show help.
  User commands:
    test      : Example command to test the OpenAI API.
    ...

Ask, Chat

ask and chat are reserved commands for running simple interaction without a predefined prompt.

  • ask

cg ask <message> returns the answer from ChatGPT for message.

  • chat

cg chat starts a chat.

:memo: In chat mode, all messages in the past, including answers from ChatGPT, will be sent each time you send a new message.

The oldest message will be dropped when the total tokens exceeds the context_window.

It means you will send almost the max length after a long conversation. Please keep the cost in mind. You may want to set context_window.

Discuss

discuss is another reserved command which start a discussion between two ChatGPTs.

Two GPTs are ask to perform as:

  • Please engage in the discussion as a supporter.
  • Please engage in the discussion as a opponent.

You input message will be a message and they start a discussion.

Please push Enter to proceed a duscussion and Ctrl-C to quit a discussion.

Configuration file

You can define your command in the configuration files.

A command can be in either ask mode, chat mode or discuss mode.

  • ask mode: Send a predefined prompt and a message from the command line and receive one answer.
  • chat mode: Start a chat with a predefined prompt if defined:
    • chat mode can be either multiline or single (no_multiline), and vim or emacs.
  • discuss mode: Start a discussion between two different ChatGPTs.

File path

The default path to the configuration file is $XDG_CONFIG_HOME/cg/config.toml.

If $XDG_CONFIG_HOME is not defined, use ~/.config/cg/config.toml.

If it does not exist and ~/.cg/config.toml exists, the existing file is used.

You can change the path by -c <file> (--conf <file>) option.

How to write the configuration file

The configuration file is written in the TOML format.

You can define a command as the top table name.

global is the special table to define global options, which are enabled for all commands if the command does not have the corresponding options.

The options for each table can be:

  • description: Description of the command.
  • mode: Set ask, chat or discuss. (default is ask mode.)
  • show_cost: Set true to show the cost at the end of the command.
  • model: The model to use (default: "gpt-3.5-turbo").
  • context_window: The maximum number of tokens the model can process at once, including both input and output. Set 0 to use the max values for the model. (default: 0)
  • max_output_tokens: The maximum of output tokens for the completion. Set 0 to use the max values for the model. Set 0 to use the max values for the model. (default: 0)
  • min_output_tokens: The minimum of output tokens for the completion. The input tokens must be less than conext_window - min_output_tokens (- a few tokens for the model to process). (default: 200)
  • temperature: Sampling temperature (0 ~ 2). (default: 1)
  • top_p: Probability (0 ~ 1) that the model will consider the top_p tokens. Do not set both temperature and top_p at the same time. (default: 1)
  • presence_penalty: The penalty for the model to return the same token (-2 ~ 2). (default: 0)
  • frequency_penalty: The penalty for the model to return the same token multiple times (-2 ~ 2). (default: 0)
  • Table of alias: Dictionary of role aliases. The default alias is: 'user' = 'User', 'system' = 'System', 'assistant' = 'Assistant'.
  • model_context_window: The context window for each model.
  • model_max_output_tokens: The maximum output tokens for each model.
  • price: Additional or updated model's price definitions.
  • List of messages: Dictionary of message, which must have role and content (message text).
    • For ask, chat modes, role must be one of system, user and assistant
    • For discuss mode, three roles, theme, gpt1 and gpt2 are needed.

The options for ask mode:

  • show: Set true to show prompt for non chat command.
  • hide: Set true to hide prompt for non chat command (default).

The options for chat mode:

  • multiline: Set true to hide prompt for non chat command.
  • no_multiline: Set true to hide prompt for non chat command (default).
  • vi: Set true to use vi mode.
  • emacs: Set true to use emacs mode (default).

You can make a example configuration file by cg init.

Here is a example configuration:

[global]
# Global configuration
# `global` is special name and not a subcommand
model = "gpt-4o"

# If model has no corresponding tiktoken's encoding, specify encoding_name
# Such a "chatgpt-4o-latest" is not supported by tiktoken.encoding_for_model
#encoding_name = "o200k_base"

# Following context_window, max_output_tokens and prices are pre-defined in
# https://github.com/rcmdnk/chatgpt-prompt-wrapper/blob/main/src/chatgpt_prompt_wrapper/chatgpt/chatgpt.py
# If you find new model or price change, you can overwrite these variables in config as below.

[global.model_context_window]
# https://platform.openai.com/docs/models/
"gpt-4o" = 128000

[global.model_max_output_tokens]
# https://platform.openai.com/docs/models/
"gpt-4o" = 16384

[global.prices]
# https://openai.com/api/pricing/
"gpt-4o" = [0.0025, 0.010]

[test]
# Example command to test the OpenAI API, taken from below.
# [Chat completion - OpenAI API](https://platform.openai.com/docs/guides/chat/introduction)

description = "Example command to test the OpenAI API."
show = true

[[test.messages]]
role = "system"
content = "You are a helpful assistant."
[[test.messages]]
role = "user"
content = "Who won the world series in 2020?"
[[test.messages]]
role = "assistant"
"content" = "The Los Angeles Dodgers won the World Series in 2020."
[[test.messages]]
role = "user"
content = "Where was it played?"

[sh]
description = "Ask a shell scripting question."
[[sh.messages]]
role = "user"
content = "You are an expert of the shell scripting. Answer the following questions."

[py]
description = "Ask a python programming question."
[[py.messages]]
role = "user"
content = "You are an expert python programmer. Answer the following questions."

[enjoy_chat]
description = "Chat example with a predefined prompt."
mode = 'chat'
[[chat.messages]]
role = "user"
content = "Let's enjoy a chat."


[dd]
mode = 'discuss'

[dd.names]
gpt1 = "pros"
gpt2 = "cons"

[[dd.messages]]
role = "theme"
content = "Discuss whether or not to actively adopt ChatGPT in the education field. Please provide one-sentence responses."
[[dd.messages]]
role = "gpt1"
content = "Please speak from the standpoint that you should actively adopt ChatGPT in the field of education."
[[dd.messages]]
role = "gpt2"
content = "Please speak from the standpoint that you should not actively adopt ChatGPT in the field of education."

These messages will be sent as an prompt before your input message.

You can give full questions and use cg w/o input messages like a first example test command.

Command examples:

  • test

test command

  • sh

sh command

  • py

py command

  • enjoy_chat

chat command

Example usage as a part of an external script

Git commit by ChatGPT

See git-gpt-commit.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

chatgpt_prompt_wrapper-0.3.0.tar.gz (429.6 kB view details)

Uploaded Source

Built Distribution

chatgpt_prompt_wrapper-0.3.0-py3-none-any.whl (26.6 kB view details)

Uploaded Python 3

File details

Details for the file chatgpt_prompt_wrapper-0.3.0.tar.gz.

File metadata

File hashes

Hashes for chatgpt_prompt_wrapper-0.3.0.tar.gz
Algorithm Hash digest
SHA256 366f96748298d663712ea718f6c552fb8380ed02233ede032f9bfa8e068bc2e1
MD5 15e5401d79bf1cc505e92c7cf4ef1050
BLAKE2b-256 f5b890fdf9d74746d8e370466f3b6d8269bb77a5ea3df0868470b874b3f9543e

See more details on using hashes here.

File details

Details for the file chatgpt_prompt_wrapper-0.3.0-py3-none-any.whl.

File metadata

File hashes

Hashes for chatgpt_prompt_wrapper-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 1c25e787e9a113613c6a9e9b55118773fed9da8f542a237abcf53ffa4b2265a6
MD5 fa3ee1087d72ae36e1c4a52d85556b47
BLAKE2b-256 70f0c384ef2de8df6afb2c2281c4be6e28181d71c7154742a61ee24310a3371d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page