Skip to main content

A handy toolkit for using LLM.

Project description

HandyLLM

A handy toolkit for using LLM.

OpenAI API Request

This toolkit uses HTTP API request instead of OpenAI's official python package to support client-side timeout control.

from handyllm import OpenAIAPI
OpenAIAPI.api_key = os.environ.get('OPENAI_API_KEY')
response = OpenAIAPI.chat(
    model="gpt-3.5-turbo",
    messages=prompt,
    timeout=10
    )

Prompt

PromptConverter can convert this text file into a structured prompt for chat API calls.

$system$
You are a helpful assistant.

$user$
Please help me merge the following two JSON documents into one.

$assistant$
Sure, please give me the two JSON documents.

$user$
{
	"item1": "It is really a good day."
}
{
	"item2": "Indeed."
}
%output_format%

It can also help substitute placeholder variables (e.g. %output_format% in this example) to make multiple prompts modular. A substitute map looks like this:

%output_format%
Please output a SINGLE JSON object that contains all items from the two input JSON objects.

%variable1%
Placeholder text.

%variable2%
Placeholder text.

%variable3%
Placeholder text.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

HandyLLM-0.1.0.tar.gz (4.9 kB view hashes)

Uploaded Source

Built Distribution

HandyLLM-0.1.0-py3-none-any.whl (5.7 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page