Skip to main content

LLMT aims to make it easy to programatically connect OpenAI and HuggingFace models to your data pipelines, CI/CD, or personal workspaces.

Project description

LLMT

PyPI version

What is this good for?

LLMT aims to make it easy to programatically connect OpenAI and HuggingFace models to your data pipelines, CI/CD, or personal workspaces.

Supports function calling, chat and context history retention. Python 3.12 and up.

Usage

Use the package in directly in your python code (pip install llmt), or as a local workspace running a container to interact with ChatGPT.

Module import

from llmt import LLMT


llmt = LLMT()
llmt.init_assistant(
    "snowflake",
    api_key="...",
    model="gpt-3.5-turbo",
    assistant_description="You are a snowflake expert. Answer questions briefly in a sentence or less."
)

llmt.init_functions(["./helper_functions.py"])
llmt.init_context("snowflake")
response = llmt.run("generate an example merge dml")

print(response)

Local workspace

Install Docker and make command. Make is not required since you can use docker compose.

  • Clone this repo.
  • If using custom functions, create your functions in the udf/ directory and import them in cli.py.
  • Update the default configuration file, or create a new one in configs/.
  • Run make run. Default config will let you use input and output files.
  • Use files/input.md to send messages.
  • Use files/output.md to receive messages.
  • CTRL + C to quit out of the container and clean up orphans.

Configuration file

If both (input_file, output_file) are ommited, then the default terminal will be used. Using the input and output files to converse with an LLM is easier than using the terminal.

  • input_file: specify a file for user input
  • output_file: specify a file for LLM response
  • assistants:
    • type: Assistant type, currently only OpenAI.
    • assistant_name: Assistant name.
    • assistant_description: Assistant description which OpenAI will use for assistant context.
    • api_key: OpenAI API key.
    • model: OpenAI model.
    • tools: Function definitions. For now, in addition to creating functions, functions must be also defined in a format which OpenAI API can understand. Functions take one object argument which must be unpacked to extract arguments within each function. Hopefully this changes in the future.

The image used for running this code has some common tools installed which I use daily in my custom functions:

  • awscli
  • cloudquery
  • numpy
  • pandas
  • psycopg2-binary
  • SQLAlchemy

Build and use your own image with additional tools for whatever your functions need.

Need help?

I help organizations build data pipelines with AI integrations. If your organization needs help building or exploring solutions, feel free to reach me at artem at outermeasure.com. The general workflow is:

  1. Fine tune a curated model with proprietary data to perform tasks specific to your pipeline.
  2. Deploy the model in your cloud environment.
  3. Connect your pipeline to the deployment via an API.
  4. Iterate and improve the model.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llmt-0.0.5.tar.gz (18.3 kB view details)

Uploaded Source

Built Distribution

llmt-0.0.5-py3-none-any.whl (19.1 kB view details)

Uploaded Python 3

File details

Details for the file llmt-0.0.5.tar.gz.

File metadata

  • Download URL: llmt-0.0.5.tar.gz
  • Upload date:
  • Size: 18.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.12.3

File hashes

Hashes for llmt-0.0.5.tar.gz
Algorithm Hash digest
SHA256 85f65d64f2c12976656d7a0ef304e9220494eeb1bc2253d5cacd96ac142054fe
MD5 8a521e3444598ed91fac9f1c411ce77f
BLAKE2b-256 7130cd97f5be7979f29dd8fe7ecccbd990d6c5c608f714825794135ef1e8349b

See more details on using hashes here.

File details

Details for the file llmt-0.0.5-py3-none-any.whl.

File metadata

  • Download URL: llmt-0.0.5-py3-none-any.whl
  • Upload date:
  • Size: 19.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.12.3

File hashes

Hashes for llmt-0.0.5-py3-none-any.whl
Algorithm Hash digest
SHA256 292d99b6bb49b6be8b9ff9b1a2faf27e72f862e3cb501df72229453330c0e01b
MD5 3a40190ff64020451dc315529a3ef5ed
BLAKE2b-256 b019c1803a8a0ba4e90470e0622f6d387b23a953d98a497cf8d3629aad2b039f

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page