Skip to main content

🐍 | Python library for langchain using RunPod API endpoint as LLM.

Project description

Runpod LLM API Endpoint Lib for LangChain

PyPI version

Installation

# Install the latest release version
pip install runpod-llm

# or

# Install the latest development version (main branch)
pip install git+https://https://github.com/tsangwailam/langchain-runpod-llm

Get Runpod API key

  1. Goto www.runpod.io. Create a RunPod account.
  2. From the portal, goto Settings>APIKeys
  3. Create a new API key by click the "+ API Key" button.

Usage

from runpod_llm import RunpodLlama2

llm = RunpodLlama2(
        apikey="YOU_RUNPOD_API_KEY",
        llm_type="7b|13b",
        config={
            "max_tokens": 500, 
            #Maximum number of tokens to generate per output sequence.
            "n": 1,  # Number of output sequences to return for the given prompt.
            "best_of": 1,  # Number of output sequences that are generated from the prompt. From these best_of sequences, the top n sequences are returned. best_of must be greater than or equal to n. This is treated as the beam width when use_beam_search is True. By default, best_of is set to n.
            "Presence penalty": 0.2,  # Float that penalizes new tokens based on whether they appear in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
            "Frequency penalty": 0.5,  # Float that penalizes new tokens based on their frequency in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
            "temperature": 0.3,  # Float that controls the randomness of the sampling. Lower values make the model more deterministic, while higher values make the model more random. Zero means greedy sampling.
            "top_p": 1,  # Float that controls the cumulative probability of the top tokens to consider. Must be in (0, 1]. Set to 1 to consider all tokens.
            "top_k": -1,  # Integer that controls the number of top tokens to consider. Set to -1 to consider all tokens.
            "use_beam_search": False,  # Whether to use beam search instead of sampling.
        },
        verbose=True, # verbose output
    )

    some_prompt_template = xxxxx
    output_chain = some_prompt_template | llm
    output_chain.invoke({"input":"some input to prompt template"})

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

langchain_runpod_llm-0.0.5.tar.gz (9.0 kB view details)

Uploaded Source

Built Distribution

langchain_runpod_llm-0.0.5-py3-none-any.whl (10.2 kB view details)

Uploaded Python 3

File details

Details for the file langchain_runpod_llm-0.0.5.tar.gz.

File metadata

  • Download URL: langchain_runpod_llm-0.0.5.tar.gz
  • Upload date:
  • Size: 9.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.9.19

File hashes

Hashes for langchain_runpod_llm-0.0.5.tar.gz
Algorithm Hash digest
SHA256 1cc87ef70c3819a1aacacdd092c80c7cb2911c3acb3ba6459452f5dcfb99aded
MD5 5ba62975d01ab6272f35d3b149ba1bca
BLAKE2b-256 6461e0db8a82ae97f0553947ba7cc98c8344c819b5a678e0fb7e82945002c5e8

See more details on using hashes here.

File details

Details for the file langchain_runpod_llm-0.0.5-py3-none-any.whl.

File metadata

File hashes

Hashes for langchain_runpod_llm-0.0.5-py3-none-any.whl
Algorithm Hash digest
SHA256 ee4248db8b46101fafb32b04821af9fb3020806cb35268d381a898112913856b
MD5 5a61f7060d9c1a6b8b0f18c569ca089b
BLAKE2b-256 e9645dbeeb2a4dff6770f572655ef1b7420e98ae2ef12e3c92e5b739393b5666

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page