🐍 | Python library for langchain using RunPod API endpoint as LLM.
Project description
Runpod LLM API Endpoint Lib for LangChain
Installation
# Install the latest release version
pip install runpod-llm
# or
# Install the latest development version (main branch)
pip install git+https://https://github.com/tsangwailam/langchain-runpod-llm
Get Runpod API key
- Goto www.runpod.io. Create a RunPod account.
- From the portal, goto Settings>APIKeys
- Create a new API key by click the "+ API Key" button.
Usage
from runpod_llm import RunpodLlama2
llm = RunpodLlama2(
apikey="YOU_RUNPOD_API_KEY",
llm_type="7b|13b",
config={
"max_tokens": 500,
#Maximum number of tokens to generate per output sequence.
"n": 1, # Number of output sequences to return for the given prompt.
"best_of": 1, # Number of output sequences that are generated from the prompt. From these best_of sequences, the top n sequences are returned. best_of must be greater than or equal to n. This is treated as the beam width when use_beam_search is True. By default, best_of is set to n.
"Presence penalty": 0.2, # Float that penalizes new tokens based on whether they appear in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
"Frequency penalty": 0.5, # Float that penalizes new tokens based on their frequency in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
"temperature": 0.3, # Float that controls the randomness of the sampling. Lower values make the model more deterministic, while higher values make the model more random. Zero means greedy sampling.
"top_p": 1, # Float that controls the cumulative probability of the top tokens to consider. Must be in (0, 1]. Set to 1 to consider all tokens.
"top_k": -1, # Integer that controls the number of top tokens to consider. Set to -1 to consider all tokens.
"use_beam_search": False, # Whether to use beam search instead of sampling.
},
verbose=True, # verbose output
)
some_prompt_template = xxxxx
output_chain = some_prompt_template | llm
output_chain.invoke({"input":"some input to prompt template"})
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file langchain_runpod_llm-0.0.5.tar.gz
.
File metadata
- Download URL: langchain_runpod_llm-0.0.5.tar.gz
- Upload date:
- Size: 9.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.9.19
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 1cc87ef70c3819a1aacacdd092c80c7cb2911c3acb3ba6459452f5dcfb99aded |
|
MD5 | 5ba62975d01ab6272f35d3b149ba1bca |
|
BLAKE2b-256 | 6461e0db8a82ae97f0553947ba7cc98c8344c819b5a678e0fb7e82945002c5e8 |
File details
Details for the file langchain_runpod_llm-0.0.5-py3-none-any.whl
.
File metadata
- Download URL: langchain_runpod_llm-0.0.5-py3-none-any.whl
- Upload date:
- Size: 10.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.9.19
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ee4248db8b46101fafb32b04821af9fb3020806cb35268d381a898112913856b |
|
MD5 | 5a61f7060d9c1a6b8b0f18c569ca089b |
|
BLAKE2b-256 | e9645dbeeb2a4dff6770f572655ef1b7420e98ae2ef12e3c92e5b739393b5666 |