llama-index llms heroku managed inference integration
Project description
Heroku Managed Inference
The llama-index-llms-heroku package contains LlamaIndex integrations for building applications with models on Heroku's Managed Inference platform. This integration allows you to easily connect to and use AI models deployed on Heroku's infrastructure.
Installation
pip install llama-index-llms-heroku
Setup
1. Create a Heroku App
First, create an app in Heroku:
heroku create $APP_NAME
2. Create and Attach AI Models
Create and attach a chat model to your app:
heroku ai:models:create -a $APP_NAME claude-3-5-haiku
3. Export Configuration Variables
Export the required configuration variables:
export INFERENCE_KEY=$(heroku config:get INFERENCE_KEY -a $APP_NAME)
export INFERENCE_MODEL_ID=$(heroku config:get INFERENCE_MODEL_ID -a $APP_NAME)
export INFERENCE_URL=$(heroku config:get INFERENCE_URL -a $APP_NAME)
Usage
Basic Usage
from llama_index.llms.heroku import Heroku
from llama_index.core.llms import ChatMessage, MessageRole
# Initialize the Heroku LLM
llm = Heroku()
# Create chat messages
messages = [
ChatMessage(
role=MessageRole.SYSTEM, content="You are a helpful assistant."
),
ChatMessage(
role=MessageRole.USER,
content="What are the most popular house pets in North America?",
),
]
# Get response
response = llm.chat(messages)
print(response)
Using Environment Variables
The integration automatically reads from environment variables:
import os
# Set environment variables
os.environ["INFERENCE_KEY"] = "your-inference-key"
os.environ["INFERENCE_URL"] = "https://us.inference.heroku.com"
os.environ["INFERENCE_MODEL_ID"] = "claude-3-5-haiku"
# Initialize without parameters
llm = Heroku()
Using Parameters
You can also pass parameters directly:
import os
llm = Heroku(
model=os.getenv("INFERENCE_MODEL_ID", "claude-3-5-haiku"),
api_key=os.getenv("INFERENCE_KEY", "your-inference-key"),
inference_url=os.getenv(
"INFERENCE_URL", "https://us.inference.heroku.com"
),
max_tokens=1024,
)
Text Completion
# Simple text completion
response = llm.complete("Explain the importance of open source LLMs")
print(response.text)
Available Models
For a complete list of available models, see the Heroku Managed Inference documentation.
Error Handling
The integration includes proper error handling for common issues:
- Missing API key
- Invalid inference URL
- Missing model configuration
Additional Information
For more information about Heroku Managed Inference, visit the official documentation.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llama_index_llms_heroku-0.2.0.tar.gz.
File metadata
- Download URL: llama_index_llms_heroku-0.2.0.tar.gz
- Upload date:
- Size: 5.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.9 {"installer":{"name":"uv","version":"0.10.9","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
0a06f55ca6c21ca46e4543bdc81bb722adfe7c1995e4b63324cc7958d5d4a08d
|
|
| MD5 |
cb0b782ff31264a3ba16db92675d68b7
|
|
| BLAKE2b-256 |
18bddc2e9f3c29c54fed1a2094d07e071328324328bbf873648e29e9c4e71a55
|
File details
Details for the file llama_index_llms_heroku-0.2.0-py3-none-any.whl.
File metadata
- Download URL: llama_index_llms_heroku-0.2.0-py3-none-any.whl
- Upload date:
- Size: 4.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.9 {"installer":{"name":"uv","version":"0.10.9","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
18c4ad62a3c8a8d78bceed4dbd5beea91ca608a385d44bcb000a2827ebd19505
|
|
| MD5 |
cccacd8bafd7be70817692d7d6206470
|
|
| BLAKE2b-256 |
c2fd46ff7c6ff88eeba5bd56b36980785378e63d358f7ea390bb2d4ba1d003fe
|