llama-index llms asi integration
Project description
ASI-1 Mini Integration for LlamaIndex
This package contains the LlamaIndex integration with ASI-1 Mini, a powerful language model designed for various natural language processing tasks.
ASI-1 Mini is the world's first Web3-native Large Language Model (LLM) developed by Fetch.ai Inc., a founding member of the Artificial Superintelligence Alliance. Unlike general-purpose LLMs, ASI-1 Mini is specifically designed and optimized for supporting complex agentic workflows.
With ASI-1 Mini, you can leverage these powerful capabilities:
- Advanced agentic reasoning with dynamic reasoning modes for complex tasks
- High performance on par with leading LLMs but with significantly lower hardware costs
- Specialized optimization for autonomous agent applications and multi-step tasks
- Seamless Web3 integration for secure and autonomous AI interactions
Want to learn more about ASI? Visit the ASI website or Fetch.ai for more information!
Installation
pip install llama-index-llms-asi
Usage
Here's an example of how to use the ASI integration with LlamaIndex:
from llama_index.llms.asi import ASI
# Initialize the ASI LLM
llm = ASI(model="asi1-mini", api_key="your_api_key")
# Generate text
response = llm.complete("Tell me about artificial intelligence.")
print(response)
# Chat completion
from llama_index.core.llms import ChatMessage, MessageRole
messages = [
ChatMessage(
role=MessageRole.SYSTEM, content="You are a helpful AI assistant."
),
ChatMessage(
role=MessageRole.USER, content="Tell me about artificial intelligence."
),
]
response = llm.chat(messages)
print(response)
Streaming Support
The ASI integration has different streaming implementations for completion and chat:
-
Streaming Completion: ASI doesn't support streaming for completions (returns 404 error). Our implementation uses a fallback mechanism that returns the complete response as a single chunk.
-
Streaming Chat: ASI supports streaming for chat, but with a unique format that includes:
- Many empty content chunks during the "thinking" phase
- Custom fields like
thoughtandinit_thoughtthat contain intermediate reasoning - Actual content appearing later in the stream
Our implementation processes this format to filter out empty chunks and extract meaningful content, providing a clean streaming experience.
# Streaming completion (falls back to regular completion)
for chunk in llm.stream_complete("Tell me about artificial intelligence."):
print(chunk.text, end="", flush=True)
# Streaming chat (handles ASI's unique streaming format)
for chunk in llm.stream_chat(messages):
if hasattr(chunk, "delta") and chunk.delta.strip():
print(chunk.delta, end="", flush=True)
Async Support
The ASI integration also supports async operations:
# Async completion
response = await llm.acomplete("Tell me about artificial intelligence.")
print(response)
# Async chat
response = await llm.achat(messages)
print(response)
# Async streaming completion (falls back to regular completion)
async for chunk in llm.astream_complete(
"Tell me about artificial intelligence."
):
print(chunk.text, end="", flush=True)
# Async streaming chat (handles ASI's unique streaming format)
async for chunk in llm.astream_chat(messages):
if hasattr(chunk, "delta") and chunk.delta.strip():
print(chunk.delta, end="", flush=True)
API Key
You need an API key to use ASI's API. You can provide it in two ways:
- Pass it directly to the ASI constructor:
ASI(api_key="your_api_key") - Set it as an environment variable:
export ASI_API_KEY="your_api_key"
Models
Currently, this integration supports the following models:
asi1-mini: A powerful language model for various natural language processing tasks.
Development
To create a development environment, install poetry then run:
poetry install --with dev
Testing
To test the integration, first enter the poetry venv:
poetry shell
Then tests can be run with make
make test
Integration tests
Integration tests will be skipped unless an API key is provided. API keys can be obtained from the Fetch.ai team. Once created, store the API key in an environment variable and run tests
export ASI_API_KEY=<your key here>
make test
Linting and Formatting
Linting and code formatting can be executed with make.
make format
make lint
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llama_index_llms_asi-0.1.0.tar.gz.
File metadata
- Download URL: llama_index_llms_asi-0.1.0.tar.gz
- Upload date:
- Size: 5.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.3 CPython/3.12.3 Linux/6.8.0-1021-azure
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
42e286202ab03712d421746db88fdbb7aacfaa4bec8a0424e3aafa9046d78719
|
|
| MD5 |
0cd5014cad6bc91923cf80fc078bca0e
|
|
| BLAKE2b-256 |
b6a428bb4bf0d1ebec9fbfbe3f56be3c70587e83e4969cb07455169b8a5ae9e0
|
File details
Details for the file llama_index_llms_asi-0.1.0-py3-none-any.whl.
File metadata
- Download URL: llama_index_llms_asi-0.1.0-py3-none-any.whl
- Upload date:
- Size: 5.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.3 CPython/3.12.3 Linux/6.8.0-1021-azure
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
14cbce7dfcb8870261a47bbe09eb01d40f3ab9e322a9a841aec54890e36b4cc5
|
|
| MD5 |
2dfc8ed2075d6e5a46159c27f231cc28
|
|
| BLAKE2b-256 |
87cbee7225060d057ced3936f1d73ad235a529def6383138d0ed5738eac6f6d3
|