Build AI workflows and agents as fully-distributed and event-driven microservices.
Project description
🐮 Calfkit SDK
The SDK to build AI agents as composable, orchestratable microservices.
Calfkit lets you compose agents with independent services—agents, tools, workflows—that communicate asynchronously. Add agent capabilities without coordination. Scale each component independently. Stream agent outputs to any downstream listener.
Agents should work like real teams, with independent, distinct roles, async communication, and the ability to onboard new teammates or tools without restructuring the whole org. Build AI employees that integrate.
pip install calfkit
Why Calfkit?
The problem
Building agents like traditional web applications—tight coupling and synchronous API calls—creates the same scalability problems that plagued early microservices:
- Tight coupling: Changing one tool or agent breaks dependent agents and tools
- Scaling bottlenecks: All agents and tools live on one runtime, so everything must scale together
- Siloed: Agent communication models are difficult to wire into existing upstream and downstream systems
- Non-streaming: Agents do not naturally follow a livestreaming pattern, making data stream consumption difficult to manage
What Calfkit provides
Calfkit is a Python SDK that builds event-stream agents out-the-box. You get the benefits of an asynchronous, distributed system without managing the infrastructure yourself.
-
Distributed to the core: Agents aren't monoliths that just sit on top of the transportation layer. Agents are decomposed into independent services — the agent itself is a deeply distributed system.
-
Independent scaling: Each service can scale on its own based on demand.
-
Livestream agents by default: Agents already listen on event streams, so consuming data streams — realtime market feeds, IoT sensors, user activity event streams — is the native pattern, not a bolted-on integration.
-
Compose agents without coupling: Compose multi-agent teams and workflows by deploying agents on communication channels that are already tapped into the messaging stream. No extra wiring, and no editing existing code — agents don't even need to know about each other.
-
Universal data flow: Agents plug into any stream — integrate and consume from any upstream data sources and publish to downstream systems like CRMs, warehouses, or even other agents.
Quick Start
Prerequisites
- Python 3.10 or later
- Docker installed and running (for testing with a local Calfkit broker)
- LLM Provider API key
1. Install
pip install calfkit
2. Start a Calfkit Broker
Option A: Local Broker (Requires Docker)
Calfkit uses Kafka as the event broker. Run the following command to clone the calfkit-broker repo and start a local Kafka broker container:
git clone https://github.com/calf-ai/calfkit-broker && cd calfkit-broker && make dev-up
Once the broker is ready, open a new terminal tab to continue with the quickstart.
Option B: ☁️ Calfkit Cloud (In Beta)
Skip the infrastructure. Calfkit Cloud is a fully-managed broker service built for Calfkit AI agents and multi-agent teams. No server infrastructure to self-host or maintain, with built-in observability and agent-event tracing.
You will be provided a Calfkit broker API to deploy your agents instead of setting up and maintaining a broker locally.
3. Define and Deploy the Tool Node
Define and deploy a tool as an independent service. Tools are not owned by or coupled to any specific agent—once deployed, any agent in your system can discover and invoke the tool. Deploy once, use everywhere.
# weather_tool.py
import asyncio
from calfkit.nodes import agent_tool
from calfkit.client import Client
from calfkit.worker import Worker
# Define a tool — the @agent_tool decorator turns any function into a deployable tool node
@agent_tool
def get_weather(location: str) -> str:
"""Get the current weather at a location"""
return f"It's sunny in {location}"
async def main():
client = Client.connect("localhost:9092") # Connect to Kafka broker
worker = Worker(client, nodes=[get_weather]) # Initialize a worker with the tool node
await worker.run() # (Blocking call) Deploy the service to start serving traffic
if __name__ == "__main__":
asyncio.run(main())
Run the file to deploy the tool service:
python weather_tool.py
4. Deploy the Agent Node
Deploy the agent as its own service. The Agent handles LLM chat, tool orchestration, and conversation management in a single node. Import the tool definition to register it with the agent—the tool definition is reusable and does not couple the agent to the tool's deployment.
# agent_service.py
import asyncio
from calfkit.nodes import Agent
from calfkit.providers import OpenAIModelClient
from calfkit.client import Client
from calfkit.worker import Worker
from weather_tool import get_weather # Import the tool definition (reusable)
agent = Agent(
"weather_agent",
system_prompt="You are a helpful assistant.",
subscribe_topics="weather_agent.input",
model_client=OpenAIModelClient(model_name="gpt-5-nano"),
tools=[get_weather], # Register tool definitions with the agent
)
async def main():
client = Client.connect("localhost:9092") # Connect to Kafka broker
worker = Worker(client, nodes=[agent]) # Initialize a worker with the agent node
await worker.run() # (Blocking call) Deploy the service to start serving traffic
if __name__ == "__main__":
asyncio.run(main())
Set your OpenAI API key:
export OPENAI_API_KEY=sk-...
Run the file to deploy the agent service:
python agent_service.py
5. Invoke the Agent
Send a request and receive the response. The Client handles broker communication and request correlation automatically.
# invoke.py
import asyncio
from calfkit.client import Client
async def main():
client = Client.connect("localhost:9092") # Connect to Kafka broker
# Send a request and await the response
result = await client.execute_node(
"What's the weather in Tokyo?",
"agent.input", # The topic the agent subscribes to
)
print(f"Assistant: {result.output}")
if __name__ == "__main__":
asyncio.run(main())
Run the file to invoke the agent:
python invoke.py
Structured Outputs (Optional)
Agents can be deployed with a final_output_type to enforce structured output from the LLM. The output is type-safe and deserialized automatically on the client side.
from dataclasses import dataclass
from calfkit.nodes import Agent
from calfkit.providers import OpenAIModelClient
@dataclass
class WeatherReport:
location: str
summary: str
agent = Agent(
"weather_agent",
system_prompt="You are a helpful assistant.",
subscribe_topics="weather_agent.input",
model_client=OpenAIModelClient(model_name="gpt-5-nano"),
final_output_type=WeatherReport, # Enforce structured output
)
When invoking, pass the matching output_type to deserialize the response:
result = await client.execute_node(
"What's the weather in Tokyo?",
"weather_agent.input",
output_type=WeatherReport,
)
print(result.output.location) # "Tokyo"
print(result.output.summary) # "It's sunny in Tokyo"
Client-Side Features (Optional)
The Client supports multi-turn conversations, runtime dependency injection, and temporary instruction overrides—all without redeploying the agent.
Multi-turn conversations — pass the message history from a previous result to maintain context:
result = await client.execute_node("What's the weather in Tokyo?", "agent.input")
# Continue the conversation with full context
result = await client.execute_node(
"How about in Osaka?",
"agent.input",
message_history=result.message_history,
)
Runtime dependency injection — pass runtime data to tools via the deps parameter:
result = await client.execute_node(
"What's my phone number?",
"agent.input",
deps={"user_id": "usr_123"}, # Available to tools via ctx.deps.provided_deps
)
Temporary instructions — temporarily add system-level instructions scoped per request:
result = await client.execute_node(
"What's the weather in Tokyo?",
"agent.input",
temp_instructions="Always respond in Japanese.",
)
Documentation
Full documentation is coming soon. In the meantime, this README serves as the primary reference for getting started with Calfkit.
Contact
Support
If you found this project interesting or useful, please consider:
- ⭐ Starring the repository — it helps others discover it!
- 🐛 Reporting issues
- 🔀 Submitting PRs
License
This project is licensed under the Apache License 2.0. See the LICENSE file for details.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file calfkit-0.2.3.tar.gz.
File metadata
- Download URL: calfkit-0.2.3.tar.gz
- Upload date:
- Size: 437.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
70e9dc4630db3d890cb9ed15a9665ac089a977444dc408f16a4d9bff917b84cb
|
|
| MD5 |
6445f9de612cfecaa549df16bcb52241
|
|
| BLAKE2b-256 |
0036b18b4f280dcbaa1661d5e87a11a98169bc743b5966776c030727cb9ecacd
|
Provenance
The following attestation bundles were made for calfkit-0.2.3.tar.gz:
Publisher:
release.yml on calf-ai/calfkit-sdk
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
calfkit-0.2.3.tar.gz -
Subject digest:
70e9dc4630db3d890cb9ed15a9665ac089a977444dc408f16a4d9bff917b84cb - Sigstore transparency entry: 1219742173
- Sigstore integration time:
-
Permalink:
calf-ai/calfkit-sdk@8b71621d73f9fba8c552c39f75d55076871dfdd1 -
Branch / Tag:
refs/heads/main - Owner: https://github.com/calf-ai
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@8b71621d73f9fba8c552c39f75d55076871dfdd1 -
Trigger Event:
push
-
Statement type:
File details
Details for the file calfkit-0.2.3-py3-none-any.whl.
File metadata
- Download URL: calfkit-0.2.3-py3-none-any.whl
- Upload date:
- Size: 558.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
56ad9c882f04c90e450fe596784d531a916e04e6884823eb968065d9e2279029
|
|
| MD5 |
bcfeb00ec1024098086d78d9d93ae7ab
|
|
| BLAKE2b-256 |
172d5c0909e27b99439bc2a8723680fc8e851d4a6fa21db843e242b43611b272
|
Provenance
The following attestation bundles were made for calfkit-0.2.3-py3-none-any.whl:
Publisher:
release.yml on calf-ai/calfkit-sdk
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
calfkit-0.2.3-py3-none-any.whl -
Subject digest:
56ad9c882f04c90e450fe596784d531a916e04e6884823eb968065d9e2279029 - Sigstore transparency entry: 1219742182
- Sigstore integration time:
-
Permalink:
calf-ai/calfkit-sdk@8b71621d73f9fba8c552c39f75d55076871dfdd1 -
Branch / Tag:
refs/heads/main - Owner: https://github.com/calf-ai
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@8b71621d73f9fba8c552c39f75d55076871dfdd1 -
Trigger Event:
push
-
Statement type: