Skip to main content

A lightweight library that orchestrates LLM completions and agents in parallel workflows

Project description

Giskard Agents

Giskard Agents is a lightweight library that orchestrates LLM completions and agents in parallel workflows. It enables multiple AI workflows to run independently while maintaining coordination and synchronization between them.

Requirements

  • Python 3.12 or higher

Installation

Using uv (recommended)

Install the package:

uv add giskard-agents

For development, install with dev dependencies:

uv add giskard-agents --dev

Telemetry

This package depends on giskard-core, which may send optional usage analytics when imported or when downstream libraries record events. See Telemetry in the giskard-core README for details and how to opt out.

Core Concepts

Three basic elements to keep in mind:

  • Generator corresponds to a conversational text generator. It represents a model with certain params and can run completions.
  • ChatWorkflow defines everything needed to run a chat with the generator. It handles templates, parsing, and tools.
  • Chat is the result of a pipeline run. It contains the generated messages and metadata.

The API is async throughout — all run methods return coroutines.

Basic usage

Running a chat

from giskard import agents

generator = agents.Generator(model="openai/gpt-4o-mini")

# generator.chat automatically creates a workflow that can be run.
chat = await generator.chat("Hello, how are you?").run()

# print the content of the last message (in this case, the assistant's response)
print(chat.last.content)

You can run multiple chats in parallel:

chats = await generator.chat("Hello, how are you?").run_many(n=3)

Or add multiple messages to the workflow:

# The chat message role is "user" by default.
chat = await (
    generator
    .chat("You are a helpful assistant.", role="system")
    .chat("Hello, how are you?")
    .chat("I'm fine, thank you!", role="assistant")
    .chat("What's your name?")
    .run()
)

Retries and rate limiting

Generators have built-in support for retries and rate limiting via dedicated fields. The correct middleware ordering is handled automatically (retry wraps rate limiting wraps custom middlewares).

Retries

By default, GiskardLLMGenerator retries failed requests with exponential backoff. You can customize the retry policy:

from giskard.agents.generators.middleware import RetryPolicy

generator = agents.Generator(
    model="openai/gpt-4o-mini",
    retry_policy=RetryPolicy(max_attempts=5, base_delay=2.0, max_delay=30.0),
)

Or use the convenience method:

generator = agents.Generator(model="openai/gpt-4o-mini").with_retries(5, base_delay=2.0, max_delay=30.0)

Rate limiting

from giskard.core import MinIntervalRateLimiter

generator = agents.Generator(
    model="openai/gpt-4o-mini",
    rate_limiter=MinIntervalRateLimiter.from_rpm(60, max_concurrent=5),
)

Or use the convenience method:

generator = generator.with_rate_limiter(MinIntervalRateLimiter.from_rpm(60, max_concurrent=5))

Custom middleware

For advanced cross-cutting concerns (logging, caching, etc.), you can write custom middleware by subclassing CompletionMiddleware and adding it to the middlewares list:

import logging
from typing import Any

from giskard.agents.generators import GenerationParams
from giskard.agents.generators.middleware import CompletionMiddleware, NextFn
from giskard.llm.types import ChatMessage, CompletionResponse

@CompletionMiddleware.register("logging")
class LoggingMiddleware(CompletionMiddleware):
    async def call(
        self,
        messages: list[ChatMessage],
        params: GenerationParams | None,
        metadata: dict[str, Any] | None,
        next_fn: NextFn,
    ) -> CompletionResponse:
        logging.info(f"Sending {len(messages)} messages")
        response = await next_fn(messages, params, metadata)
        logging.info(f"Got response: {response.choices[0].finish_reason}")
        return response

generator = agents.Generator(
    model="openai/gpt-4o-mini",
    middlewares=[LoggingMiddleware()],
)

Custom middlewares run after the built-in retry and rate-limiter middleware.

Structured output

You can specify the output model for the workflow, and this will be passed to each completion call:

from pydantic import BaseModel

class SimpleOutput(BaseModel):
    mood: str
    greeting: str

chat = await (
    generator.chat("Hello!")
    .with_output(SimpleOutput)
    .run()
)

assert isinstance(chat.output, SimpleOutput)
assert chat.output.mood == "happy"

Inputs and templates

Inline templates

You can associate input variables to a workflow, and use them in the messages thanks to jinja2 templating. By default, strings passed to chat() are literal text; pass as_template=True when the string should be rendered as Jinja2.

Here's an example:

# This will run a chat with the message "Hello Test Bot, how are you?"
chat = await (
    generator.chat(
        "Hello {{ name_of_the_bot }}, how are you?", as_template=True
    )
    .with_inputs(name_of_the_bot="Test Bot")
    .run()
)

External templates

For more complicated prompts you can define your template in a separate file. First tell giskard.agents where to find the templates (you probably want to do this in your __init__.py file):

agents.set_prompts_path("path/to/the/prompts")

Write your templates in jinja2:

Hello {{ name_of_the_bot }}, how are you?
chat = await (
    generator.template("hello_template.j2")
    .with_inputs(name_of_the_bot="Test Bot")
    .run()
)

Multi-message templates

Sometimes you may want to use more complex, multi-message prompts. This is particularly useful when you need a few-shots chat that includes examples. For this need, giskard.agents provides a special syntax to define multi-message prompts.

{% message system %}
You are an impartial evaluator of scientific theories. Your only job is to rate them on a scale of 1-5, where:
1 = "This theory belongs in the same category as 'the Earth is flat'"
2 = "More holes than Swiss cheese, but at least it's creative"
3 = "Could be true, could be false, Schrödinger's theory"
4 = "Almost as solid as the theory of gravity"
5 = "This theory is so good, even the experimentalists are convinced!"

The user will provide you with a scientific theory to evaluate. Respond with ONLY a number from 1-5.
{% endmessage %}

{# Example #}
{% message user %}
The universe is actually a giant simulation running on a quantum computer in a higher dimension, and we're all just NPCs in someone's cosmic video game.
{% endmessage %}

{% message assistant %}
3
{% endmessage %}

{# Actual input #}
{% message user %}
{{ theory }}
{% endmessage %}

You can then load the template as usual:

chat = await (
    generator.template("evaluators.scientific_theory")
    .with_inputs(theory="Normandy is actually the center of the universe because its perfect balance of rain, cheese, and cider creates a quantum field that bends space-time, making it the most harmonious place on Earth.")
    .run()
)

score = int(chat.last.content)
assert score == 5

Input batches

You can run multiple chats with different inputs by passing a list of inputs to the run_batch method.

chats = await (
    generator.chat("What's the weather in {{ city }}?", as_template=True)
    .run_batch([{"city": "Paris"}, {"city": "London"}])
)
assert len(chats) == 2

Tools

You can define tools using the @agents.tool decorator. Tools will be automatically called when the workflow is run.

When defining tools, you need to make sure that all tool arguments have type hints. These will be used to define the tool schema. You must also provide a docstring, which will be used to describe the tool to the LLM. If you include the parameters in the docstring, their descriptions will be automatically added to the tool schema.

This can be combined with all functionalities described earlier. Here's an example:

from giskard import agents

@agents.tool
def get_weather(city: str) -> str:
    """Get the weather in a city.

    Parameters
    ----------
    city: str
        The city to get the weather for.
    """
    if city == "Paris":
        return f"It's raining in {city}."

    return f"It's sunny in {city}."

# Run parallel chats with tools
chats = await (
    generator.chat("Hello, what's the weather in {{ city }}?", as_template=True)
    .with_tools(get_weather)
    .run_batch([{"city": "Paris"}, {"city": "London"}])
)

assert "rain" in chats[0].last.content
assert "sun" in chats[1].last.content

Run context

Tools can access a RunContext object that acts as a storage memory for the run. This can be useful to store information that is needed for the next tool calls.

The RunContext object will be automatically passed to the tool if you specify the type hint.

@agents.tool
def get_weather(city: str, run_context: agents.RunContext) -> str:
    previously_asked_cities = run_context.get("previously_asked_cities", [])

    if city in previously_asked_cities:
        return f"I've already asked this!"

    run_context.set("previously_asked_cities", previously_asked_cities + [city])
    return f"It's raining in {city}."

The run context will be shared between all tool calls in the same run.

You can also retrieve it after the run is complete:

chat = await (
    generator.chat("Hello, what's the weather in {{ city }}?", as_template=True)
    .with_tools(get_weather)
    .with_inputs(city="Paris")
    .run()
)

assert "Paris" in chat.context.get("previously_asked_cities")

To initialize the run context, you can pass it to the workflow with the with_context method:

run_context = agents.RunContext()
run_context.set("previously_asked_cities", ["Paris"])

chat = await (
    generator.chat("Hello, what's the weather in {{ city }}?", as_template=True)
    .with_context(run_context)
    .with_tools(get_weather)
    .run()
)

Error handling

Errors during workflow execution

You can specify the error handling policy for the workflow. By default, the workflow will raise an error if an error occurs. You can change this behavior by passing the on_error method.

You can choose to:

  • Raise an error (ErrorPolicy.RAISE)
  • Return the chat with the error (ErrorPolicy.RETURN). The chat will have a failed attribute set to True, and an error attribute with a serializable error message.
  • For multi-run methods (e.g. run_many or run_batch), you can discard the failed chats (ErrorPolicy.SKIP). You will then only get the successful chats (potentially an empty list).

Note: when running a single chat (workflow.run(...)), error policy SKIP behaves as RETURN, returning a Chat object with the error.

from giskard.agents import ErrorPolicy

# This may return fewer than 3 chats if some fail.
chats = await generator.chat("Hello!", role="user").on_error(ErrorPolicy.SKIP).run_many(n=3)

# This will return 3 chats, some may be in failed state.
chats = await generator.chat("Hello!", role="user").on_error(ErrorPolicy.RETURN).run_many(n=3)

for chat in chats:
    if chat.failed:
        print("CHAT FAILED:", chat.error.message)

Errors during tool calls

By default, giskard.agents will catch errors during tool calls and return the error message as a tool result. This will let the agent decide what to do with the error (whether retrying or moving on). You can change this behavior by passing the catch=None on the tool decorator. In this case, the error will be raised and passed to the workflow, which will then handle it according to the workflow error handling policy.

# Default behavior, will catch errors
@agents.tool
def get_weather(city: str) -> str:
    raise ValueError("City not found")

result = await get_weather.run(arguments={"city": "Paris"})
print(result) # "ERROR: City not found"


# Opt out of the catch
@agents.tool(catch=None)
def get_weather(city: str) -> str:
    raise ValueError("City not found")

# This will raise an exception
result = await get_weather.run(arguments={"city": "Paris"})

Development

Quick Setup

For quick development setup, use the provided Makefile:

make setup  # Install deps + tools
make help   # See all available commands

Manual Setup

Install the project dependencies:

uv sync

Install development tools:

uv tool install ruff
uv tool install vermin
uv tool install pre-commit --with pre-commit-uv

Note: pytest and pip-audit are included in dev dependencies since they need access to the project code.

Common Tasks

make test          # Run tests
make lint          # Run linting
make format        # Format code
make check-format  # Check if code is formatted
make check         # Run all checks (lint + format + compatibility + security)
make ci            # Simulate CI locally
make clean         # Clean build artifacts

Python Compatibility

This project maintains compatibility with Python 3.12+. We use vermin to ensure code compatibility:

# Check Python 3.12 compatibility
make check-compat
# or manually:
uv tool run vermin --target=3.12- --no-tips --violations .

Setting up Pre-commit Hooks

To automatically check compatibility on every commit:

# Quick setup (if you ran `make setup` this is already done)
make pre-commit-install

# Or manually:
uv tool install pre-commit --with pre-commit-uv
pre-commit install

# Run on all files
make pre-commit-run

The hooks will now run automatically on git commit and prevent commits that don't meet Python 3.12 compatibility requirements.

Security

We use pip-audit to scan for known security vulnerabilities in dependencies:

# Check for security vulnerabilities
make security
# or manually:
uv run pip-audit

Both compatibility and security checks are automatically run in CI for every pull request.

License

This project is licensed under the MIT License - see the LICENSE file for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

giskard_agents-1.0.2b2.tar.gz (46.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

giskard_agents-1.0.2b2-py3-none-any.whl (36.2 kB view details)

Uploaded Python 3

File details

Details for the file giskard_agents-1.0.2b2.tar.gz.

File metadata

  • Download URL: giskard_agents-1.0.2b2.tar.gz
  • Upload date:
  • Size: 46.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.11.8 {"installer":{"name":"uv","version":"0.11.8","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for giskard_agents-1.0.2b2.tar.gz
Algorithm Hash digest
SHA256 6de1fbcebdc2547889e8c2cd79c7519acdb59017ea786083382798a782263ac9
MD5 4009d8f9a7cc16bb436e5e026946a526
BLAKE2b-256 6344647fb3129b4c92aa88d0ac7a0d8277ab8114c07fbd4849433d2516709943

See more details on using hashes here.

File details

Details for the file giskard_agents-1.0.2b2-py3-none-any.whl.

File metadata

  • Download URL: giskard_agents-1.0.2b2-py3-none-any.whl
  • Upload date:
  • Size: 36.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.11.8 {"installer":{"name":"uv","version":"0.11.8","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for giskard_agents-1.0.2b2-py3-none-any.whl
Algorithm Hash digest
SHA256 c22b4d106f4002c241ce32b6db2ea1270f0d8cbbe4551f84fdb19db1e68172b1
MD5 e4bd2f3d80a125a3eaa73568bace9a5a
BLAKE2b-256 ff9ddab98d01aa19d542472307c6fbcd4939ba4a97e316310cd8be132405fe4d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page