Skip to main content

Python client library for the AgentLab evaluation platform using Connect RPC

Project description

AgentLab Python Client

PyPI version Python 3.8+ License: MIT

A Python client library for the AgentLab evaluation platform using Connect RPC. This library provides a simple and intuitive interface for running AI agent evaluations, managing evaluators, and accessing evaluation results.

🚀 Quick Start

pip install agentlab-py

Set your API token as an environment variable:

export AGENTLAB_API_TOKEN=your-api-token-here
from agentlab import AgentLabClient, AgentLabClientOptions, CreateEvaluationOptions

# Initialize the client (automatically loads AGENTLAB_API_TOKEN), or pass the token via api_token parameter
client = AgentLabClient(AgentLabClientOptions())

# Run an evaluation
evaluation = client.run_evaluation(CreateEvaluationOptions(
    agent_name='my-agent',
    agent_version='1.0.0',
    project_id='your-project-id',
    evaluator_names=['correctness-v1'],
    user_question='What is the capital of France?',
    agent_answer='The capital of France is Paris.',
    ground_truth='Paris is the capital of France',
    metadata={'confidence': 0.95}  # Optional metadata for tracking
))

print(f"Evaluation completed: {evaluation.name}")

Retrieving Results

# Get evaluation run details
evaluation_run = client.get_evaluation_run('evaluation-run-id')

# Get structured results with parsed JSON
result_data = client.get_evaluation_result('evaluation-run-id')
print(result_data['results'])  # Parsed evaluator outputs

# Access raw evaluator results
for evaluator_name, result in evaluation_run.evaluator_results.items():
    print(f"{evaluator_name}: {result.output}")

Listing Evaluation Runs

# List recent evaluation runs
runs_response = client.list_evaluation_runs('project-123')
for run in runs_response.evaluation_runs:
    print(f"Run: {run.name} - Question: {run.user_question}")

🔧 Development

Setting up the development environment

# Clone the repository
git clone https://github.com/VectorLabsCZ/agentlab-py.git
cd agentlab-py

# Create virtual environment
python -m venv venv
source venv/bin/activate  # On Windows: venv\Scripts\activate

# Install development dependencies
pip install -r requirements-dev.txt

# Install the package in development mode
pip install -e .

Running Examples

# Basic usage example
python examples/basic_usage.py

# Async usage example
python examples/async_usage.py

Running Tests

# Run tests
pytest

# Run tests with coverage
pytest --cov=agentlab

# Run type checking
mypy agentlab/

📦 Building and Publishing

Building the package

# Install build tools
pip install build

# Build the package
python -m build

Publishing to PyPI

# Install twine
pip install twine

# Upload to PyPI
twine upload dist/*

🌟 Examples

Complete Evaluation Workflow

import json
from agentlab import AgentLabClient, AgentLabClientOptions, CreateEvaluationOptions

def main():
    # Initialize client
    client = AgentLabClient(AgentLabClientOptions(
        api_token='your-api-token'
    ))
    
    project_id = 'your-project-id'
    
    try:
        # 1. List available evaluators
        print("📋 Available evaluators:")
        evaluators = client.list_evaluators(project_id)
        for evaluator in evaluators.evaluators[:3]:  # Show first 3
            print(f"  - {evaluator.name}: {evaluator.display_name}")
        
        # 2. Run evaluation
        print("\n🚀 Running evaluation...")
        evaluation = client.run_evaluation(CreateEvaluationOptions(
            agent_name='demo-agent',
            agent_version='1.0.0',
            project_id=project_id,
            evaluator_names=['correctness-v1'],
            user_question='What is the square root of 16?',
            agent_answer='The square root of 16 is 4.',
            ground_truth='4',
            metadata={'confidence': 1.0}  # Additional context/scores
        ))
        
        # 3. Get results
        print(f"\n✅ Evaluation completed: {evaluation.name}")
        result_data = client.get_evaluation_result(evaluation.name)
        
        print("\n📊 Results:")
        print(json.dumps(result_data['results'], indent=2))
        
        # 4. List recent runs
        print("\n📈 Recent evaluation runs:")
        runs = client.list_evaluation_runs(project_id)
        for run in runs.evaluation_runs[:3]:  # Show first 3
            print(f"  - {run.name}: {run.user_question[:50]}...")
            
    except Exception as e:
        print(f"❌ Error: {e}")

if __name__ == '__main__':
    main()

Error Handling

from agentlab import AgentLabClient, AgentLabError, AuthenticationError, APIError

try:
    client = AgentLabClient(AgentLabClientOptions(api_token='invalid-token'))
    evaluation = client.run_evaluation(options)
    
except AuthenticationError as e:
    print(f"Authentication failed: {e}")
    
except APIError as e:
    print(f"API error (status {e.status_code}): {e}")
    
except AgentLabError as e:
    print(f"AgentLab error: {e}")
    
except Exception as e:
    print(f"Unexpected error: {e}")

🆚 Comparison with JavaScript SDK

Feature Python JavaScript
Installation pip install agentlab-py npm install @vectorlabs/agentlab-js
Import from agentlab import AgentLabClient const { AgentLabClient } = require('@vectorlabs/agentlab-js')
Async Support ✅ Built-in support ✅ Native async/await
Type Safety ✅ Full type hints ✅ TypeScript definitions
Error Handling ✅ Custom exception classes ✅ Custom error classes

🤝 Contributing

We welcome contributions! Please see our Contributing Guide for details.

Issues and Feature Requests

📄 License

This project is licensed under the MIT License - see the LICENSE file for details.

🔗 Links

🏢 About VectorLabs

AgentLab is developed by VectorLabs, a company focused on advancing AI agent evaluation and development tools.


Made with ❤️ by the VectorLabs team

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agentlab_py-0.3.0.tar.gz (59.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agentlab_py-0.3.0-py3-none-any.whl (51.5 kB view details)

Uploaded Python 3

File details

Details for the file agentlab_py-0.3.0.tar.gz.

File metadata

  • Download URL: agentlab_py-0.3.0.tar.gz
  • Upload date:
  • Size: 59.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.7

File hashes

Hashes for agentlab_py-0.3.0.tar.gz
Algorithm Hash digest
SHA256 445e373d3b4d080e9fde096b61f9a856bdb65864cf84c24adaa181812c66f832
MD5 68784fe5249679e2bdfefce7dfc39ea4
BLAKE2b-256 320acfd68c2998d6de37eb4a5deccafba88b724a92e7c62cfaf5139d42fb71b7

See more details on using hashes here.

File details

Details for the file agentlab_py-0.3.0-py3-none-any.whl.

File metadata

  • Download URL: agentlab_py-0.3.0-py3-none-any.whl
  • Upload date:
  • Size: 51.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.7

File hashes

Hashes for agentlab_py-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 64ffc196a8f1a6ca9fda60761a086d156cbf0174e55e1fee4d24481d43fe0ce8
MD5 b9d844c995d6f322f25039fc5264fcad
BLAKE2b-256 ea27a219d7e94e3ccd770d4b891cf09fc94528800ae30a67e55e125b309eb077

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page