Skip to main content

Keiro client — call the EB1 multi-model ensemble API.

Project description

Keiro

EB1 multi-model ensemble inference. Run multiple frontier models in parallel and synthesize the best response.

Quick start

pip install keiro
keiro setup
from keiro import models

print(models("eb1-preview", "What is machine learning?"))

How it works

EB1 sends your prompt to multiple frontier models (Claude, GPT, Gemini) in parallel, then a judge synthesizes the strongest elements into a single response. The result is more accurate and more complete than any individual model.

Models

Model Description
eb1 (default) 3-model ensemble with synthesis
eb1-preview Preview ensemble (GPT-5.2, Gemini, Claude)
eb1-pro 4-model ensemble for harder tasks
claude-opus-4-6 Direct passthrough (no ensemble)
gpt-5.2 Direct passthrough
from keiro import models

# Default ensemble
answer = models("eb1", "Solve this step by step: what is 23 * 47?")

# Specific model
answer = models("claude-opus-4-6", "Write a haiku")

Prompt-first API

from keiro import models

reply = models.response("eb1-preview", "Explain quantum computing in one paragraph.")
print(reply.text)

creative = models.instance("eb1-preview", temperature=0.8)
print(creative("Write a limerick about debugging."))

for chunk in models.stream("eb1-preview", "Draft a launch email."):
    print(chunk, end="")

complete(...) is still available as the smallest one-liner, but models is the preferred external interface because it matches Ember's public prompt-first API more closely.

Configuration

Interactive setup (recommended):

keiro setup

This validates your API key against the gateway and saves credentials to ~/.keiro/credentials.

Environment variables:

export KEIRO_API_KEY="your-api-key"
export KEIRO_BASE_URL="http://54.202.103.124:8080"  # optional

Explicit arguments:

from keiro import ModelsAPI

models = ModelsAPI(api_key="your-key", base_url="http://54.202.103.124:8080")
print(models("eb1-preview", "Hello"))

Precedence: explicit arguments > environment variables > credentials file.

Requirements

  • Python 3.11+
  • No GPU required (inference runs on Keiro's hosted infrastructure)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

keiro-0.5.1.tar.gz (34.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

keiro-0.5.1-py3-none-any.whl (38.4 kB view details)

Uploaded Python 3

File details

Details for the file keiro-0.5.1.tar.gz.

File metadata

  • Download URL: keiro-0.5.1.tar.gz
  • Upload date:
  • Size: 34.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.2

File hashes

Hashes for keiro-0.5.1.tar.gz
Algorithm Hash digest
SHA256 b9bfd07a9196d6598fd10ba46791695a51992782ea66b5ed51c2479c92618264
MD5 6cdb18ce0f2af8ad9c6ac561132a363d
BLAKE2b-256 a8b683596ffe19a079eb9b470f01311089e95f9bf8eda678912c143c3a7b1f32

See more details on using hashes here.

File details

Details for the file keiro-0.5.1-py3-none-any.whl.

File metadata

  • Download URL: keiro-0.5.1-py3-none-any.whl
  • Upload date:
  • Size: 38.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.2

File hashes

Hashes for keiro-0.5.1-py3-none-any.whl
Algorithm Hash digest
SHA256 aa9e62e95be828641f19b3682c2c4af79efc4e70d45f09f93d887cf8bdf2eafd
MD5 1ae02f963a17d4931d37af085131af09
BLAKE2b-256 3dc66ea4f8bb692a85cccf3756f43169e8b6e1dc8cedb52d2901a6a40fe8dbf2

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page