Skip to main content

Open source, type-safe primitives for multi-modal AI. All capabilities, all providers, one interface

Project description

Celeste AI

Celeste Logo

The primitive layer for multi-modal AI

All modalities. All providers. One interface.

Primitives, not frameworks.

Python License PyPI

Follow @withceleste on LinkedIn

Quick StartRequest Provider

🚀 This is the v1 Beta release. We're validating the new architecture before the stable v1.0 release. Feedback welcome!

Celeste AI

Type-safe, modality/provider-agnostic primitives.

  • Unified Interface: One API for OpenAI, Anthropic, Gemini, Mistral, and 14+ others.
  • True Multi-Modal: Text, Image, Audio, Video, Embeddings, Search —all first-class citizens.
  • Type-Safe by Design: Full Pydantic validation and IDE autocomplete.
  • Zero Lock-In: Switch providers instantly by changing a single config string.
  • Primitives, Not Frameworks: No agents, no chains, no magic. Just clean I/O.
  • Lightweight Architecture: No vendor SDKs. Pure, fast HTTP.

🚀 Quick Start

import celeste

# One SDK. Every modality. Any provider.
text   = await celeste.text.generate("Explain quantum computing", model="claude-opus-4-5")
image  = await celeste.images.generate("A serene mountain lake at dawn", model="flux-2-pro")
speech = await celeste.audio.speak("Welcome to the future", model="eleven_v3")
video  = await celeste.videos.analyze(video_file, prompt="Summarize this clip", model="gemini-3-pro")
embeddings = await celeste.text.embed(["lorep ipsum", "dolor sit amet"], model="gemini-embedding-001")

15+ providers. Zero lock-in.

Google OpenAI Mistral Anthropic Cohere xAI DeepSeek Ollama Groq ElevenLabs BytePlus Black Forest Labs

and many more

Missing a provider? Request it – ⚡ we ship fast.


Operations by Domain

Action Text Images Audio Video
Generate
Edit
Analyze
Upscale
Speak
Transcribe
Embed

✓ Available · ○ Planned

🔄 Switch providers in one line

from pydantic import BaseModel

class User(BaseModel):
    name: str
    age: int

# Model IDs
anthropic_model_id = "claude-4-5-sonnet"
google_model_id = "gemini-2.5-flash"
# ❌ Anthropic Way
from anthropic import Anthropic
import json

client = Anthropic()
response = client.messages.create(
    model=anthropic_model_id,
    messages=[
        {"role": "user",
         "content": "Extract user info: John is 30"}
    ],
    output_format={
        "type": "json_schema",
        "schema": User.model_json_schema()
    }
)
user_data = json.loads(response.content[0].text)
# ❌ Google Gemini Way
from google import genai
from google.genai import types

client = genai.Client()
response = await client.aio.models.generate_content(
    model=gemini_model_id,
    contents="Extract user info: John is 30",
    config=types.GenerateContentConfig(
        response_mime_type="application/json",
        response_schema=User
    )
)
user = response.parsed
# ✅ Celeste Way
import celeste

response = await celeste.text.generate(
    "Extract user info: John is 30",
    model=google_model_id,  # <--- Choose any model from any provider
    output_schema=User,  # <--- Unified parameter working across all providers
)
user = response.content  # Already parsed as User instance

⚙️ Advanced: Create Client

For explicit configuration or client reuse, use create_client with modality + operation. This is modality-first: you choose the output type and operation explicitly.

from celeste import create_client, Modality, Operation, Provider

client = create_client(
    modality=Modality.TEXT,
    operation=Operation.GENERATE,
    provider=Provider.OLLAMA,
    model="llama3.2",
)
response = await client.generate("Extract user info: John is 30", output_schema=User)

capability is still supported but deprecated. Prefer modality + operation.


🪶 Install

uv add celeste-ai
# or
pip install celeste-ai

🔧 Type-Safe by Design

# Full IDE autocomplete
import celeste

response = await celeste.text.generate(
    "Explain AI",
    model="gpt-4o-mini",
    temperature=0.7,    # ✅ Validated (0.0-2.0)
    max_tokens=100,     # ✅ Validated (int)
)

# Typed response
print(response.content)              # str (IDE knows the type)
print(response.usage.input_tokens)   # int
print(response.metadata["model"])     # str

Catch errors before production.


🤝 Contributing

We welcome contributions! See CONTRIBUTING.md.

Request a provider: GitHub Issues Report bugs: GitHub Issues


📄 License

MIT license – see LICENSE for details.

Get StartedDocumentationGitHub

Made with ❤️ by developers tired of framework lock-in

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

celeste_ai-0.9.2.tar.gz (191.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

celeste_ai-0.9.2-py3-none-any.whl (260.2 kB view details)

Uploaded Python 3

File details

Details for the file celeste_ai-0.9.2.tar.gz.

File metadata

  • Download URL: celeste_ai-0.9.2.tar.gz
  • Upload date:
  • Size: 191.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for celeste_ai-0.9.2.tar.gz
Algorithm Hash digest
SHA256 1a5181662e7e9bd99b43594e14b96b8d4e34658fdf82f82e9d97dad97ec7f182
MD5 b6637f83394eaf7f990b0c0dc128fa00
BLAKE2b-256 d9cd9ce0d43819468f2ac8487d0fc65f1c2d7ebccf26d357a8333a121f758f0d

See more details on using hashes here.

Provenance

The following attestation bundles were made for celeste_ai-0.9.2.tar.gz:

Publisher: publish.yml on withceleste/celeste-python

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file celeste_ai-0.9.2-py3-none-any.whl.

File metadata

  • Download URL: celeste_ai-0.9.2-py3-none-any.whl
  • Upload date:
  • Size: 260.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for celeste_ai-0.9.2-py3-none-any.whl
Algorithm Hash digest
SHA256 2e09e2794105755af6671dfb7b9f01119aaab7e2efa3336a53e375bdcac8b188
MD5 b55b43adbe1a4cf024747e2ed616da44
BLAKE2b-256 2dc50d412ecbadc423da2280deadcf5c263a1e54f61fc43c73b3b9b66227ea12

See more details on using hashes here.

Provenance

The following attestation bundles were made for celeste_ai-0.9.2-py3-none-any.whl:

Publisher: publish.yml on withceleste/celeste-python

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page