Skip to main content

Call LLMs and embedding models from a Polars DataFrame, one row at a time, using native Polars expressions. Powered by LangChain.

Project description

polars-llm

PyPI version Python versions Build status codecov License

Call OpenAI, Anthropic, and Gemini models from a Polars DataFrame, one row at a time, using native Polars expressions.

polars-llm registers an .llm namespace on Polars expressions so you can call any LangChain-supported chat model or embedding model on every row of a DataFrame — synchronously or asynchronously — and pipe the responses straight back into your data pipeline.

import polars as pl
import polars_llm  # noqa: F401  — registers the `.llm` namespace

(
    pl.DataFrame({"user_prompt": ["Summarise polars in one sentence."]})
      .with_columns(
          pl.col("user_prompt").llm.openai(model="gpt-4o-mini").alias("answer")
      )
)

Why polars-llm?

  • Expression-native — works inside with_columns, select, and any other Polars expression context. No Python for loops over rows, no notebook glue.
  • Sync and async — every provider verb has an a-prefixed async sibling that fans out concurrently with asyncio.gather and an optional max_concurrency cap.
  • Per-row prompts and system messages — both the prompt and the system message can be Polars expressions, so you can build them from other columns.
  • Structured outputs — pass a Pydantic model as schema= to get a struct column back, parsed via LangChain's with_structured_output.
  • Embeddings, tooopenai_embed and gemini_embed return List[Float64] columns ready for vector search.
  • Powered by LangChain — you get the same retries, batching, and observability primitives the rest of the LangChain ecosystem uses, plumbed straight into a DataFrame.

Common use cases:

  • Summarise, classify, translate, or extract structured fields from a column of text.
  • Score rows against a custom rubric using an LLM-as-judge.
  • Build embeddings for a corpus directly from a DataFrame, ready to write to a vector database.
  • Mix LLM calls with the rest of your pipeline (joins, filters, group-bys) without leaving Polars.

Installation

polars-llm keeps its base install light. Pick the providers you need as extras:

# Just one provider
pip install "polars-llm[openai]"
pip install "polars-llm[anthropic]"
pip install "polars-llm[gemini]"

# Or all of them
pip install "polars-llm[all]"

# uv
uv add "polars-llm[all]"

Requires Python 3.9+ and Polars 1.0+.

Authentication follows LangChain conventions — set OPENAI_API_KEY, ANTHROPIC_API_KEY, or GOOGLE_API_KEY in your environment before importing.

Quickstart

1. Chat completion per row

import polars as pl
import polars_llm  # noqa: F401

df = (
    pl.DataFrame({"user_prompt": [
        "What is the capital of Spain?",
        "What is the capital of France?",
    ]})
    .with_columns(
        pl.col("user_prompt").llm.openai(model="gpt-4o-mini").alias("answer")
    )
)

2. System prompt — literal or per-row

# Same system prompt for every row
pl.col("user_prompt").llm.anthropic(
    model="claude-sonnet-4-6",
    system="Answer in fewer than 10 words.",
)

# Per-row system prompt from another column
pl.col("user_prompt").llm.gemini(
    model="gemini-2.5-pro",
    system=pl.col("system_prompt"),
)

3. Async for throughput

The a-prefixed verbs run concurrently across the batch, capped at max_concurrency:

df.with_columns(
    pl.col("user_prompt").llm.aopenai(
        model="gpt-4o-mini",
        max_concurrency=20,
    ).alias("answer")
)

4. Structured output with Pydantic

from pydantic import BaseModel

class Sentiment(BaseModel):
    label: str  # "positive" | "neutral" | "negative"
    confidence: float

df.with_columns(
    pl.col("review").llm.openai(
        model="gpt-4o-mini",
        schema=Sentiment,
    ).alias("sentiment")
).unnest("sentiment")

5. Embeddings

df.with_columns(
    pl.col("text").llm.openai_embed(
        model="text-embedding-3-small",
    ).alias("vector")
)

6. Retries, caching, metadata

pl.col("user_prompt").llm.aanthropic(
    model="claude-sonnet-4-6",
    retries=3,
    backoff=0.5,
    max_concurrency=10,
    cache=True,            # dedupe identical prompts within a batch
    with_metadata=True,    # struct {content, elapsed_ms, error}
)

API reference

All methods live under the .llm namespace on any Polars expression that resolves to a string column.

Chat verbs

Method Provider Mode
openai / aopenai OpenAI sync / async
anthropic / aanthropic Anthropic sync / async
gemini / agemini Google Gemini sync / async

Embedding verbs

Method Provider Mode
openai_embed / aopenai_embed OpenAI Embeddings sync / async
gemini_embed / agemini_embed Google Gemini sync / async

Anthropic does not currently offer a first-party embeddings API.

Common arguments

All verbs are keyword-only and accept:

  • model (str) — model name forwarded to LangChain (e.g. "gpt-4o-mini", "claude-sonnet-4-6", "gemini-2.5-pro").
  • system (chat only) — literal string or pl.Expr for a per-row system prompt.
  • schema (chat only) — a Pydantic model class. Returns a struct column with the schema fields, via with_structured_output.
  • client — a pre-configured LangChain chat or embeddings instance (skips the in-tree constructor and is handy for advanced configuration like custom base URLs).
  • retries (int, default 0) — retry on any exception raised by the provider call.
  • backoff (float, default 0.0) — exponential backoff base (seconds).
  • max_concurrency (async only, int) — cap on in-flight requests via asyncio.Semaphore.
  • cache (bool, default False) — memoise identical inputs within a batch.
  • with_metadata (bool, default False) — return a struct column with timing and error metadata instead of just the content / vector.
  • on_error ("null" | "raise", default "null") — when with_metadata=False, what to do on errors. "null" replaces failures with None and emits a warning; "raise" re-raises immediately.
  • **model_kwargs — any additional keyword arguments forwarded to the underlying LangChain class (e.g. temperature=, max_tokens=, timeout=).

Return types

Mode Default dtype With with_metadata=True
Chat (no schema) Utf8 Struct{content: Utf8, elapsed_ms: Float64, error: Utf8}
Chat (with schema) Struct{...} matching the Pydantic model Same struct; content JSON-serialised under content
Embeddings List[Float64] Struct{vector: List[Float64], dim: Int64, elapsed_ms: Float64, error: Utf8}

Tips and patterns

  • Build prompts from columns with pl.format("Translate to {}: {}", pl.col("language"), pl.col("text")).
  • Bring your own client to share a single ChatOpenAI (with custom base_url, organization, etc.) across many calls — pass it as client=.
  • Watch the warning — when a request fails and is silently nulled, polars-llm emits a UserWarning so you don't ship a column of nulls by accident. Pass with_metadata=True to inspect per-row errors instead.
  • Combine with lazy frames — every verb is an expression, so it composes inside LazyFrame.with_columns(...).

Contributing

Contributions are welcome — see CONTRIBUTING.md. Please open an issue before starting on larger changes.

License

MIT © Diego Garcia Lozano


Inspired by and patterned after polars-api.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

polars_llm-0.1.0.tar.gz (18.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

polars_llm-0.1.0-py3-none-any.whl (11.5 kB view details)

Uploaded Python 3

File details

Details for the file polars_llm-0.1.0.tar.gz.

File metadata

  • Download URL: polars_llm-0.1.0.tar.gz
  • Upload date:
  • Size: 18.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for polars_llm-0.1.0.tar.gz
Algorithm Hash digest
SHA256 2c65b6c23368432622693805ef80ab96968237d609634d6b993e904d3e02b5bf
MD5 340c0df30d23924aac2624e95c10e437
BLAKE2b-256 fda25aa109356cabf9e04d1710f5c69dd1489e1d524369b4130e9c297f8c3397

See more details on using hashes here.

File details

Details for the file polars_llm-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: polars_llm-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 11.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for polars_llm-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 b4b7e3b46b982fa526a6e785ae549b9b6cc751b9a2b6a5ee262897128acf687d
MD5 e84035cde93e88453316eacbfbf7aaba
BLAKE2b-256 f7fe4924829287e30cc09d7557ab8c77e579475ca5a95523cefac0bb46f94720

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page