Skip to main content

🔖 write prompts as python functions

Project description

funcchain

Version tests PyVersion License Downloads Discord GitHub Contributors GitHub Last Commit Pydantic v2 Twitter Follow

pip install funcchain

Introduction

funcchain is the most pythonic way of writing cognitive systems. Leveraging pydantic models as output schemas combined with langchain in the backend allows for a seamless integration of llms into your apps. It utilizes OpenAI Functions or LlamaCpp grammars (json-schema-mode) for efficient structured output. In the backend it compiles the funcchain syntax into langchain runnables so you can easily invoke, stream or batch process your pipelines.

Open in GitHub Codespaces

Simple Demo

from funcchain import chain
from pydantic import BaseModel

# define your output shape
class Recipe(BaseModel):
    ingredients: list[str]
    instructions: list[str]
    duration: int

# write prompts utilising all native python features
def generate_recipe(topic: str) -> Recipe:
    """
    Generate a recipe for a given topic.
    """
    return chain() # <- this is doing all the magic

# generate llm response
recipe = generate_recipe("christmas dinner")

# recipe is automatically converted as pydantic model
print(recipe.ingredients)

Complex Structured Output

from pydantic import BaseModel, Field
from funcchain import chain

# define nested models
class Item(BaseModel):
    name: str = Field(description="Name of the item")
    description: str = Field(description="Description of the item")
    keywords: list[str] = Field(description="Keywords for the item")

class ShoppingList(BaseModel):
    items: list[Item]
    store: str = Field(description="The store to buy the items from")

class TodoList(BaseModel):
    todos: list[Item]
    urgency: int = Field(description="The urgency of all tasks (1-10)")

# support for union types
def extract_list(user_input: str) -> TodoList | ShoppingList:
    """
    The user input is either a shopping List or a todo list.
    """
    return chain()

# the model will choose the output type automatically
lst = extract_list(
    input("Enter your list: ")
)

# custom handler based on type
match lst:
    case ShoppingList(items=items, store=store):
        print("Here is your Shopping List: ")
        for item in items:
            print(f"{item.name}: {item.description}")
        print(f"You need to go to: {store}")

    case TodoList(todos=todos, urgency=urgency):
        print("Here is your Todo List: ")
        for item in todos:
            print(f"{item.name}: {item.description}")
        print(f"Urgency: {urgency}")

Vision Models

from funcchain import Image
from pydantic import BaseModel, Field
from funcchain import chain, settings

# set global llm using model identifiers (see MODELS.md)
settings.llm = "openai/gpt-4-vision-preview"

# everything defined is part of the prompt
class AnalysisResult(BaseModel):
    """The result of an image analysis."""

    theme: str = Field(description="The theme of the image")
    description: str = Field(description="A description of the image")
    objects: list[str] = Field(description="A list of objects found in the image")

# easy use of images as input with structured output
def analyse_image(image: Image) -> AnalysisResult:
    """
    Analyse the image and extract its
    theme, description and objects.
    """
    return chain()

result = analyse_image(Image.open("examples/assets/old_chinese_temple.jpg"))

print("Theme:", result.theme)
print("Description:", result.description)
for obj in result.objects:
    print("Found this object:", obj)

Seamless local model support

from pydantic import BaseModel, Field
from funcchain import chain, settings

# auto-download the model from huggingface
settings.llm = "ollama/openchat"

class SentimentAnalysis(BaseModel):
    analysis: str
    sentiment: bool = Field(description="True for Happy, False for Sad")

def analyze(text: str) -> SentimentAnalysis:
    """
    Determines the sentiment of the text.
    """
    return chain()

# generates using the local model
poem = analyze("I really like when my dog does a trick!")

# promised structured output (for local models!)
print(poem.analysis)

Features

  • 🐍 pythonic
  • 🔀 easy swap between openai or local models
  • 🔄 dynamic output types (pydantic models, or primitives)
  • 👁️ vision llm support
  • 🧠 langchain_core as backend
  • 📝 jinja templating for prompts
  • 🏗️ reliable structured output
  • 🔁 auto retry parsing
  • 🔧 langsmith support
  • 🔄 sync, async, streaming, parallel, fallbacks
  • 📦 gguf download from huggingface
  • ✅ type hints for all functions and mypy support
  • 🗣️ chat router component
  • 🧩 composable with langchain LCEL
  • 🛠️ easy error handling
  • 🚦 enums and literal support
  • 📐 custom parsing types

Documentation

Checkout the docs here 👈

Also highly recommend to try and run the examples in the ./examples folder.

Contribution

You want to contribute? Thanks, that's great! For more information checkout the Contributing Guide. Please run the dev setup to get started:

git clone https://github.com/shroominic/funcchain.git && cd funcchain

./dev_setup.sh

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

funcchain-0.3.6.tar.gz (409.2 kB view details)

Uploaded Source

Built Distribution

funcchain-0.3.6-py3-none-any.whl (43.3 kB view details)

Uploaded Python 3

File details

Details for the file funcchain-0.3.6.tar.gz.

File metadata

  • Download URL: funcchain-0.3.6.tar.gz
  • Upload date:
  • Size: 409.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.9.20

File hashes

Hashes for funcchain-0.3.6.tar.gz
Algorithm Hash digest
SHA256 6137919652e8b72faf3dfe16822bce5cbcb062ba85b74632eb68ff4faefc5f34
MD5 ce6bbff67537cf630dcfe69bf8f0799f
BLAKE2b-256 1b69851578d96bb359c86c343a98a47e1154981c34f6cb8405062ee4884b3a50

See more details on using hashes here.

File details

Details for the file funcchain-0.3.6-py3-none-any.whl.

File metadata

  • Download URL: funcchain-0.3.6-py3-none-any.whl
  • Upload date:
  • Size: 43.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.9.20

File hashes

Hashes for funcchain-0.3.6-py3-none-any.whl
Algorithm Hash digest
SHA256 572e4dc1a8670a905a27a12c9a1f5dd242d47c9facd0d9d39ada366a83ec0d51
MD5 836b01899995e93cc8af04fcd61bbb42
BLAKE2b-256 34f30b462a8b92fedc0ba2c40675b17574c940d71cb019fafed417ba47694c12

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page