Skip to main content

No project description provided

Project description

PromptoGen

Bridging LLMs and Python Seamlessly.

Releases Test Package version Supported Python versions


Documentation: https://promptogen.zawakin.dev

Source Code: https://github.com/zawakin/promptogen

Getting Started: https://promptogen.zawakin.dev/getting-started/installation


PromptoGen: Achieving efficient and expandable communication with LLM

🚀 Vision

"Seamlessly bridge the gap between LLM and Python, ensuring efficient, future-ready communication."

💡 Key Challenges in LLM Libraries:

  • Lack of an ecosystem for prompt engineering, making prompt creation and sharing difficult.
  • Strong dependency on specific LLM versions, making them vulnerable to LLM updates.
  • Complex implementations, hindering customization.

🎯 Why Choose PromptoGen?

  1. Seamless Conversion between LLM I/O and Python Objects: Streamlining LLM interactions.
  2. Flexible & Unique Interface: Guaranteeing user customizability and extensibility.
  3. Future-Proof Design: Stay ahead with reduced dependency on LLM evolutions.

Compared to Other Libraries: Many are tied to specific LLM versions, lacking the adaptability that PromptoGen offers. With a dependency only on the Pydantic data class library, PromptoGen serves as the ideal bridge between LLM strings and Python objects.

🛠️ Core Features:

  • Prompt Data Class: Standardizing LLM communication and supporting prompt engineering.
  • TextLLM Interface: Independence from LLM specifics.
  • PromptFormatter Interface: High customizability for users.

🎉 Benefits:

  • 🧩 Modular & Extendable: Flexibly mix, match, and add custom components.
  • 🛡️ Future-Proof: Stand strong against new model updates.
  • 🔧 Maintainability: Ensuring easy debugging and minimal adjustments for different LLMs.

⚠️ Unsupported Features:

  • Direct LLM Communication: We prioritize efficient interfacing over direct LLM conversations.
  • Prompt Version Management: To keep things streamlined, we avoid adding versioning complexities.
  • Specific LLM Optimization: Our focus is on adaptability across LLMs rather than optimizing for any single one.

📚 Learn More

Dive deep into the documentation for a comprehensive understanding.


Requirements

Python 3.8 or above

Installation

pip install promptogen

Importing

import promptogen as pg

How to Use

Creating your Prompt

import promptogen as pg

summarizer = pg.Prompt(
    name="Text Summarizer and Keyword Extractor",
    description="Summarize text and extract keywords.",
    input_parameters=[
        pg.ParameterInfo(name="text", description="Text to be summarized"),
    ],
    output_parameters=[
        pg.ParameterInfo(name="summary", description="Summarized text"),
        pg.ParameterInfo(name="keywords", description="Extracted keywords from the text"),
    ],
    template=pg.IOExample(
        input={'text': "This is a sample text for summarization."},
        output={
            'summary': "This is a summary of the text.",
            'keywords': ["sample", "text", "summarization"],
        },
    ),
    examples=[
        pg.IOExample(
            input={
                'text': "One sunny afternoon, a group of friends opted to meet at the nearby park to indulge in various sports and activities. They engaged in soccer, badminton, and basketball, reveling in the joy of camaraderie and crafting unforgettable moments together."},
            output={
                'summary': "A bunch of friends relished an afternoon of sports and bonding at a neighborhood park.",
                'keywords': ["friends", "park", "sports", "moments"],
            },
        )
    ],
)

Saving the Prompt

summarizer.to_json_file("summarizer.json")

Loading the Prompt

import promptogen as pg

summarizer = pg.Prompt.from_json_file("summarizer.json")

For more information, please refer to Prompt.

Format your Prompt as a String

To send an instance of the Prompt class to the LLM (Large Language Model) in actual use, you need to convert it into a string. With PromptoGen, you can use pg.PromptFormatter to convert the prompt into a string in any desired format.

import promptogen as pg

summarizer = pg.Prompt(
    name="Text Summarizer and Keyword Extractor",
    # ...
)

formatter = pg.KeyValuePromptFormatter()

input_value = {
    "text": "In the realm of software engineering, ...",
}
print(formatter.format_prompt(summarizer, input_value))

Console Output:

Summarize text and extract keywords.

Input Parameters:
  - text: Text to summarize

Output Parameters:
  - summary: Summary of text
  - keywords: Keywords extracted from text

Template:
Input:
text: "This is a sample text to summarize."
Output:
summary: "This is a summary of the text."
keywords: ['sample', 'text', 'summarize']

Example 1:
Input:
text: "One sunny afternoon, a group of friends decided to gather at the nearby park to engage in various games and activities. They played soccer, badminton, and basketball, laughing and enjoying each other's company while creating unforgettable memories together."
Output:
summary: "A group of friends enjoyed an afternoon playing sports and making memories at a local park."
keywords: ['friends', 'park', 'sports', 'memories']

--------

Input:
text: "In the realm of software engineering, ..."
Output:

Parsing Outputs from Large Language Models

After receiving the prompt string as input, you obtain an output from a large language model (like GPT-3.5, GPT-4).

LLM Output:

summary: "This is a summary of the text."
keywords: ['sample', 'text', 'summarize']

You can parse this output as:

import promptogen as pg

formatter = pg.KeyValuePromptFormatter()

raw_resp = """summary: "This is a summary of the text."
keywords: ['sample', 'text', 'summarize']"""
summarized_resp = formatter.parse(summarizer, raw_resp)
print(summarized_resp)
{'summary': 'This is a summary of the text.', 'keywords': ['sample', 'text', 'summarize']}

TextLLM: Flexible LLM Integration

Through pg.TextLLM, PromptoGen achieves collaboration with a variety of large-scale language models (LLM).

import promptogen as pg

class YourTextLLM(pg.TextLLM):
    def __init__(self, model: str):
        self.model = model

    def generate(self, text: str) -> str:
        return generate_by_your_text_llm(text, self.model)

text_llm = YourTextLLM(model="your-model")

By adopting this interface, PromptoGen can seamlessly incorporate different LLMs and their versions. Users can utilize various LLMs in a consistent manner regardless of the specific LLM.

For more information, please refer to TextLLM.

PromptRunner: Execute Prompts Efficiently

pg.PromptRunner supports the execution of prompts simply and efficiently.

import promptogen as pg

# Prepare an LLM that implements the `pg.TextLLM` interface
text_llm = YourTextLLM(model="your-model")

formatter = pg.KeyValuePromptFormatter()
runner = pg.TextLLMPromptRunner(llm=text_llm, formatter=formatter)

summarizer = pg.Prompt(
    name="Text Summarizer and Keyword Extractor",
    # ...
)

input_value = {
    "text": "In the realm of software engineering, ...",
}

output_value = runner.run_prompt(summarizer, input_value)
print(output_value)

pg.PromptRunner is a key tool for making prompt execution more intuitive and efficient using PromptoGen.

For more information, please refer to PromptRunner.

Quick Start Guide

Please refer to the Quick Start Guide.

Application Examples

Refer to Application Examples.

Dependent Libraries

PromptoGen only depends on Pydantic to define the data class.

Limitations

  • With updates to PromptoGen, compatibility with prompts outputted in JSON may be lost.
  • The large language models tested for operation are OpenAI's gpt-3.5-turbo, gpt-4, and Meta's Llama 2. Other large language models have not been tested for operation. In particular, there may be cases where the parser does not work correctly, so please be cautious.

Contribution

Bug reports, proposals for new features, pull requests, etc., are all welcome! For more details, please see Contribution.

License

MIT License

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

promptogen-0.0.6.tar.gz (24.9 kB view details)

Uploaded Source

Built Distribution

promptogen-0.0.6-py3-none-any.whl (33.8 kB view details)

Uploaded Python 3

File details

Details for the file promptogen-0.0.6.tar.gz.

File metadata

  • Download URL: promptogen-0.0.6.tar.gz
  • Upload date:
  • Size: 24.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.5.1 CPython/3.11.4 Linux/5.15.0-1041-azure

File hashes

Hashes for promptogen-0.0.6.tar.gz
Algorithm Hash digest
SHA256 ec6e7451f38cbfb967bde8c47bc94313d9fb1681e42c879ff61d0e85e6e06251
MD5 ad1317e15b64bbf49a67bf9021ae9270
BLAKE2b-256 fda09bb3e56dfee288fc23a5cbebb60064082d0c0673243f48f3f837d71be211

See more details on using hashes here.

File details

Details for the file promptogen-0.0.6-py3-none-any.whl.

File metadata

  • Download URL: promptogen-0.0.6-py3-none-any.whl
  • Upload date:
  • Size: 33.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.5.1 CPython/3.11.4 Linux/5.15.0-1041-azure

File hashes

Hashes for promptogen-0.0.6-py3-none-any.whl
Algorithm Hash digest
SHA256 332ba17c5f5deecbe0092af66a29d6e8024c0b6165554f4d787335612cefb826
MD5 d858d617f4a3f2b4641341310586b059
BLAKE2b-256 73494f8b314781cf355a62f0c4be67b07112ec4fe8289cf59c2a89e41f3a079d

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page