Skip to main content

Autopilot enhances OpenAI interactions with concurrent task processing, faster data handling, reliable progress tracking, and transparency.

Project description

Autopilot logo

OpenAI Autopilot

PyPI - Version PyPI - Python Version

Autopilot simplifies your OpenAI interactions by concurrently processing multiple tasks, significantly speeding up data handling. It uses temporary files for checkpointing, ensuring no progress is lost during unexpected interruptions. Plus, with its real-time progress tracking, you're always informed about your task completion status, making your workflow not only faster but also more reliable and transparent.


Table of Contents

Features

  • Concurrent Processing: Efficiently handles multiple OpenAI tasks in parallel, greatly speeding up data processing.
  • Checkpointing with Temporary Files: Safely saves progress using temporary files, allowing continuation from the last checkpoint in case of interruptions.
  • Real-Time Progress Tracking: Features a progress bar that provides immediate updates on the status of ongoing tasks, enhancing transparency and planning.
  • Asynchronous Operation: Leverages Python's asyncio for non-blocking task execution, improving overall performance.
  • Error Handling: Robust exception management to ensure smooth operation under various scenarios.
  • Type Checking with Pydantic: Utilizes Pydantic for rigorous data validation and type checking, ensuring that the inputs and outputs of tasks adhere to defined schemas. This enhances the reliability and consistency of the data being processed.

How it works

The Autopilot framework streamlines asynchronous data processing with OpenAI's API. It queues data items and processes them using multiple concurrent workers, each executing a user-defined function that interacts with the OpenAI API. Importantly, if a temporary file for a task already exists, Autopilot skips reprocessing it, efficiently avoiding redundant work. As tasks are completed, results are saved in temporary files, allowing the process to resume seamlessly from the last checkpoint in case of interruptions. This setup ensures efficient handling and error management. Progress is visually tracked, and upon completion, results are compiled into a structured format for final use, simplifying the complexity of asynchronous operations.

Installation

pip install openai-autopilot

Usage

In this section, we demonstrate how to utilize the Autopilot for efficient asynchronous data processing with OpenAI's API. The process involves defining a custom asynchronous function that interfaces with OpenAI, initializing the OpenAI client, setting up the Autopilot with the processing function, and executing the process with your data. The steps are as follows:

  1. Define the Processing Function: Create an async function named process, which encapsulates the logic for interacting with OpenAI's API.

  2. Initialize OpenAI Client: Set up the AsyncOpenAI client using your OpenAI API key.

  3. Configure Autopilot: Instantiate the Autopilot class with the initialized OpenAI client and the custom processing function.

  4. Run the Process: Use the run method of the Autopilot instance, passing in your data encapsulated in AutopilotDataList.

  5. Access Processed Data: The run method returns the processed data, which can be used as needed.

This setup is designed to offer a straightforward approach to handling asynchronous tasks with OpenAI, enhancing both efficiency and ease of use.

from typing import Coroutine, List
from openai import AsyncOpenAI
from openai_autopilot import Autopilot, AutopilotMessage, AutopilotDataList

# Define the processing function, wrapping around the normal OpenAI async API usage
async def process(
    worker_id: int, client: AsyncOpenAI, data_id: int, messages: List[AutopilotMessage]
) -> Coroutine[None, None, str]:
    chat_completion = await client.chat.completions.create(
        # ...insert your configuration here...
    )

    completion_text = chat_completion.choices[0].message.content

    # Return the string to be added to the response of your data
    return completion_text

# Initialize the OpenAI asynchronous client
api_key = "...your api key..."
client = AsyncOpenAI(api_key=api_key)
autopilot = Autopilot(client=client, process_fn=process)

processed_data = autopilot.run(
    AutopilotDataList(
        # ...insert your data here...
    )
)

Configuration Options

When initializing the Autopilot class, you can configure it using the following parameters:

  1. client (AsyncOpenAI): The OpenAI client instance. It should be an instance of AsyncOpenAI. If not provided, it defaults to None.

  2. process_fn (Callable): The processing function that defines how each item in the data list is processed. It should be an asynchronous function that accepts parameters worker_id, client, data_id, and messages, and returns a coroutine with a string response. If not provided, it defaults to None.

  3. concurrency (int): Determines the number of concurrent workers that process the data. The default value is 5.

  4. tmp_dir (str): The directory path for storing temporary files used in checkpointing. Defaults to "tmp".

  5. tmp_file_prefix (str): The prefix for naming temporary files. Defaults to "data".

  6. verbose (bool): A boolean flag to enable verbose logging. Useful for debugging. Defaults to False.

These parameters allow you to customize the behavior of the Autopilot framework according to your specific requirements.

Example

In our example, we demonstrate how to use the Autopilot for processing requests with OpenAI's GPT-4 model. This example generates recipes for various food items. The results are outputted in a JSON format, providing an easy-to-read display of the generated recipes.

export OPENAI_API_KEY='...your api key...'
python examples/simple_process.py

License

openai-autopilot is distributed under the terms of the MIT license.

Acknowledgments

Thanks to ChatGPT for invaluable assistance in creating this document.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

openai_autopilot-0.0.4.tar.gz (1.5 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

openai_autopilot-0.0.4-py3-none-any.whl (7.4 kB view details)

Uploaded Python 3

File details

Details for the file openai_autopilot-0.0.4.tar.gz.

File metadata

  • Download URL: openai_autopilot-0.0.4.tar.gz
  • Upload date:
  • Size: 1.5 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: python-httpx/0.26.0

File hashes

Hashes for openai_autopilot-0.0.4.tar.gz
Algorithm Hash digest
SHA256 3e04196d260628f2a203ca23749c2b6bd3d34b38d7ee5d5112356aeeda407f31
MD5 75277dd9cebd7fbfb675ac4c6ac5bc57
BLAKE2b-256 26f458a96cd829af37e92da5e31fa37826ff417eb59482fb41444c5a14b77d61

See more details on using hashes here.

File details

Details for the file openai_autopilot-0.0.4-py3-none-any.whl.

File metadata

File hashes

Hashes for openai_autopilot-0.0.4-py3-none-any.whl
Algorithm Hash digest
SHA256 67fe0363cbdca3d1bc4d33dd35b96b2a525d7ff09a1a9472fa9c94bb4499ad0d
MD5 3f9c335d7cd764b4589cdd683560c3e7
BLAKE2b-256 88e452c7268bbe2755938af7c4c71708cd5ba837b0c9b9c4d6da4d7b906aaf35

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page