Skip to main content

Import wisdom, export code.

Project description

AutoGenLib

The only library you'll need ever.

Import wisdom, export code.

AutoGenLib is a Python library that automatically generates code on-the-fly using OpenAI's API. When you try to import a module or function that doesn't exist, AutoGenLib creates it for you based on a high-level description of what you need.

Features

  • Dynamic Code Generation: Import modules and functions that don't exist yet
  • Context-Aware: New functions are generated with knowledge of existing code
  • Progressive Enhancement: Add new functionality to existing modules seamlessly
  • No Default Caching: Each import generates fresh code for more varied and creative results
  • Full Codebase Context: LLM can see all previously generated modules for better consistency
  • Caller Code Analysis: The LLM analyzes the actual code that's importing the module to better understand the context and requirements
  • Automatic Exception Handling: All exceptions are sent to LLM to provide clear explanation and fixes for errors.

Installation

pip install autogenlib

Or install from source:

git clone https://github.com/cofob/autogenlib.git
cd autogenlib
pip install -e .

Requirements

  • Python 3.12+
  • OpenAI API key

Quick Start

Set OpenAI API key in OPENAI_API_KEY env variable.

from autogenlib import init
init("Library for generating secure random tokens")

# Import a function that doesn't exist yet - it will be automatically generated
from autogenlib.tokens import generate_token

# Use the generated function
token = generate_token(length=32)
print(token)

How It Works

  1. You initialize AutoGenLib with a description of what you need
  2. When you import a module or function under the autogenlib namespace, the library:
    • Checks if the module/function already exists
    • If not, it analyzes the code that's performing the import to understand the context
    • It sends a request to OpenAI's API with your description and the context
    • The API generates the appropriate code
    • The code is validated and executed
    • The requested module/function becomes available for use

Examples

Generate a TOTP Generator

from autogenlib import init
init("Library for generating TOTP codes")
from autogenlib.totp import totp_generator
print(totp_generator("SECRETKEY123"))

Using Context-Awareness

from autogenlib import init
init("Library for data processing")

# Define your data structure
data = [{"user": "Alice", "score": 95}, {"user": "Bob", "score": 82}]

# Import a function - AutoGenLib will see how your data is structured
from autogenlib.processor import get_highest_score

# The function will work with your data structure without you having to specify details
print(get_highest_score(data))  # Will correctly extract the highest score

Add a Verification Function Later

# Later in your application, you need verification:
from autogenlib.totp import verify_totp
result = verify_totp("SECRETKEY123", "123456")
print(f"Verification result: {result}")

Create Multiple Modules

from autogenlib import init
init("Cryptographic utility library")

# Generate encryption module
from autogenlib.encryption import encrypt_text, decrypt_text
encrypted = encrypt_text("Secret message", "password123")
decrypted = decrypt_text(encrypted, "password123")
print(decrypted)

# Generate hashing module
from autogenlib.hashing import hash_password, verify_password
hashed = hash_password("my_secure_password")
is_valid = verify_password("my_secure_password", hashed)
print(f"Password valid: {is_valid}")

Configuration

Setting the OpenAI API Key

Set your OpenAI API key as an environment variable:

export OPENAI_API_KEY="your-api-key-here"
# Optional
export OPENAI_API_BASE_URL="https://openrouter.ai/api/v1"  # Use OpenRouter API
export OPENAI_API_KEY="openai/gpt-4.1"

Or in your Python code (not recommended for production):

import os
os.environ["OPENAI_API_KEY"] = "your-api-key-here"

Caching Behavior

By default, AutoGenLib does not cache generated code. This means:

  • Each time you import a module, the LLM generates fresh code
  • You get more varied and often funnier results due to LLM hallucinations
  • The same import might produce different implementations across runs

If you want to enable caching (for consistency or to reduce API calls):

from autogenlib import init
init("Library for data processing", enable_caching=True)

Or toggle caching at runtime:

from autogenlib import init, set_caching
init("Library for data processing")

# Later in your code:
set_caching(True)  # Enable caching
set_caching(False)  # Disable caching

When caching is enabled, generated code is stored in ~/.autogenlib_cache.

Limitations

  • Requires internet connection to generate new code
  • Depends on OpenAI API availability
  • Generated code quality depends on the clarity of your description
  • Not suitable for production-critical code without review

Advanced Usage

Inspecting Generated Code

You can inspect the code that was generated for a module:

from autogenlib.totp import totp_generator
import inspect
print(inspect.getsource(totp_generator))

How AutoGenLib Uses the OpenAI API

AutoGenLib creates prompts for the OpenAI API that include:

  1. The description you provided
  2. Any existing code in the module being enhanced
  3. The full context of all previously generated modules
  4. The code that's importing the module/function (new feature!)
  5. The specific function or feature needed

This comprehensive context helps the LLM generate code that's consistent with your existing codebase and fits perfectly with how you intend to use it.

Contributing

Contributions are not welcome! This is just a fun PoC project.

License

MIT License


Note: This library is meant for prototyping and experimentation. Always review automatically generated code before using it in production environments.

Note: Of course 100% of the code of this library was generated via LLM

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

autogenlib-0.1.2.tar.gz (24.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

autogenlib-0.1.2-py3-none-any.whl (17.3 kB view details)

Uploaded Python 3

File details

Details for the file autogenlib-0.1.2.tar.gz.

File metadata

  • Download URL: autogenlib-0.1.2.tar.gz
  • Upload date:
  • Size: 24.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.5.29

File hashes

Hashes for autogenlib-0.1.2.tar.gz
Algorithm Hash digest
SHA256 f4e46ab7aa1a13bb78b84a050e285c8fbbc4c23ea5a28651e3051eddd40f266d
MD5 e1f601de1b1c6af1eb3d523b5d35a83e
BLAKE2b-256 f4fadde2bedc2ee396b5f2484a338c03b87fb800ccbfeed87ef62a9ce64e0a25

See more details on using hashes here.

File details

Details for the file autogenlib-0.1.2-py3-none-any.whl.

File metadata

  • Download URL: autogenlib-0.1.2-py3-none-any.whl
  • Upload date:
  • Size: 17.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.5.29

File hashes

Hashes for autogenlib-0.1.2-py3-none-any.whl
Algorithm Hash digest
SHA256 9d0968c4e24e4a6bc90e51aa5e5ff4d94fb14afb4a842d198230c4e6cf1d40d6
MD5 406162e2a857b25178a7dd9f4d27a94b
BLAKE2b-256 eab8e7cd285fab2fe6a5e4e5336ad050d86d03b063b6bdb0791ce342a8877942

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page