Skip to main content

Add your description here

Project description

Experiment Configuration Agent for AutoGluon

This agent uses a Large Language Model to recommend optimal configurations for AutoGluon's TabularPredictor based on your machine learning problem context. By providing details about your domain, use case, and dataset, the agent will generate a set of TabularPredictor parameters designed to optimize for performance and efficiency.

Features

  • Intelligent Configuration: Leverages LLMs to recommend eval_metric, presets, time_limit, and ensembling parameters.
  • Context-Aware: Considers the business domain, specific use case, ML methodology (e.g., classification, regression), and dataset characteristics.
  • Flexible Backend: Powered by sfn-blueprint, allowing for a configurable LLM backend.
  • Multiple Scenarios: Provides recommendations for different optimization goals, such as maximizing accuracy, balancing performance and speed, or fast prototyping.

Installation

This project uses uv for dependency management and requires Python 3.10 or higher.

  1. Clone the repository:

    git clone https://github.com/stepfnAI/experiment_config_agent.git
    cd experiment-configuration-agent
    
  2. Set up the environment and install dependencies: It is recommended to use a virtual environment. uv can create one for you.

    # Create a virtual environment and install dependencies
    uv sync --extra dev
     source .venv/bin/activate
    

Usage

Basic usage

python ./examples/basic_usage.py

To get a configuration recommendation, instantiate the AutoGluonConfigAgent and pass a dictionary containing the problem context.

  1. Create a .env file in the project root to configure the LLM provider. See the Configuration section for more details.

    PROVIDER="openai"
    MODEL="gpt-4-turbo"
    # Add your API key, e.g., OPENAI_API_KEY="sk-..."
    
  2. Create your Python script:

    from experiment_configuration_agent.agent import AutoGluonConfigAgent
    
    # 1. Define the problem context
    task_data = {
        "domain": {
            "name": "Manufacturing",
            "description": "An automotive parts manufacturing facility with multiple production lines."
        },
        "use_case": {
            "name": "Predictive Maintenance",
            "description": "Detect unusual temporal patterns in sensor data to predict equipment failure and prevent breakdowns."
        },
        "methodology": "binary_classification",
        "dataset_insights": {
            "num_samples": 5000,
            "num_features": 10,
            "target": {
                "name": "failure_flag",
                "imbalance_ratio": 0.05 # Highly imbalanced
            },
            "feature_summary": {
                "sensor_A": {"min": 0.1, "max": 100.5, "dtype": "float"},
                "production_line_id": {"unique_count": 3, "dtype": "category"}
            }
        }
    }
    
    # 2. Initialize the agent
    agent = AutoGluonConfigAgent()
    
    # 3. Get the configuration recommendation
    result = agent(task_data)
    
    # 4. Print the result
    print("Recommended AutoGluon Configuration:")
    print(result.get("configuration"))
    print("\nCost Summary:")
    print(result.get("cost_summary"))
    

Configuration

The agent is configured via environment variables, which can be placed in a .env file in the project root. The primary configurations are inherited from the GluonConfig class.

  • PROVIDER: The LLM provider to use (e.g., "openai", "anthropic").
  • MODEL: The specific model to use (e.g., "gpt-4-turbo", "claude-3-opus-20240229").
  • TEMPERATURE: The model's temperature setting (e.g., 0.3).
  • MAX_TOKENS: The maximum number of tokens for the response (e.g., 4000).

You will also need to set the API key for your chosen provider, for example OPENAI_API_KEY="your-key-here".

Testing

This project uses pytest. To run the test suite, execute the following command from the project root:

pytest

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

experiment_configuration_agent-0.1.7.tar.gz (10.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

File details

Details for the file experiment_configuration_agent-0.1.7.tar.gz.

File metadata

File hashes

Hashes for experiment_configuration_agent-0.1.7.tar.gz
Algorithm Hash digest
SHA256 dfa15a2335be9974ed17a08a8797fe5d50b21723efd4ddb04ec61344776e0b71
MD5 ae2536717be992dbc9168dc87f7d9ea5
BLAKE2b-256 46372681374e20f2bf042a71157dad03ccd6cf77fc96a067eb8ffc88fc9ab66a

See more details on using hashes here.

File details

Details for the file experiment_configuration_agent-0.1.7-py3-none-any.whl.

File metadata

File hashes

Hashes for experiment_configuration_agent-0.1.7-py3-none-any.whl
Algorithm Hash digest
SHA256 0958d4098f8fad44d5089a808268a45bd2e06288cf193ce6b71e863e1648eee9
MD5 1a9b1a5948328bbd4368800b1614ceaf
BLAKE2b-256 03cad1dd4f363f9d1a557ffe36ea1cac749ef3e9eaf3e2a75d3676dc4eed02c9

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page