Skip to main content

In-context Freeze-Thaw Bayesian Optimization for Hyperparameter Optimization

Reason this release was yanked:

Broken download URL. Please upgrade to >=0.4.0

Project description

arXiv

ifBO: In-context Freeze-Thaw Bayesian Optimization for Hyperparameter Optimization

Open in Spaces

This repository contains the official code for our ICML 2024 paper. ifBO is an efficient Bayesian Optimization algorithm that dynamically selects and incrementally evaluates candidates during the optimization process. It uses a model called the Freeze-Thaw surrogate (FT-PFN) to predict the performance of candidate configurations as more resources are allocated. The main branch includes the necessary API to use FT-PFN. Refer to the following sections:

To reproduce experiments from the above paper version, please refer to the branch icml-2024.

Installation

Requires Python 3.11.

pip install -U ifBO

Usage

Surrogate API

Checkout out this notebook.

Initializing the model

from ifbo.surrogate import FTPFN
from ifbo import Curve, PredictionResult

model = FTPFN(version="0.0.1")

This creates a .model/ directory in the current working directory for the surrogate model. To have control over this, specify a target_path: Path when initializing.

Supported versions:

Version Identifier Notes
0.0.1 ICML '24 submission Supports up to 1000 unique configurations in the context, with each configuration having a maximum of 10 dimensions.

Creating context and query points

The code snippet below demonstrates how to create instances of learning curves using ifbo.Curve class. Each curve represents the performance over time of a configuration (vector of hyperparameter values). These instances are used to form the context and query points for the model:

  • context: known data points with both time (t) and observed values (y).
  • query: points where predictions are needed, with only time (t) provided.

Note: All values (hyperparameters, performances, and times) must be normalized to the range $[0, 1]$.

import torch

context = [
  Curve(
    hyperparameters=torch.tensor([0.2, 0.1, 0.5]), 
    t=torch.tensor([0.1, 0.2, 0.3]), 
    y=torch.tensor([0.1, 0.15, 0.3])
  ),
  Curve(
    hyperparameters=torch.tensor([0.2, 0.3, 0.25]), 
    t=torch.tensor([0.1, 0.2, 0.3, 0.4]), 
    y=torch.tensor([0.2, 0.5, 0.6, 0.75])
  ),
]
query = [
  Curve(
    hyperparameters=torch.tensor([0.2, 0.1, 0.5]), 
    t=torch.tensor([0.3, 0.4, 0.5, 0.6, 0.7, 0.9])
  ),
  Curve(
    hyperparameters=torch.tensor([0.2, 0.3, 0.25]), 
    t=torch.tensor([0.4, 0.5, 0.6, 0.7, 0.8, 0.9])
  ),
]

Making predictions

Use the model to predict performances at the query points.

predictions: list[PredictionResult] = model.predict(context=context, query=query)

# Get predictions for the first curve
prediction: PredictionResult = predictions[0]

# Print the 5% and 95% percentiles of the predictive posterior distribution
print(prediction.quantile(0.05), prediction.quantile(0.95))

Following the PFN approach, the FT-PFN model outputs the Predictive Posterior Distribution (PPD) of the performances for each query point. Each PPD is encapsulated in an ifbo.PredictionResult object, which provides an interface to compute various quantities from the distribution, including:

  • likelihood(y_test: torch.Tensor): Computes the negative log-likelihood of the test targets (y_test).
  • ucb(): Computes the upper confidence bound.
  • ei(y_best: torch.Tensor): Computes the expected improvement over y_best.
  • pi(y_best: torch.Tensor): Computes the probability of improvement over y_best.
  • quantile(q: float): Computes the value at the specified quantile level q.

Bayesian Optimization with ifBO

To use the ifBO algorithm in practice, refer to NePS, a package for hyperparameter optimization that includes the latest and improved version of ifBO. Below is a template example of how to use ifBO with NePS. For a complete Python script, see the full example.

import neps

def training_pipeline(
    num_layers,
    num_neurons,
    epochs,
    learning_rate,
    weight_decay
):
    # Training logic and checkpoint loading here
    pass

pipeline_space = {
    "learning_rate": neps.Float(1e-5, 1e-1, log=True),
    "num_layers": neps.Integer(1, 5),
    "num_neurons": neps.Integer(64, 128),
    "weight_decay": neps.Float(1e-5, 0.1, log=True),
    "epochs": neps.Integer(1, 10, is_fidelity=True),
}

neps.run(
    pipeline_space=pipeline_space,
    run_pipeline=training_pipeline,
    searcher="ifbo",
    max_evaluations_total=50,
    step_size=1,
    surrogate_model_args=dict(
        version="0.0.1",
        target_path=None,
    ),
)

Citation

If using our surrogate, code, experiment setup, kindly cite using:

@inproceedings{
  rakotoarison-icml24,
  title={In-Context Freeze-Thaw Bayesian Optimization for Hyperparameter Optimization},
  author={H. Rakotoarison and S. Adriaensen and N. Mallik and S. Garibov and E. Bergman and F. Hutter},
  booktitle={Forty-first International Conference on Machine Learning},
  year={2024},
  url={https://openreview.net/forum?id=VyoY3Wh9Wd}
}

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ifbo-0.3.11.tar.gz (710.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

ifBO-0.3.11-py3-none-any.whl (713.4 kB view details)

Uploaded Python 3

File details

Details for the file ifbo-0.3.11.tar.gz.

File metadata

  • Download URL: ifbo-0.3.11.tar.gz
  • Upload date:
  • Size: 710.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.11.10

File hashes

Hashes for ifbo-0.3.11.tar.gz
Algorithm Hash digest
SHA256 473da139225ec2ffdbf7f92718acb049b20f0919a4e00ae9173a99134c3fc3d4
MD5 b80bf5f65b48583cbbae826aec23f4e5
BLAKE2b-256 bb312c8b7271097629d474bac7c46f9caad673e33e2d7d1fb35e84655358e366

See more details on using hashes here.

File details

Details for the file ifBO-0.3.11-py3-none-any.whl.

File metadata

  • Download URL: ifBO-0.3.11-py3-none-any.whl
  • Upload date:
  • Size: 713.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.11.10

File hashes

Hashes for ifBO-0.3.11-py3-none-any.whl
Algorithm Hash digest
SHA256 1927aae7fc52bcdab899bdef9ee7377e9f8dc4010bf2830b203d2e3748245d4d
MD5 fb096ec74e25b9ba846dc87569e3e4d7
BLAKE2b-256 0d5bcf97e109e572a3f204175dffe19cdc7101c229d335ba4c7ebe0cb6073277

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page