Skip to main content

Bayesian regression for low-noise data using POPS algorithm

Project description

POPSRegression

Linear regression scheme from the paper

Parameter uncertainties for imperfect surrogate models in the low-noise regime

TD Swinburne and D Perez, Machine Learning: Science and Technology 2025

@article{swinburne2025,
	author={Swinburne, Thomas and Perez, Danny},
	title={Parameter uncertainties for imperfect surrogate models in the low-noise regime},
	journal={Machine Learning: Science and Technology},
	doi={10.1088/2632-2153/ad9fce},
	year={2025}
}

Installation

There will be a PR on scikit-learn soon, but in the meantime

pip install POPSRegression

What is POPSRegression?

Bayesian regression for low-noise data (vanishing aleatoric uncertainty).

Fits the weights of a regression model using BayesianRidge, then estimates weight uncertainties (sigma_ in BayesianRidge) accounting for model misspecification using the POPS (Pointwise Optimal Parameter Sets) algorithm [1]. The alpha_ attribute which estimates aleatoric uncertainty is not used for predictions as correctly it should be assumed negligable.

Bayesian regression is often used in computational science to fit the weights of a surrogate model which approximates some complex calculation. In many important cases the target calculation is near-deterministic, or low-noise, meaning the true data has vanishing aleatoric uncertainty. However, there can be large misspecification uncertainty, i.e. the model weights are instrinsically uncertain as the model is unable to exactly match training data.

Existing Bayesian regression schemes based on loss minimization can only estimate epistemic and aleatoric uncertainties. In the low-noise limit, weight uncertainties (sigma_ in BayesianRidge) are significantly underestimated as they only account for epistemic uncertainties which decay with increasing data. Predictions then assume any additional error is due to an aleatoric uncertainty (alpha_ in BayesianRidge), which is erroneous in a low-noise setting. This has significant implications on how uncertainty is propagated using weight uncertainties.

Example usage

Here, usage follows sklearn.linear_model, inheriting BayesianRidge

After running BayesianRidge.fit(..), the alpha_ attribute is not used for predictions.

The sigma_ matrix still contains epistemic weight uncertainties, whilst misspecification_sigma_ contains the POPS uncertainties.

from POPSRegression import POPSRegression

X_train,X_test,y_train,y_test = ...

# Sobol resampling of hypercube with 1.0 samples / training point
model = POPSRegression(resampling_method='sobol',resample_density=1.)

# fit the model, sample POPS hypercube
model.fit(X_train,y_train)

# Return mean and hypercube std
y_pred, y_std = model.predict(X_test,return_std=True)

# can also return max/min 
y_pred, y_std, y_max, y_min = model.predict(X_test,return_std=True,return_bounds=True)

# can also return the epistemic uncertainty seperately
y_pred, y_std, y_max, y_min, y_epistemic_std = model.predict(X_test,return_std=True,return_bounds=True,return_epistemic_std=True)

Toy example and MACE foundation model

Extreme low-dimensional case, fitting N data points to a quartic polynomial (P=5 parameters) to a complex oscillatory function. Green: two sigma of sigma_ weight uncertainty from Bayesian Regression (i.e. without alpha_ term for aleatoric error) Orange: two sigma of sigma_ and misspecification_sigma_ posterior from POPS Regression Gray: min-max of posterior from POPS Regression

As can be seen, the final error bars give very good coverage of the test output

As a more realistic example, below shows POPS applied to a P=256 linear corrector to the MACE-MP-0 foundation model. Following the POPS paper, the left panel shows the true test error histogram (black) and the predicted test error (green). The right panel shows the probabilty that the true test error lies outside of the POPS max/min bounds as a function of N/P.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

popsregression-0.3.1.tar.gz (10.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

POPSRegression-0.3.1-py3-none-any.whl (10.1 kB view details)

Uploaded Python 3

File details

Details for the file popsregression-0.3.1.tar.gz.

File metadata

  • Download URL: popsregression-0.3.1.tar.gz
  • Upload date:
  • Size: 10.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.6

File hashes

Hashes for popsregression-0.3.1.tar.gz
Algorithm Hash digest
SHA256 e79bba0391fa3e34080e194134f1e8260fd59e000a2fd68806a4e9c77da1b054
MD5 5fe2e3c8eb5b4528826af67ebf57e664
BLAKE2b-256 2bf3a370c60a6f387ac82ad8b222831b28a971b8b3a8dff9ab70aac54be89847

See more details on using hashes here.

File details

Details for the file POPSRegression-0.3.1-py3-none-any.whl.

File metadata

  • Download URL: POPSRegression-0.3.1-py3-none-any.whl
  • Upload date:
  • Size: 10.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.6

File hashes

Hashes for POPSRegression-0.3.1-py3-none-any.whl
Algorithm Hash digest
SHA256 1888181f4bf15fd4ada34c5d3c1be8ae8f3d4574c44413e3698a691c60071552
MD5 075b516c8b4e6b2ae18024fe3cf1d595
BLAKE2b-256 1ca2d0df2eeba8165c5d5e9622f087aea030f3772e18175c252239158e593e14

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page