Skip to main content

Fairness oriented Multi-objective Optimization

Project description

Overview

Fomo is a Fairness Oriented Multiobjective Optimization toolbox for training regression and classification models. Fomo focuses on optimizing the trade-offs between various metrics in ML fairness that are often in direct conflict. The goal is to assist decision makers in weighing multiple criteria by generating good solution sets, rather than singular models.

Cite

If you use Fomo please cite the following:

  • W.G. La Cava (2023). Optimizing fairness tradeoffs in machine learning with multiobjective meta-models. GECCO '23. arXiv:2304.12190

Introduction

Fomo is designed to be incredibly general. It works with any ML model that has a scikit-learn interface (i.e. fit() and predict() methods) and takes sample weights as part of its loss function. Specifically, the fit() method should optionally take an argument, sample_weight, that provides a weight to each observation in X,y. That covers nearly all estimators in sklearn, including linear models (linear and logistic regression, lasso), SVMs, neural nets, decision trees, and ensemble methods like random forests, gradient boosting, and XGBoost.

In addition, Fomo works with many different metrics of fairness and accuracy. It currently supports:

  • Subgroup Fairness (False Positive, False Negative, and Demographic Parity)
  • Differential Fairness (Demographic Parity and Calibration)
  • Multicalibration
  • Proportional Multicalibration

In addition, users can specify any callable function they would like to be optimized, as long as it matches the call signature of these functions. Users can specify the combination of performance metrics and fairness metrics that best suit the task they are studying. You can specify any number and combinatoin of these metrics.

Finally, Fomo works with many different optimization methods available from pymoo, including NSGA-II, NSGA-III, MOEAD, and others.

Quick Start

Installation

pip install pyfomo

Dependencies

The requirements for fomo are listed in environment.yml.

Note on pymoo If you are working in linux and get a warning about pymoo, is recommended that you manually install it from the github repo rather than pip:

git clone https://github.com/anyoptimization/pymoo
cd pymoo
make compile
pip install .

Development

To install a development version, do the following:

git clone https://github.com/cavalab/fomo
cd fomo
pip install . 

Basic Usage

Here is an example of training a fair classifier on the adult income prediction dataset.

from fomo import FomoClassifier
from pmlb import pmlb
dataset = pmlb.fetch_data('adult')
X = dataset.drop('target',axis=1)
y = dataset['target']
groups = ['race','sex']
est = FomoClassifier()
est.fit(X,y, protected_features=groups)

License

See LICENSE.

Contact

  • William La Cava: william dot lacava at childrens dot harvard dot edu

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pyfomo-0.0.17.tar.gz (20.2 kB view details)

Uploaded Source

Built Distribution

pyfomo-0.0.17-py3-none-any.whl (27.1 kB view details)

Uploaded Python 3

File details

Details for the file pyfomo-0.0.17.tar.gz.

File metadata

  • Download URL: pyfomo-0.0.17.tar.gz
  • Upload date:
  • Size: 20.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.0

File hashes

Hashes for pyfomo-0.0.17.tar.gz
Algorithm Hash digest
SHA256 38f40352aa5fa2ae5c447145556d9cb557205e5ce56f7a1980f9893bd59f7b41
MD5 ef554c7e96306c2bbb7c785a469a7336
BLAKE2b-256 90afa87092c050a596e2f4642314cee9bd6b629bbfb15539480e26fc84b0c033

See more details on using hashes here.

File details

Details for the file pyfomo-0.0.17-py3-none-any.whl.

File metadata

  • Download URL: pyfomo-0.0.17-py3-none-any.whl
  • Upload date:
  • Size: 27.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.0

File hashes

Hashes for pyfomo-0.0.17-py3-none-any.whl
Algorithm Hash digest
SHA256 f2354464386ecfb385fa9752e64a801acaa82ac198a2f4acdbd05a8f45648425
MD5 ae7ab3ef2bbe09bfe311eafe07091335
BLAKE2b-256 33bbc0cd7a7e5d83847f7ddff0cd528ca6995831ede2c9d364c563bd97db3469

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page