Skip to main content

An implementation of the BatchBALD algorithm

Project description

BatchBALD Redux

Clean reimplementation of "BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning"

For an introduction & more information, see http://batchbald.ml/. The paper can be found at http://arxiv.org/abs/1906.08158.

The original implementation used in the paper is available at https://github.com/BlackHC/BatchBALD.

We are grateful for fastai's nbdev which is powering this package.

For more information, explore the sections and notebooks in the left-hand menu. The code is available on https://github.com/BlackHC/batchbald_redux, and the website on https://blackhc.github.io/batchbald_redux.

Install

pip install batchbald_redux

Motivation

BatchBALD is an algorithm and acquisition function for Active Learning in a Bayesian setting using BNNs and MC dropout.

The aquisition function is the mutual information between the joint of a candidate batch and the model parameters $\omega$:

{% raw %} $$a_{\text{BatchBALD}}((y_b)_B) = I[(y_b)_B;\omega]$$ {% endraw %}

The best candidate batch is one that maximizes this acquisition function.

In the paper, we show that this function satisfies sub-modularity, which provides us an optimality guarantee for a greedy algorithm. The candidate batch is selected using greedy expansion.

Joint entropies are hard to estimate and, for everything to work, one also has to use consistent MC dropout, which keeps a set of dropout masks fixed while scoring the pool set.

To aid reproducibility and baseline reproduction, we provide this simpler and clearer reimplementation.

Please cite us

@misc{kirsch2019batchbald,
    title={BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning},
    author={Andreas Kirsch and Joost van Amersfoort and Yarin Gal},
    year={2019},
    eprint={1906.08158},
    archivePrefix={arXiv},
    primaryClass={cs.LG}
}

How to use

We provide a simple example experiment that uses this package here.

To get a candidate batch using BatchBALD, we provide a simple API in batchbald_redux.batchbald:

get_batchbald_batch[source]

get_batchbald_batch(log_probs_N_K_C:Tensor, batch_size:int, num_samples:int, dtype=None, device=None)

We also provide a simple implementation of consistent MC dropout in batchbald_redux.consistent_mc_dropout.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

batchbald_redux-2.0.5.tar.gz (17.7 kB view details)

Uploaded Source

Built Distribution

batchbald_redux-2.0.5-py3-none-any.whl (16.7 kB view details)

Uploaded Python 3

File details

Details for the file batchbald_redux-2.0.5.tar.gz.

File metadata

  • Download URL: batchbald_redux-2.0.5.tar.gz
  • Upload date:
  • Size: 17.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.6.1 requests/2.25.1 setuptools/51.0.0.post20201207 requests-toolbelt/0.9.1 tqdm/4.55.1 CPython/3.8.3

File hashes

Hashes for batchbald_redux-2.0.5.tar.gz
Algorithm Hash digest
SHA256 21add2fa90233073d15048a72b7ca01127233b0e1aa2f258f9653ab74377a3cc
MD5 8eadc6ecf3b47b80069e2cb273ca5834
BLAKE2b-256 77a138f193a66a600c868b98aa67ff2ec6fcf38b6bbfbe71e748a79f6d859fca

See more details on using hashes here.

File details

Details for the file batchbald_redux-2.0.5-py3-none-any.whl.

File metadata

  • Download URL: batchbald_redux-2.0.5-py3-none-any.whl
  • Upload date:
  • Size: 16.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.6.1 requests/2.25.1 setuptools/51.0.0.post20201207 requests-toolbelt/0.9.1 tqdm/4.55.1 CPython/3.8.3

File hashes

Hashes for batchbald_redux-2.0.5-py3-none-any.whl
Algorithm Hash digest
SHA256 d971fd63d4358f5350ca79ddd057dbb6dfe5af7e8a0dc7f3829f6aeba72132ed
MD5 58ffd2b96beee4d96028b3c0110ff04c
BLAKE2b-256 17873fe347be2e834908071820f40b93f8997107668818177cf5ac09d4e186c5

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page