The Extreme Value Machine
Project description
This package contains source code for the Extreme Value Machine (EVM). The EVM is a model to compute probabilities for samples to belong to a certain class. In opposition to other closed-set classification models, the EVM enables open-set classification, i.e., test samples might be of classes that are not contained in the training set. Hence, the probability of this test sample should be low for any of the classes.
The EVM was introduced in the paper The Extreme Value Machine. If you are using the EVM in a scientific publication, please cite:
@article{rudd2018evm,
author={Rudd, Ethan M. and Jain, Lalit P. and Scheirer, Walter J. and Boult, Terrance E.},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)},
title={The Extreme Value Machine},
year={2018},
volume={40},
number={3},
pages={762-768},
}
Installation
Pip
This package is on PyPI, so you can simply call:
pip install evm
to install this package. All required dependencies will be automatically installed, too.
Local build
The EVM package is using the setuptools installation package. In order to compile the package locally, you can use:
python setup.py build
which will install this package into the build/lib directory. You can also call:
python setup.py install
to install the package system-wide.
Usage
This package provides mainly two classes, for different purposes.
Training a Single EVM
The first class is the EVM class, which holds all information that is required to model data from a specific class of your data. This is done by computing distances from all instances of your class (positives) to instances of other classes (negatives).
First, you need to create an object of the EVM class. Here, you need to specify several parameters:
tailsize : An integral parameter that defines, how many negative samples are used to estimate the model parameters.
cover_threshold : If not None (the default), specifies the probability threshold used to eliminate extreme vectors if they are covered by other extreme vectors with that probability. If None, no model reduction will be performed.
distance_multiplier : The multiplier to compute margin distances. 0.5 by default.
distance_function : The distance function used to compute the distance between two samples; defaults to scipy.spatial.distance.cosine
include_cover_probabilities : (experimental) include cover probabilities that are calculated between samples of the same class. At least three samples per class are required to be able to compute cover probabilities.
log_level : defines the verbosity of the EVM training. Possible values: 'info' (the default), 'debug'.
import EVM, numpy, scipy
evm = EVM.EVM(tailsize=10, cover_threshold = 0.7, distance_function=scipy.spatial.distance.euclidean)
After creating an EVM object, it needs to be trained. You can train the model in two different ways. In the most commonly used way, you pass both positive and negative data samples of your class and other classes, respectively:
class1 = numpy.random.normal((0,0),3,(50,2))
class2 = numpy.random.normal((-10,10),3,(50,2))
class3 = numpy.random.normal((10,-10),3,(50,2))
evm.train(positives = class1, negatives = numpy.concatenate((class2, class3)))
Alternatively, you can also pre-compute the distances between all positive and all negative samples and pass the distance matrix in:
distances = scipy.spatial.distance.cdist(class1, numpy.concatenate((class2, class3)), 'euclidean')
evm.train(positives = class1, distances = distances)
Now, you can compute the probability of any data point to belong to this class (actually, the function expects a list of points, so here we have to wrap it in square brackets):
probabilities = evm.probabilities([[0,0]])[0]
which will return the probability of inclusion for each of the extreme vectors. Alternatively, you might be interested in the maximum probability, i.e., the probability that the test sample belongs to your class. You can call:
probability, evm_index = evm.max_probabilities([[0,0]])
which will return the maximum probability over all extreme vectors, and the index of the extreme vectors that was the maximum.
Training Multi-Class EVMs
The second class that you can use to train EVMs is the MultipleEVM class. For a given set of samples of several classes, it will compute an EVM model for each of the classes, taking all other classes as negatives. The parameters are similar to the EVM class.
mevm = EVM.MultipleEVM(tailsize=10, cover_threshold = 0.7, distance_function=scipy.spatial.distance.euclidean)
mevm.train([class1, class2, class3])
You can obtain the trained EVM models for each of the classes separately, if you want:
evm1 = mevm.evms[0]
evm2 = mevm.evms[1]
evm3 = mevm.evms[2]
For a given test sample, you can compute the probabilities for all extreme vectors in all EVMs by calling:
probabilities = mevm.probabilities([[0,0]])[0]
Similarly, you can compute the class with the maximum probability:
probabilities, indexes = mevm.max_probabilities([[0,0]])
where indexes contain both the index of the maximum class, and the index of the extreme vector inside that class.
Parallelism
Any of the public API functions to train or test EVM or MultipleEVM have a parallel parameter, where you can specify the maximum number of parallel threads to compute the function.
Example
In the root directory of this package, there is an example.py script that will show an exemplary usage of the EVM and MultipleEVM classes in order to display probabilities obtained on a sample 2D dataset. You can run the example via
python example.py
It might run for some time, and it will create a multi-page PDF file containing 4 plots:
Training a single EVM class from samples, and showing the extreme vectors.
Provide the probability of the class for inputs in a grid in color coding.
Training multiple EVM classes from the same samples, and showing the extreme vectors.
Provide the maximum probability of the classes for inputs in a grid in color coding.
In the color coding, black means that none of the classes had a high probability. Since most of the regions are black, this indicates that the EVM has learned to classify unknown samples as unknown.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file EVM-0.1.2.zip
.
File metadata
- Download URL: EVM-0.1.2.zip
- Upload date:
- Size: 21.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ec8b36adbed3d326496b59e45e075d8c5a0fad0a093481a8b48ef88145a8b6ef |
|
MD5 | aac43e931e4ed786fe85cd9f664d79a1 |
|
BLAKE2b-256 | a5ee187044f9158fcfd4cf808a5ca58115a987f79228b76bf230e1abac0c2636 |