Pairwise difference learning library is a scikit learn compatible library for learning from pairwise differences.
Project description
Pairwise difference learning library (pdll)
Pairwise Difference Learning (PDL) library is a python module. It contains a scikit-learn compatible implementation of PDL Classifier, as described in Belaid et al. 2024
PDL Classifier or PDC is a meta learner that can reduce multiclass classification problem into a binary classification problem (similar/different).
Installation
To install the package, run the following command:
pip install -U pdll
Usage
from pdll import PairwiseDifferenceClassifier
from sklearn.ensemble import RandomForestClassifier
from sklearn.datasets import make_blobs
# Generate random data with 2 features, 10 points, and 3 classes
X, y = make_blobs(n_samples=10, n_features=2, centers=3, random_state=0)
pdc = PairwiseDifferenceClassifier(estimator=RandomForestClassifier())
pdc.fit(X, y)
print('score:', pdc.score(X, y))
y_pred = pdc.predict(X)
proba_pred = pdc.predict_proba(X)
Please consult examples/
directory for more examples.
How does it work?
The PDL algorithm works by transforming the multiclass classification problem into a binary classification problem. The algorithm works as follows:
Example 1: Graphical abstract
Example 2: PDC trained on the Iris dataset
Clic to show
We provide a minimalist classification example using the Iris dataset. The dataset is balanced, so the prior probabilities of each of the 3 classes are equal: p(Setosa) = p(Versicolour) = p(Virginica) = 1/3Three Anchor Points
- Flower 1:
y1 = Setosa
- Flower 2:
y2 = Versicolour
- Flower 3:
y3 = Virginica
One Query Point
- Flower Q:
yq
(unknown target)
Pairwise Predictions The model predicts the likelihood that both points have a similar class:
- g_sym(Flower Q, Flower 1) = 0.6
- g_sym(Flower Q, Flower 2) = 0.3
- g_sym(Flower Q, Flower 3) = 0.0
Given the above data, the first step is to update the priors.
Posterior using Flower 1:
- p_post,1(Setosa) = 0.6
- p_post,1(Versicolour) = (1/3 * (1 - 0.6)) / (1 - 1/3) = 0.2
- p_post,1(Virginica) = (1/3 * (1 - 0.6)) / (1 - 1/3) = 0.2
Similarly, we calculate for anchors 2 and 3:
-
p_post,2(Setosa) = 0.35
-
p_post,2(Versicolour) = 0.30
-
p_post,2(Virginica) = 0.35
-
p_post,3(Setosa) = 0.5
-
p_post,3(Versicolour) = 0.5
-
p_post,3(Virginica) = 0.0
Averaging over the three predictions:
Finally, the predicted class is the most likely prediction:
ŷq = arg max{y ∈ Y} p_post(y) = Setosa
Evaluation
To reproduce the experiment of the paper, please run run_benchmark.py
with a base learner and a dataset number, between 0 and 99. Example:
python run_benchmark.py --model DecisionTreeClassifier --data 0
Scores will be stored in ./results/tmp/
directory.
Experiment
We use 99 datasets from the OpenML repository. We compare the performance of the PDC algorithm with 7 base learners. We use the macro F1 score as a metric. The search space is inspired from TPOT a state-of-the-art library in optimizing Sklearn pipelines
Description of the search space per estimator
Estimator | # parameters | # combinations |
---|---|---|
DecisionTree | 4 | 350 |
RandomForest | 7 | 1000 |
ExtraTree | 6 | 648 |
HistGradientBoosting | 6 | 486 |
Bagging | 6 | 96 |
ExtraTrees | 7 | 1000 |
GradientBoosting | 5 | 900 |
Search space per estimator
Estimator | Parameter | Values |
---|---|---|
DecisionTreeClassifier | criterion | gini, entropy |
max depth | None, 1, 2, 4, 6, 8, 11 | |
min samples split | 2, 4, 8, 16, 21 | |
min samples leaf | 1, 2, 4, 10, 21 | |
RandomForestClassifier | criterion | gini, entropy |
min samples split | 2, 4, 8, 16, 21 | |
max features | sqrt, 0.05, 0.17, 0.29, 0.41, 0.52, 0.64, 0.76, 0.88, 1.0 | |
min samples leaf | 1, 2, 4, 10, 21 | |
bootstrap | True, False | |
ExtraTreeClassifier | criterion | gini, entropy |
min samples split | 2, 5, 10 | |
min samples leaf | 1, 2, 4 | |
max features | sqrt, log2, None | |
max leaf nodes | None, 2, 12, 56 | |
min impurity decrease | 0.0, 0.1, 0.5 | |
HistGradientBoostingClassifier | max iter | 100, 10 |
learning rate | 0.1, 0.01, 1 | |
max leaf nodes | 31, 3, 256 | |
min samples leaf | 20, 4, 64 | |
l2 regularization | 0, 0.01, 0.1 | |
max bins | 255, 2, 64 | |
BaggingClassifier | n estimators | 10, 5, 100, 256 |
max samples | 1.0, 0.5 | |
max features | 0.5, 0.9, 1.0 | |
bootstrap | True, False | |
bootstrap features | False, True | |
ExtraTreesClassifier | criterion | gini, entropy |
max features | sqrt, 0.05, 0.17, 0.29, 0.41, 0.52, 0.64, 0.76, 0.88, 1.0 | |
min samples split | 2, 4, 8, 16, 21 | |
min samples leaf | 1, 2, 4, 10, 21 | |
bootstrap | False, True | |
GradientBoostingClassifier | learning rate | 0.1, 0.01, 1 |
min samples split | 2, 4, 8, 16, 21 | |
min samples leaf | 1, 2, 4, 10, 21 | |
subsample | 1.0, 0.05, 0.37, 0.68 | |
max features | None, 0.15, 0.68 |
OpenML benchmark datasets
data_id | NumberOfClasses | NumberOfInstances | NumberOfFeatures | NumberOfSymbolicFeatures | NumberOfFeatures_post_processing | MajorityClassSize | MinorityClassSize |
---|---|---|---|---|---|---|---|
43 | 2 | 306 | 4 | 2 | 3 | 225 | 81 |
48 | 3 | 151 | 6 | 3 | 5 | 52 | 49 |
59 | 2 | 351 | 35 | 1 | 34 | 225 | 126 |
61 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
164 | 2 | 106 | 58 | 58 | 57 | 53 | 53 |
333 | 2 | 556 | 7 | 7 | 6 | 278 | 278 |
377 | 6 | 600 | 61 | 1 | 60 | 100 | 100 |
444 | 2 | 132 | 4 | 4 | 3 | 71 | 61 |
464 | 2 | 250 | 3 | 1 | 2 | 125 | 125 |
475 | 4 | 400 | 6 | 5 | 5 | 100 | 100 |
714 | 2 | 125 | 5 | 3 | 4 | 76 | 49 |
717 | 2 | 508 | 11 | 1 | 10 | 286 | 222 |
721 | 2 | 200 | 11 | 1 | 10 | 103 | 97 |
733 | 2 | 209 | 7 | 1 | 6 | 153 | 56 |
736 | 2 | 111 | 4 | 1 | 3 | 58 | 53 |
744 | 2 | 250 | 6 | 1 | 5 | 141 | 109 |
750 | 2 | 500 | 8 | 1 | 7 | 254 | 246 |
756 | 2 | 159 | 16 | 1 | 15 | 105 | 54 |
766 | 2 | 500 | 51 | 1 | 50 | 262 | 238 |
767 | 2 | 475 | 4 | 3 | 3 | 414 | 61 |
768 | 2 | 100 | 26 | 1 | 25 | 55 | 45 |
773 | 2 | 250 | 26 | 1 | 25 | 126 | 124 |
779 | 2 | 500 | 26 | 1 | 25 | 267 | 233 |
782 | 2 | 120 | 3 | 1 | 2 | 63 | 57 |
784 | 2 | 140 | 4 | 2 | 3 | 70 | 70 |
788 | 2 | 186 | 61 | 1 | 60 | 109 | 77 |
792 | 2 | 500 | 6 | 1 | 5 | 298 | 202 |
793 | 2 | 250 | 11 | 1 | 10 | 135 | 115 |
811 | 2 | 264 | 3 | 2 | 2 | 163 | 101 |
812 | 2 | 100 | 26 | 1 | 25 | 53 | 47 |
814 | 2 | 468 | 3 | 1 | 2 | 256 | 212 |
824 | 2 | 500 | 11 | 1 | 10 | 274 | 226 |
850 | 2 | 100 | 51 | 1 | 50 | 51 | 49 |
853 | 2 | 506 | 14 | 2 | 13 | 297 | 209 |
860 | 2 | 380 | 3 | 1 | 2 | 195 | 185 |
863 | 2 | 250 | 11 | 1 | 10 | 133 | 117 |
870 | 2 | 500 | 6 | 1 | 5 | 267 | 233 |
873 | 2 | 250 | 51 | 1 | 50 | 142 | 108 |
877 | 2 | 250 | 51 | 1 | 50 | 137 | 113 |
879 | 2 | 500 | 26 | 1 | 25 | 304 | 196 |
880 | 2 | 284 | 11 | 1 | 10 | 142 | 142 |
889 | 2 | 100 | 26 | 1 | 25 | 50 | 50 |
895 | 2 | 222 | 3 | 1 | 2 | 134 | 88 |
896 | 2 | 500 | 26 | 1 | 25 | 280 | 220 |
902 | 2 | 147 | 7 | 5 | 6 | 78 | 69 |
906 | 2 | 400 | 8 | 1 | 7 | 207 | 193 |
909 | 2 | 400 | 8 | 1 | 7 | 203 | 197 |
911 | 2 | 250 | 6 | 1 | 5 | 140 | 110 |
915 | 2 | 315 | 14 | 4 | 13 | 182 | 133 |
918 | 2 | 250 | 51 | 1 | 50 | 135 | 115 |
925 | 2 | 323 | 5 | 1 | 4 | 175 | 148 |
932 | 2 | 100 | 51 | 1 | 50 | 56 | 44 |
933 | 2 | 250 | 26 | 1 | 25 | 136 | 114 |
935 | 2 | 250 | 11 | 1 | 10 | 140 | 110 |
936 | 2 | 500 | 11 | 1 | 10 | 272 | 228 |
937 | 2 | 500 | 51 | 1 | 50 | 282 | 218 |
969 | 2 | 150 | 5 | 1 | 4 | 100 | 50 |
973 | 2 | 178 | 14 | 1 | 13 | 107 | 71 |
974 | 2 | 132 | 5 | 1 | 4 | 81 | 51 |
1005 | 2 | 214 | 10 | 1 | 9 | 138 | 76 |
1011 | 2 | 336 | 8 | 1 | 7 | 193 | 143 |
1012 | 2 | 194 | 29 | 27 | 28 | 125 | 69 |
1054 | 2 | 161 | 40 | 1 | 39 | 109 | 52 |
1063 | 2 | 522 | 22 | 1 | 21 | 415 | 107 |
1065 | 2 | 458 | 40 | 1 | 39 | 415 | 43 |
1073 | 2 | 274 | 9 | 1 | 8 | 140 | 134 |
1100 | 3 | 478 | 11 | 5 | 10 | 247 | 90 |
1115 | 3 | 151 | 7 | 5 | 6 | 52 | 49 |
1413 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
1467 | 2 | 540 | 21 | 1 | 20 | 494 | 46 |
1480 | 2 | 583 | 11 | 2 | 10 | 416 | 167 |
1488 | 2 | 195 | 23 | 1 | 22 | 147 | 48 |
1490 | 2 | 182 | 13 | 1 | 12 | 130 | 52 |
1499 | 3 | 210 | 8 | 1 | 7 | 70 | 70 |
1510 | 2 | 569 | 31 | 1 | 30 | 357 | 212 |
1511 | 2 | 440 | 9 | 2 | 8 | 298 | 142 |
1523 | 3 | 310 | 7 | 1 | 6 | 150 | 60 |
1554 | 5 | 500 | 13 | 5 | 12 | 192 | 43 |
1556 | 2 | 120 | 7 | 6 | 6 | 61 | 59 |
1600 | 2 | 267 | 45 | 1 | 44 | 212 | 55 |
4329 | 2 | 470 | 17 | 14 | 16 | 400 | 70 |
40663 | 5 | 399 | 33 | 21 | 32 | 96 | 44 |
40681 | 2 | 128 | 7 | 7 | 6 | 64 | 64 |
41568 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
41977 | 2 | 156 | 91 | 1 | 90 | 98 | 58 |
41978 | 2 | 156 | 81 | 1 | 80 | 94 | 62 |
42011 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
42021 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
42026 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
42051 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
42066 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
42071 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
42186 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
42700 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
43859 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
44149 | 2 | 296 | 14 | 1 | 18 | 159 | 137 |
44151 | 3 | 149 | 5 | 0 | 4 | 50 | 49 |
44344 | 3 | 150 | 5 | 1 | 4 | 50 | 50 |
45711 | 2 | 530 | 14 | 3 | 13 | 354 | 176 |
Score comparison
2D datasets Examples
Here we see the difference in the learned patterns between PDL and the base learner. In case PDL is compatible with the base learner (DecisionTree, RandomForest) then the scores improves. In case the base learner is not compatible with PDL (SVC, AdaBoost, ...) then the scores gets lower.
Reference
Please cite us if you use this library in your research:
@article{belaid2024pairwise,
title={Pairwise Difference Learning for Classification},
author={Belaid, Mohamed Karim and Rabus, Maximilian and H{\"u}llermeier, Eyke},
journal={arXiv preprint arXiv:2406.20031},
year={2024}
}
The first commit correspond to the original implementation of the PDC algorithm
Acknowledgments: We would like to thank Tim Wibiral, Dorra ElMekki, Viktor Bengs, Muhammad Zeeshan Anwer, Muhammad Hossein Shaker, Alireza Javanmardi, Patrick Kolpaczki, and Maximilian Muschalik for their early comments on this work. We also acknowledge LRZ and IDIADA for computational resources.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file pdll-0.3.1.tar.gz
.
File metadata
- Download URL: pdll-0.3.1.tar.gz
- Upload date:
- Size: 27.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.11.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 5e4393946674c3113cd74f9396204faac7512b1b88830c08cf8e854268d13fb1 |
|
MD5 | 86a9e06e5c717f32a0afe91d649fe1d7 |
|
BLAKE2b-256 | 29449b93c5fd82150e7b6e1340b5207aaa2abfce0b86aabef1e4b379d92af42b |
File details
Details for the file pdll-0.3.1-py3-none-any.whl
.
File metadata
- Download URL: pdll-0.3.1-py3-none-any.whl
- Upload date:
- Size: 22.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.11.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e42f26f1c19a132e2c8a60f67546da245d458e3dedc7b3c5ab627511167fdafd |
|
MD5 | d37e232136f79d6cae4b99c3b9df45a1 |
|
BLAKE2b-256 | 62911fc76950a301476ee6fc9fef03545012c7c36187c8672ae3d148a15e73e5 |