Provides utilities for the training and evaluation of machine learning algorithms
Project description
"MLRL-Testbed": Utilities for Evaluating Multi-label Rule Learning Algorithms
Important links: Documentation | Issue Tracker | Changelog | Contributors | Code of Conduct | License
This software package provides utilities for training and evaluating machine learning algorithms, including classification and regression problems.
Functionalities
Most notably, the package includes a command line API that allows configuring and running machine learning algorithms. For example, the BOOMER algorithm integrates with the command line API out-of-the-box. For using other algorithms only a few lines of Python code are necessary.
The command line API allows applying machine learning algorithms to different datasets and evaluating their predictive performance in terms of commonly used measures (provided by the scikit-learn framework). In detail, it supports the following functionalities:
- Single- and multi-output datasets in the Mulan and Meka format are supported.
- Datasets can automatically be split into training and test data, including the possibility to use cross validation. Alternatively, predefined splits can be used by supplying the data as separate files.
- One-hot-encoding can be applied to nominal or binary features.
- Binary predictions, scores, or probability estimates can be obtained from a machine learning algorithm, including classification and regression algorithms. Evaluation measures that are suited for the respective type of predictions are picked automatically.
- Evaluation scores can be saved to output files and printed on the console.
- Ensemble models can be evaluated incrementally, i.e., they can be evaluated repeatedly using only a subset of their members with increasing size.
- Textual representations of models can be saved to output files and printed on the console. In addition, the characteristics of models can also be saved and printed.
- Characteristics of datasets can be saved to output files and printed on the console.
- Unique label vectors contained in a classification dataset can be saved to output files and printed on the console.
- Predictions can be saved to output files and printed on the console. In addition, characteristics of predictions can also be saved and printed.
- Models for the calibration of probabilities can be saved to output files and printed on the console.
- Models can be saved on disk in order to be reused by future experiments.
- Algorithmic parameters can be read from configuration files instead of providing them via command line arguments. When providing parameters via the command line, corresponding configuration files can automatically be saved on disk.
License
This project is open source software licensed under the terms of the MIT license. We welcome contributions to the project to enhance its functionality and make it more accessible to a broader audience. A frequently updated list of contributors is available here.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distributions
Built Distribution
File details
Details for the file mlrl_testbed-0.11.1-py3-none-any.whl
.
File metadata
- Download URL: mlrl_testbed-0.11.1-py3-none-any.whl
- Upload date:
- Size: 60.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.12.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | c0c1f5f8fdde4431fa0c28b71fee7e97835b926929d3fd7ecb23e7e79849eb8a |
|
MD5 | b2390211f44ceca48189a9a96e558a08 |
|
BLAKE2b-256 | c03736059129d94cee9eca47cb590ad18642e2ff159657a9290df0829fc842b3 |