Skip to main content

No project description provided

Project description

Cross Validation Package

Python package for plug and play cross validation techniques. If you like the idea or you find usefull this repo in your job, please leave a star to support this personal project.

At the moment the package is not available using pip install <PACKAGE-NAME>.

For the installation from the source code click here.

Each method returns the confusion matrix and some performance metrics for each itheration and for the overall result. The performance metrics are:

  • Balanced Accuracy;
  • F1 Score;
  • Matthews Correlation Coefficient.

K-fold

K-fold consists of partitioning the dataset into k subsets; iteratively one of the k subsets is the test set and the others are the training set. The value of k could be chosen according to the amount of available data. Increasing the value of k the result is enlarging the training set and decreasing the size of the test set. Tipically, the default value of k is between 5 to 10, this is a good trade of between a robust validation and computational time. After a k-fold cross validation all the data set has been tested and it is possible to generate a confusion matrix and compute some performance metrics to validate the generalization capabilities of your model.

k-fold-cv-image K-fold cross-validation concept illustration Each row represents an iteration of the cross-validation; in blue, there are the subsets labeled as training set and in orange, the subset defined as test set for the i-th iteration. At the end, each subset has been tested getting the outcome, that could be compared to the real outputs of the instances

Example

from cross_validation.cross_validation import kfold

clf = RandomForestClassifier()
[cm, perf] = kfold(clf, X, y, verbose=True)

Leave One Out (LOO)

Leave-one-out (LOO) is a particular case of the k-fold when the value of k is equal to the number of data points in the dataset. This method should be used when the data set has few samples; this guarantees to have enough data point for the model training; after the training phase only one point will be evaluated by the model.

Example

from cross_validation.cross_validation import leave_one_out

clf = RandomForestClassifier()
[cm, perf] = leave_one_out(clf, X, y, verbose=True)

Leave One Subject Out (LOSO)

This method could be considered as a different version of the leave-one-out cross-validation. This method works leaving as a test set not a single example, but the entire examples that belong to a specific subject. The other subjects’ instances are used to train the learning algorithm. The main advantage of the LOSO is the removal of the subject bias because all the instances of the are the test set. This technique of cross-validation is widely used in the biomedical field where the the main task is to predict a disease or a condition of a patient using data of other patients.

Example

from cross_validation.cross_validation import leave_one_subject_out

clf = RandomForestClassifier()
[cm, perf] = leave_one_subject_out(clf, X, y, subject_ids, verbose=True):

Installation

For the installation from the source code type this command into your terminal window:

pip install git+<repository-link>

or

python -m pip install git+<repository-link>

or

python3 -m pip install git+<repository-link>

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

cross-validation-package-1.0.0.tar.gz (8.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

cross_validation_package-1.0.0-py3-none-any.whl (9.2 kB view details)

Uploaded Python 3

File details

Details for the file cross-validation-package-1.0.0.tar.gz.

File metadata

File hashes

Hashes for cross-validation-package-1.0.0.tar.gz
Algorithm Hash digest
SHA256 b81e684c9e5d5a0e39d44c8257bf0a814934d09953ce4a86bb0abd780a1830f0
MD5 ced6ad1796199598e65042ad21feea24
BLAKE2b-256 6e881032d5509016cbe953020beade33497301b5f2082adb983ea3af73317413

See more details on using hashes here.

File details

Details for the file cross_validation_package-1.0.0-py3-none-any.whl.

File metadata

File hashes

Hashes for cross_validation_package-1.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 0fabb0690a8f830ce4da0d0bce354fe705722376df4710b87da2f15dea4bf80d
MD5 91896cecf9ab861333427c540aadc530
BLAKE2b-256 0b0eaa7c54d1650c182a64c820d26370b97b72ebcd96d8e87cfeb992b7894277

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page