No project description provided
Project description
Data generator for synthetic data including artificial classes, intraclass correlations, pseudo-classes and random data - Sphinx Doc
Table of Contents
Purpose
In order to develop new methods or to compare existing methods for feature selection, reference data with known dependencies and importance of the individual features are needed. This data generator can be used to simulate biological data for example artificial high throughput data including artificial biomarkers. Since commonly not all true biomarkers and internal dependencies of high-dimensional biological datasets are known with certainty, artificial data enables to know the expected outcome in advance. In synthetic data, the feature importances and the distribution of each class are known. Irrelevant features can be purely random or belong to a pseudo-class. Such data can be used, for example, to make random effects observable.
Data structure
Different parts of the data set
The synthetic-data-generator produces data sets consisting of up to three main parts:
- Relevant features belonging to an artificial class (for example artificial biomarkers)
- [optional] Pseudo-classes (for example a patient's height or gender, which have no association with a particular disease)
- [optional] Random data representing the features (for example biomarker candidates) that are not associated with any class
The number of artificial classes is not limited. Each class is generated individually and then combined with the others. In order to simulate artificial biomarkers in total, all individual classes have the same number of features in total.
This is an example of simulated binary biological data including artificial biomarkers:
Data distribution and effect sizes
For each class, either the normal distribution or the log normal distribution can be selected. The different classes can be shifted to regulate the effect sizes and to influence the difficulty of data analysis.
The normally distributed data could, for example, represent the range of values of healthy individuals. In the case of a disease, biological systems are in some way out of balance. Extreme changes in values as well as outliers can then be observed (Concordet et al., 2009). Therefore, the values of a diseased individual could be simulated with a lognormal distribution.
Example of log-normal and normal distributed classes:
Correlations
Intra-class correlation can be generated for each artificial class. Any number of groups containing correlated features can be combined with any given number of uncorrelated features.
However, a high correlation within a group does not necessarily lead to a high correlation to other groups or features of the same class. An example of a class with three highly correlated groups but without high correlations between all groups:
It is probably likely that biomarkers of healthy individuals usually have a relatively low correlation. On average, their values are within a usual "normal" range. In this case, one biomarker tends to be in the upper normal range and another biomarker in the lower normal range. However, individually it can also be exactly the opposite, so that the correlation between healthy individuals would be rather low. Therefore, the values of healthy people could be simulated without any special artificially generated correlations.
In the case of a disease, however, a biological system is brought out of balance in a certain way and must react to it. For example, this reaction can then happen in a coordinated manner involving several biomarkers, or corresponding cascades (e.g. pathways) can be activated or blocked. This can result in a rather stronger correlation of biomarkers in patients suffering from a disease. To simulate these intra-class correlations, a class is divided into a given number of groups with high internal correlation (the respective strength can be defined).
Pseudo-classes
One option for an element of the generated data set is a pseudo-class. For example, this could be a patient's height or gender, which are not related to a specific disease.
The generated pseudo-class contains the same number of classes with identical distributions as the artificial biomarkers. But after the generation of the individual classes, all samples (rows) are randomly shuffled. Finally, combining the shuffled data with the original, unshuffled class labels, the pseudo-class no longer has a valid association with any class label. Consequently, no element of the pseudo-class should be recognized as relevant by a feature selection algorithm.
Random Features
The artificial biomarkers and, if applicable, the optional pseudo-classes can be combined with any number of random features. Varying the number of random features can be used, for example, to analyze random effects that occur in small sample sizes with a very large number of features.
Installation
The artificial-data-generator is available at the Python Package Index (PyPI). It can be installed with pip:
$ pip install artificial-data-generator
Project Setup
We recommend to do the setup in a text console and not with a GUI tool. This offers better control and transparency.
We use Poetry and pyenv. Not Conda, Anaconda or pip directly.
1. Get Project Source
First you have to clone the project with GIT.
If you want to make a pull request, you must clone your previously forked project and
not the original project.
After the project has been cloned, use cd
to change into the project directory.
2. Install Poetry
We use Poetry for dependency management and packaging in this project. The next step is the installation of Poetry, if you do not already have it. Poetry offers different installation options. We recommend the option "with the official installer". But it does not matter. It's your choice.
3. Configure Poetry
We suggest the following two config options. These are not mandatory but useful.
Set virtualenvs.prefer-active-python
to true
.
With this setting Poetry uses the currently activated Python version to create a new virtual environment.
If set to false, the Python version used during Poetry installation is used.
This makes it possible to determine the exact Python version for development.
This can be done global or locale.
We suggest to do this setting as global.
- global setting:
poetry config virtualenvs.prefer-active-python true
- locale setting:
poetry config virtualenvs.prefer-active-python true --local
- this will create or change thepoetry.toml
file
Set virtualenvs.options.always-copy
to true
.
When the new virtual environment is created (later) all needed files are copied into it instead of symlinked.
The advantage is that you can delete the old globally installed Python version later without breaking the Python in
the locale virtual environment.
The disadvantage is that we waste some disk space.
This can be done global or locale.
We suggest to do this setting as global.
- global setting:
poetry config virtualenvs.options.always-copy true
- locale setting:
poetry config virtualenvs.options.always-copy true --local
- this will create or change thepoetry.toml
file
4. Set the Python Version (pyenv)
We recommend pyenv to install and manage different Python versions. First install pyenv if you do not already have it.
Next install the appropriate Python version.
We recommend the development on the oldest still permitted Python version of the project.
This version number can be found in the pyproject.toml
file in the setting called
tool.poetry.dependencies.python
. If this is set like python = "^3.8"
we use pyenv to install Python 3.8:
pyenv install 3.8
This installs the latest 3.8 Python version.
If the Python installation was successful we use pyenv versions
to see which exact Version is installed.
Then we activate this version with pyenv local <version>
.
This command will create a .python-version
file in the project directory.
Make sure that you are still in the project directory.
For example execute: pyenv local 3.8.17
5. Install the Project with Poetry
Execute poetry install --all-extras
to install the project.
This installs all dependencies, optional (extra) dependencies and
needed linting, testing and documentation dependencies.
With this method, the sources are also implicitly installed in
editable mode.
Licensing
Copyright (c) 2022 Sigrun May, Helmholtz-Zentrum für Infektionsforschung GmbH (HZI)
Copyright (c) 2022 Sigrun May, Ostfalia Hochschule für angewandte Wissenschaften
Licensed under the MIT License (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License by reviewing the file LICENSE in the repository.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file artificial_data_generator-0.0.3.tar.gz
.
File metadata
- Download URL: artificial_data_generator-0.0.3.tar.gz
- Upload date:
- Size: 19.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.2 CPython/3.8.18 Linux/6.5.0-1016-azure
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | dcd1acb8f6e19fb6ddfb0fe8dbfc94fe2f5e785a655e127e541fee5be834ad11 |
|
MD5 | 480e5f1f85ca81f2c2c0ac6b08bb43cb |
|
BLAKE2b-256 | ffae1c03171f9ec785c9230343a35deb1433baf6776ff658335fcf8056ab6c81 |
File details
Details for the file artificial_data_generator-0.0.3-py3-none-any.whl
.
File metadata
- Download URL: artificial_data_generator-0.0.3-py3-none-any.whl
- Upload date:
- Size: 19.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.8.2 CPython/3.8.18 Linux/6.5.0-1016-azure
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ccfc70d3927a753af64b0b7302f5ad48c708b077d5494e2df7f9fc6fcead5be6 |
|
MD5 | ec3bc240ede46cd31da3dac9d912ead1 |
|
BLAKE2b-256 | 7bc2d84e17d1075a43802b8069f974149fa84e7ab840c4d0635dab79b0354319 |