In silico Validation of Disease and Gene Sets, Clusterings or Subnetworks (DIGEST)
Project description
biodigest
The python package for DIGEST (validation of disease and gene sets, clusterings or subnetworks). It greatly facilitates in silico validation of gene and disease sets, clusterings or subnetworks via fully automated validation pipelines comprising disease and gene ID mapping, enrichment analysis, comparisons of shared genes and variants, and background distribution estimation. Moreover, functionality is provided to automatically update the external databases used by the pipelines.
A light version excluding the subnetwork option end therefore reducing the needed requirements for installing more complex python packages can be found as biodigest-light.
Setup for proper usage
After installing biodigest, you need to install graph-tools package.
import biodigest
Before you can run the validation, you need to download precalculated mappings and distance matrices. These can be downloaded in two ways:
1. [Recommended] Get data from api
The API keeps all data up to date and checks for updates at regular intervals. This process takes 1-5 minutes depending on the internet connection.
from biodigest import setup
setup.main(setup_type="api")
2. Create data from scratch
Any mappings are freshly fetched from the databases and the distance matrices are calculated. Be aware that this can take up to 3 hours.
from biodigest import setup
setup.main(setup_type="create")
Run validation
from biodigest.single_validation import single_validation
results = single_validation(tar: Union[pd.DataFrame, set], tar_id: str, mode: str, distance: str = "jaccard",
ref: Union[str, set] = None, ref_id: str = None, enriched: bool = False,
network_data: dict = None, mapper: Mapper = FileMapper(), runs: int = config.NUMBER_OF_RANDOM_RUNS,
background_model: str = "complete", replace=100, verbose: bool = False)
All results that can later be saved and visualize are saved in results
as data type dict()
.
Parameters
Required parameters
- tar: Target input you want to be validated
- a cluster should be of type
pd.DataFrame()
withcolumns=["id","cluster"]
- a set should be of type
set()
- a cluster should be of type
- tar_id: Is the id type of the target (see possible options below)
- ref: Reference, to which tar will be compared (Only for mode id-set and set-set)
- a single id should be of type
str
- a set should be of type
set()
- a single id should be of type
- ref_id: Is the id type of the reference (see possible options below)
- mode: Desired mode. See possible options below.
Optional parameters
- distance: Distance measure used for pairwise comparison
- enriched: Set
True
, if only enriched attributes of the reference set should be used (Only for set-set) - network_data: Only for "subnetwork" and "subnetwork-set" mode. Dictionary consisting of {"network_file": path to network file, "prop_name": name of vertex property with ids if network file of type graphml or gt, "id_type": id type of network ids}
- background_model: Model defining how random values should be picked. See possible options below.
- mapper: Mapper object indicating where all files from the setup are saved.
[Default=FileMapper()]
- runs: Number of runs with random target values for p-value calculation
- replace: Percentage of how many of the original ids should be replaced with random ids
- verbose: get additional information during the run
Supported types
- gene types: entrez, ensembl, symbol, uniprot
- disease types: mondo, omim, snomedct, umls, orpha, mesh, doid, ICD-10
Modes
- set: Compare similarity inside the set using the mean of all pairwise comparisons
- id-set: Compare target set to reference set
- set-set: Compare target set to reference id
- clustering: Compare cluster quality inside clustering. Either genes or diseases
- subnetwork: Compare similarity inside the subnetwork nodes. Either genes or diseases
- subnetwork-set: Compare target subnetwork to reference set. Both either genes or diseases
Background models
- complete: Random ids will be picked completely randomly
- term-pres: Random ids will preserve the number of mapped terms for the replaced ids
- network: Random ids will preserve the number of connected components in given network.
Result
The method call returns the result in a json format of datatype dict which consists of the following elements:
result = {'status': 'Status text',
'input_values': {'values': dict(), 'mapped_ids': list()},
'random_values': {'values': dict()},
'p_values': {'values': dict()}}
- status: contains either an error message if a mapping failed or "ok" if IDs could be mapped
- input_values:
- values: table in dict format with the functional or genetic relevance score(s) determined for solely their input
- mapped_ids: list containing the IDs with non empty annotations per functional or genetic annotation type
- random_values:
- values: table in dict format with the functional or genetic relevance score(s) determined for all random runs
- p_values: table in dict format with the calculated empirical P-values using the selected background model and other parameters that indicate the significance of the calculated relevance scores derived from the input
Save and visualize results
from biodigest.single_validation import save_results
from biodigest.evaluation.d_utils.plotting_utils import create_plots, create_extended_plots
# Save results into json file and 2 .csv table files
save_results(results: dict, prefix: str, out_dir)
# Generate and save plots based on results
# Consisting of p-value plot and mappability plot
create_plots(results, mode, tar, tar_id, out_dir, prefix, file_type: str = "pdf")
# Generate and save extended plots based on results
# Consisting of distribution and sankey plots
create_extended_plots(results, mode, tar, out_dir, prefix, file_type: str = "pdf", mapper:Mapper=FileMapper())
Parameters
Required parameters
- results: Is the output created with method
single_validation
as data typedict()
- prefix: Prefix for file names
- out_dir: Output directory for the generated files
Additional required parameters for create_plots
- tar: Target input you want to be validated
- a cluster should be of type
pd.DataFrame()
withcolumns=["id","cluster"]
- a set should be of type
set()
- a cluster should be of type
- tar_id: Is the id type of the target (see possible options above)
- mode: Desired mode. See possible options above.
Optional parameters for create_plots
- file_type: Type of the plots image files.
Run significance contribution calculation
If you are interested in how the single ids from the target set are contribution to the final calculated empirical P-values, you can run the significance contribution calculations. Keep in mind the runtime will increase in a linear way based on the number of ids in the input target set.
from biodigest.single_validation import significance_contributions
results_sig = significance_contributions(results: dict,
tar: Union[pd.DataFrame, set], tar_id: str, mode: str, distance: str = "jaccard",
ref: Union[str, set] = None, ref_id: str = None, enriched: bool = False,
mapper: Mapper = FileMapper(), runs: int = config.NUMBER_OF_RANDOM_RUNS,
background_model: str = "complete", replace=100, verbose: bool = False)
Parameters
- results: This is the output generated from
single_validation
on the full input set. This will be used to calculate the significance contribution of the single ids from the input set. - all other paramters: look up parameters under "Run validation"
Save and visualize results
from biodigest.single_validation import save_contribution_results
from biodigest.evaluation.d_utils.plotting_utils import create_contribution_plots, create_contribution_graphs
# Save results into json file and .csv table files for each validation type
save_contribution_results(results: dict, prefix: str, out_dir)
# Generate and save plots based on results
# Consisting an overview heatmap of the top 15 ids
# with the largest absolute overall significance contribution
# and top 10 largest positive and negative
# significance contribution per annotation type
create_contribution_plots(result_sig, out_dir, prefix, file_type: str = "pdf")
# Generate a graph visualization from subnetwork constructed
# by the input set and the given (or default) network
# with nodes colored by their significance contribution
create_contribution_graphs(result_sig, tar_id, network_data, out_dir, prefix,
file_type: str = "pdf", mapper: Mapper = FileMapper())
Parameters
Required parameters
- results: Is the output created with method
significance_contributions
as data typedict()
- prefix: Prefix for file names
- out_dir: Output directory for the generated files
Additional required parameters
- input_type: "genes" or "diseases" based on id type of target set
- network_data: Only for "subnetwork" and "subnetwork-set" mode. Dictionary consisting of {"network_file": path to network file, "prop_name": name of vertex property with ids if network file of type graphml or gt, "id_type": id type of network ids}
- mapper: Mapper object indicating where all files from the setup are saved.
[Default=FileMapper()]
Optional parameters for create_plots
- file_type: Type of the plots image files.
Example runs
Check out the tutorial to see examples of usage in a script.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.