Skip to main content

A deep-learning based multi-omics bulk sequencing data integration suite with a focus on (pre-)clinical endpoint prediction.

Project description

logo

Downloads benchmarks tutorials

flexynesis

A deep-learning based multi-omics bulk sequencing data integration suite with a focus on (pre-)clinical endpoint prediction. The package includes multiple types of deep learning architectures such as simple fully connected networks, supervised variational autoencoders, graph convolutional networks, multi-triplet networks different options of data layer fusion, and automates feature selection and hyperparameter optimisation. The tools are continuosly benchmarked on publicly available datasets mostly related to the study of cancer. Some of the applications of the methods we develop are drug response modeling in cancer patients or preclinical models (such as cell lines and patient-derived xenografts), cancer subtype prediction, or any other clinically relevant outcome prediction that can be formulated as a regression, classification, survival, or cross-modality prediction problem.

workflow

Citing our work

In order to refer to our work, please cite our manuscript currently available at BioRxiv.

Getting started with Flexynesis

Command-line tutorial

Jupyter notebooks for interactive usage

Benchmarks

For the latest benchmark results see: https://bimsbstatic.mdc-berlin.de/akalin/buyar/flexynesis-benchmark-datasets/dashboard.html

The code for the benchmarking pipeline is at: https://github.com/BIMSBbioinfo/flexynesis-benchmarks

Defining Kernel for the Jupyter Notebook

For interactively using flexynesis on Jupyter notebooks, one can define the kernel to make flexynesis and its dependencies available on the jupyter session.

Assuming you have already defined an environment and installed the package:

conda activate flexynesisenv 
python -m ipykernel install --user --name "flexynesisenv" --display-name "flexynesisenv"

Compiling Notebooks

papermill can be used to compile the tutorials under examples/tutorials.

If the purpose is to quickly check if the notebook can be run; set HPO_ITER to 1. This sets hyperparameter optimisation steps to 1. For longer training runs to see more meaningful results from the notebook, increase this number to e.g. 50.

Example:

papermill examples/tutorials/brca_subtypes.ipynb brca_subtypes.ipynb -p HPO_ITER 1 

The output from papermill can be converted to an html file as follows:

jupyter nbconvert --to html brca_subtypes.ipynb 

Documentation

Documentation generated using mkdocs

pip install mkdocstrings[python]
mkdocs build --clean

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

flexynesis-0.2.10.tar.gz (76.7 kB view details)

Uploaded Source

Built Distribution

flexynesis-0.2.10-py3-none-any.whl (95.0 kB view details)

Uploaded Python 3

File details

Details for the file flexynesis-0.2.10.tar.gz.

File metadata

  • Download URL: flexynesis-0.2.10.tar.gz
  • Upload date:
  • Size: 76.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.11.9

File hashes

Hashes for flexynesis-0.2.10.tar.gz
Algorithm Hash digest
SHA256 e6faea383272ee0bdd2792a2d11e4c1682d898eea09008990aa1132a93e72b34
MD5 d440cd7fe118671599cd052236a60499
BLAKE2b-256 ec36bfec16887b7586521038d1e26a7d3ddfa82c161ef8d80d01922d60526579

See more details on using hashes here.

File details

Details for the file flexynesis-0.2.10-py3-none-any.whl.

File metadata

  • Download URL: flexynesis-0.2.10-py3-none-any.whl
  • Upload date:
  • Size: 95.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.0 CPython/3.11.9

File hashes

Hashes for flexynesis-0.2.10-py3-none-any.whl
Algorithm Hash digest
SHA256 7be36485e7bf0b2554c6a6b00be2738198b221b3100619315ec567f461f6ac70
MD5 57f0ed3e4ee5e854a4e791617c65fd7d
BLAKE2b-256 7a826ae287e45e028dbf2ac2fc20190e6874d25c382415853492e2a52912d5ca

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page