Skip to main content

Targeted ortholog search for miRNAs

Project description

ncOrtho

PyPI version License: GPL v3

NcOrtho is a tool for the targeted search of orthologous micro RNAs (miRNAs) throughout the tree of life. Conceptually, it works similar to the program fDOG in that a probabilistic model of a reference miRNA is created. For training the model, orthologs of the reference sequence are first identified in a set of taxa that are more closely related to the reference species. In contrast to fDOG, ncOrtho does not train hidden Markov Models but covariance models (CMs) (Eddy & Durbin, 1994) which also model conservation of the miRNA's secondary structure.

workflow

Getting Started

NcOrtho depends on multiple third party applications, some of which are Linux specific. All dependencies can be installed with Anaconda. It is recommended to create a new Anaconda environment for this. For example:

conda create --name ncOrtho python=3.8
conda activate ncOrtho

Prerequisites

  • Operating System: Linux (tested on: Ubuntu 20.04)
  • Python: version 3 or higher (tested with v3.8)
Tool Tested version Anaconda installation
BLASTn v2.7.1 conda install -c kantorlab blastn
Infernal v1.1.4 conda install -c bioconda infernal
t_coffee v13.45 conda install -c bioconda t-coffee
MUSCLE v5.1 conda install -c bioconda muscle

Installing

After installing all three dependencies, ncOrtho can be installed with pip:

 pip install ncOrtho

Usage

CM construction

As a targeted search for orthologs, ncOrtho's biggest strength is its flexibility to change the taxonomic scope of an analysis according to the research question at hand.

For this reason, a few questions need to be answered, before we can start constructing covariance models:

  1. What is the reference species?
  2. How phylogenetically diverse will my set of target species be?
  3. From which species should the core set of miRNA orthologs be extracted, which will be used for training the CMs?
  4. Which miRNAs are going to be used for the ortholog search?

To identify suitable core species for a given reference species you can calculate an estimate of conserved synteny given a set of pairwise ortholog predictions with:

ncCheck -p <parameters.yaml> -o <outdir>

You can find additional information about ncCheck in the WIKI.

As soon as you know what your core species are going to be, you will need to collect the following data:

  • Genomic sequence in FASTA format (e.g "genomic.fna" from RefSeq)
  • Genome annotation in GFF3 format (e.g. "genomic.gff" from RefSeq)
  • Pairwise orthologs of all proteins between the reference and each core species (more information here

Modify the example parameters file to contain all relevant paths to your input files. The "name" property of your reference and core species has to merely be a unique identifier. It is however recommended to use whitespace-free species names to increase readability.

Additional to the parameters file, you will need a tab separated file containing the position and sequence of each miRNA for which a model should be constructed (more information here).

You can then start CM construction with:

ncCreate -p <parameters.yaml> -n <mirnas.tsv> -o <outdir>

If you encounter errors, make sure that:

  • The identifiers in the pairwise orthologs files match the ones in the gff files (use the -idtype= flag to use other ID types)
  • The contig/chromosome column in tab separated miRNA input file match the contig/chromosome id in the reference gff file

Use ncCreate -h to see all available options for CM construction.

Orthology search

You can start the orthology search with:

ncSearch -m <CMs/> -n <mirnas.tsv> -q <query_genome.fa> -r <reference_genome.fa> -o <outdir>

Use ncSearch -h to see all available options for the orthology search or have a look at the WIKI.

Phylogenetic Analysis

To facilitate the downstream analyses of miRNA orthologs, we also supply the ncAnalyze function:

ncAnalzye -r <result directory of ncOrtho> -o <output_dir> -m <mappingfile>

This will create a phylogenetic Profile ready for visualisation in PhyloProfile as well as calculate a supermatrix species tree based on the miRNA orthologs.

More information can be found with ncAnalyze -h or the WIKI

Support

Please refer to our Wiki Page of known issues first, then consider opening an issue on GitHub or contacting me directly via mail

Contributors

Dept. for Applied Bioinformatics Institute for Cell Biology and Neurosciences, Goethe University, Frankfurt am Main

License

This project is licensed under the GNU General Public License v3.0 - see the LICENSE.md file for details

Acknowledgments

Contact

For support or bug reports please contact: langschied@bio.uni-frankfurt.de

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ncortho-1.0.0.tar.gz (58.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

ncortho-1.0.0-py3-none-any.whl (68.3 kB view details)

Uploaded Python 3

File details

Details for the file ncortho-1.0.0.tar.gz.

File metadata

  • Download URL: ncortho-1.0.0.tar.gz
  • Upload date:
  • Size: 58.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.10.20

File hashes

Hashes for ncortho-1.0.0.tar.gz
Algorithm Hash digest
SHA256 05c53d174efcf85a44a17b454cd92ee00d683b2170edaded25a66785c863af0c
MD5 296024a7a188ea12c89536ae35a39fa5
BLAKE2b-256 70d6170500cf10e1510cd9839c14c73fb176b3cc2e7ad2b76cbc654013e2dad0

See more details on using hashes here.

File details

Details for the file ncortho-1.0.0-py3-none-any.whl.

File metadata

  • Download URL: ncortho-1.0.0-py3-none-any.whl
  • Upload date:
  • Size: 68.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.10.20

File hashes

Hashes for ncortho-1.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 30171c9a9aded62859e03deec8842adae1878764e9a3caa1a507a6c5d097021c
MD5 da1a04fef75d40b19c34c4de530a6d7b
BLAKE2b-256 d0450d622b0a20964053ebbe4637f7ea305f2523c7061a497b18811b159e732e

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page