Various tools and scripts used in the GHFC lab
Project description
ghfc-utils
set of small tools designed to help automatize simple task locally or on Pasteur's cluster.
ghfc-reannotate for the postprocessing of slivar files including filtering and geneset reannotation.
Installation
pip install git+https://{token_username}:{generated_token}@gitlab.pasteur.fr/ghfc/ghfc-utils.git
when on maestro:
module load Python/3.9.18
pip install --user git+https://maestro_tok_FC:rSUDdAJWZutsZTJseJ3V@gitlab.pasteur.fr/ghfc/ghfc-utils.git
slivar reannotator
A tool to filter and reannotate slivar files according to various parameters and genesets. The goal is to produce a more generic kind of slivar files and to use this for the user to run their own filtering.
usage: ghfc-reannotate [-h] configuration slivar output
positional arguments:
configuration config file
slivar slivar file to reannotate
output annotated slivar file
optional arguments:
-h, --help show this help message and exit
--chunksize CHUNKSIZE
size of the chunks read from the input (default 100000)
- This tools read the slivar file before decomposing the impacts by transcripts.
- It then filters all line using, in this order following the config file parameters:
- the geneset (based on the ENSG, mind the GRCh37/GRCh38 differences in ENSG)
- the impact / impact-categories
- if missense are kept, filtering them on their impact (using scores such as the mpc or the cadd)
- the gnomad frequency
- the variant/transcript are then sorted according to criteria given by the user in the config file from the most important to the least important
- for each sample, variant and gene (ENSG) the first transcript (most important given the config criteria) is kept.
yaml config file
This section is listing the accepted options in the config file, but example files are provided.
- (optional) geneset-file: path to the file containing the list of ENSG for out geneset of interest
- ordering-priority: the list (ordered) of criteria to use to rank the importance of the transcripts to output
- impact-categories-filter: the impact categories to keep during the filtering. The categories are defined in this package in impacts.yaml.
- impact-filter: the impacts to keep during the filtering process. the name of the impacts are visible in impacts.yaml.
- (optional) missense-filter: this section will define how the missense variant are further filtered
- first a subcategory is used for the score used, e.g. mpc, cadd
- for each subcategory, 3 fields are expected:
- field: the name of the slivar column containing the value
- min: the minimal value to keep (included)
- max: the maximal value to keep (excluded)
- in addition to the subcategories, a condition field is expected to specify how the subcategories are used. Possible values are:
- cadd_if_no_mpc: use the mpc and when not available (-1) uses the cadd
- cadd_and_mpc
- cadd_or_mpc
- mpc_only
- cadd_only
- (optional) gnomad-filter: to filter further on an included gnomAD column. 3 fields are expected here:
- field: the name of the slivar column containing the gnomad value to filter on
- min: the minimal value to keep (included)
- max: the maximal value to keep (included)
- (optional) pext-filter: to annotate each transcript and filter them using a pext file:
- file: path to the pext file to use
- field: nme of the outputed column
- min: the minimal value to keep (included). can put -1 to annotate and not filter on it.
Finally, some more global slivar parameters that are not likely to change a lot:
- slivar-field-name: the name of the slivar column that contains the list of all vep impacts per transcripts
- slivar-field-decomposed: the list of each field when they are decomposed. some of those fields are expected with the following names:
- impact
- ENSG
- canonical: the vep columns containing "YES" for the canonical transcripts
- loftee: the loftee "LoF" column
pext file
The pext is a bed file with the following columns (order important, there must be some header):
chr start end max_brain ensg symbol
Need to have the genome version to match the data (GRCh37/38 and using the chr or not in the chromosome names)
TODO
- offer to prefix geneset columns?
- offer to keep some of the original columns (impact/transcript)
- possibility to run on stdin / stdout?
- refining DP and AB
- needs for automated submission on the cluster? (means user has permission to use it)
- possibility to automate splitting in chunks and merging back?
slivar de novo ML
Moving the machine learning validator for de novo variants to this tool.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file ghfc_utils-0.0.1.tar.gz.
File metadata
- Download URL: ghfc_utils-0.0.1.tar.gz
- Upload date:
- Size: 8.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.9.6
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
8dd214a06852934b12597d8183622b2d8abc9ef71ec20ceaeea3cabb7a7f1e8c
|
|
| MD5 |
d0304fbc715865e7b15999629bcf8f23
|
|
| BLAKE2b-256 |
68fff2c9b4a3234b2caed41b3a8fab5efa173d843687931e58187c3dbad44c09
|
File details
Details for the file ghfc_utils-0.0.1-py3-none-any.whl.
File metadata
- Download URL: ghfc_utils-0.0.1-py3-none-any.whl
- Upload date:
- Size: 7.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.9.6
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
cf68b554d7b14688b22993285ab78959996d45260c16d387c259fbbc1615e78d
|
|
| MD5 |
919264b4ddddbc3c0c380add4a6507f1
|
|
| BLAKE2b-256 |
1e1852fc9aa62756f62b544b6ce3c7a915f42cb34e75b0526379053028468b6d
|