Skip to main content

NIPT analysis pipeline

Project description

Build codecov

FluFFyPipe

NIPT analysis pipeline, using WisecondorX for detecting aneuplodies and large CNVs, Spiky for FFY and PREFACE for FF prediction (optional). wcx2cytosure is used to convert WisecondorX output into CytoSure Interpret Software (by OGT) format (.cgh file). FluFFYPipe produces a variety of output files, as well as a per batch csv summary.

fluffythesnail

Run FluFFyPipe

Run NIPT analysis, using a previously comnputed reference:

fluffy --sample <samplesheet>  --project <input_folder> --out <output_folder> --analyse

Run NIPT analysis, using an internally computed reference (i.e the reference is built using all samples listed in samplesheet):

fluffy --sample <samplesheet>  --project <input_folder> --out <output_folder> --analyse --batch-ref

optionally, skip preface:

fluffy --sample <samplesheet>  --project <input_folder> --out <output_folder> --skip_preface --analyse

All output will be written to the output folder, this output includes:

bam files
wisecondorX output
tiddit coverage summary
Fetal fraction estimation
cgh file with probes for coverage per bin and aberrations for visualization in CytoSure Interpret Software (by OGT).

as well as a summary csv and multiqc html (per batch)

the input folder is a project folder containing one folder per sample, each of these subfolders contain the fastq file(s). The samplesheet contains at least a "sampleID" column, the sampleID should match the subfolders in the input folder. The samplesheet may contain other columns, such as flowcell and index folder: such columns will be printed to the summary csv. If the samplesheet contains a SampleName column, fluffy will name the output according to SampleName

Create a WisecondorX reference

fluffy --sample <samplesheet>  --project <input_folder> --out <output_folder> --reference

samplesheet should contain atleast a "sampleID" column. All samples in the samplesheet will be used to construct the reference, visit the WisecondorX manual for more information.

Optional fluffy parameters:

Analysis mode:
	--dry_run - run the pipeline without generating files
	-l	-	add paramters to the slurm header of the script, should be given on the following format parameter:value
			example: qos:high 

Reference mode:
	--dry_run - run the pipeline without generating files
	
Rerun mode:
	--dry_run - run the pipeline without generating files

Troubleshooting and rerun

There are three statuses of the fluffy pipeline: running, complete, and failed

The status of a fluffy run is found in the

<output_folder>/analysis_status.json

The status of all jobs are listed in

<output_folder>/sacct/fluffy_<date>.log.status

Where is the timepoint when the jobs were submitted Use grep to find the failed jobs:

grep -v COMPLETE <output_folder>/sacct/fluffy_<date>.log.status

The output logs are stored in:

 <output_folder>/logs

Before continuing, you may want to generate the summary csv for all completed cases:

bash <output_folder>/scripts/summarizebatch-<hash>

where is a randomly generated string.

use the rerun module to rerun failed fluffy analyses:

fluffy --sample <samplesheet>  --project <input_folder> --out <output_folder> --skip_preface rerun

Install FluFFyPipe

FluFFyPipe requires python 3, slurm, slurmpy, and singularity, python-coloredlogs.

fluffy may be installed using pip:

pip install fluffy-cg

alternatively, fluffy is cloned and installed from github: git clone https://github.com/Clinical-Genomics/fluffy cd fluffy pip install -e .

Next download the FluFFyPipe, wcx2cytosure, blastp, bowtie2, and Spiky singularity container:

 singularity pull library://jeisfeldt/default/fluffy:sha256.dbef92cd5eab8558c2729f73a191d73a7576a24e9bb44dde7372c0cd405c4ef6 
 singularity pull --arch amd64 library://ravinale/wcx2cytosure/wcx2cytosure:latest
 singularity pull --arch amd64 library://jeisfeldt/spiky/spiky:latest
 singularity pull docker://quay.io/biocontainers/fastp:1.1.0--heae3180_0
 singularity pull docker://quay.io/biocontainers/bowtie2:2.5.5--ha27dd3b_0

copy the example config (found in example_config), and edit the variables. You will need to download/create the following files:

Reference fasta (indexed using bwa and bowtie2)

WisecondorX reference files (created using the reference mode)

PREFACE model file (optional)

blacklist bed file (used by wisecondorX)

FluFFyPipe singularity collection (singularity pull --name FluFFyPipe.sif shub://J35P312/FluFFyPipe)

wcx2cytosure singularity container

    Spiky regions and model

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

cg_fluffy-5.0.0.tar.gz (26.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

cg_fluffy-5.0.0-py3-none-any.whl (34.9 kB view details)

Uploaded Python 3

File details

Details for the file cg_fluffy-5.0.0.tar.gz.

File metadata

  • Download URL: cg_fluffy-5.0.0.tar.gz
  • Upload date:
  • Size: 26.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.25

File hashes

Hashes for cg_fluffy-5.0.0.tar.gz
Algorithm Hash digest
SHA256 b0f957d55689ed5339c66302d3bae29f60998baa958e1cf4feb5ee1b3c1557c4
MD5 f170d492497387ef5e15938d9c384408
BLAKE2b-256 50640d6769021070c94f2d1449d37ba3f5e60681f191962eda14d2d9af93700f

See more details on using hashes here.

File details

Details for the file cg_fluffy-5.0.0-py3-none-any.whl.

File metadata

  • Download URL: cg_fluffy-5.0.0-py3-none-any.whl
  • Upload date:
  • Size: 34.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.25

File hashes

Hashes for cg_fluffy-5.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 459b640d9ff4da32d9a38e3c2f30dec1984404fdc76e6f85f6a91d2d7728fefc
MD5 500b21ce27c404763d4db83f969f11d7
BLAKE2b-256 29da171c7aa794a00be130b726d4251fb5cde5c2c4fecb2e481ca6b047effde7

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page