Skip to main content

Quantile normalization

Project description

qnorm

PyPI version Anaconda version tests DOI monthd totald

Quantile normalization made easy! This tool was developed as the current (Python) implementations scattered across the web do not correctly resolve collisions/ties in the ranks. Properly resolving rank ties is important when ties happen frequently, such as when working with discrete numbers (integers) in count tables. This implementation should be relatively fast, and can use multiple cores to sort the columns and tie-resolvement is accelerated by numba.

Code example

We recreate the example of Wikipedia:

import pandas as pd
import qnorm

df = pd.DataFrame({'C1': {'A': 5, 'B': 2, 'C': 3, 'D': 4},
                   'C2': {'A': 4, 'B': 1, 'C': 4, 'D': 2},
                   'C3': {'A': 3, 'B': 4, 'C': 6, 'D': 8}})

print(qnorm.quantile_normalize(df, axis=1))

which is what we expect:

         C1        C2        C3
A  5.666667  5.166667  2.000000
B  2.000000  2.000000  3.000000
C  3.000000  5.166667  4.666667
D  4.666667  3.000000  5.666667

Qnorm accepts an (optional) axis argument, which is used to normalize along. If axis=1 (default), standardize each sample (column), if axis=0, standardize each feature (row).

  • note: pandas is an optional dependency of qnorm, and if you want to quantile normalize dataframes make sure to install pandas yourself (conda/pip install pandas).

  • note: you can also pass numpy arrays as input to qnorm.quantile_normalize.

Multicore support

To accelerate the computation you can pass a ncpus argument to the function call and qnorm will be run in parallel:

qnorm.quantile_normalize(df, ncpus=8)  

Normalize onto distribution

You can also use the quantile_normalize function to normalize "onto" a distribution, by passing a target along to the function call.

import pandas as pd
import qnorm

df = pd.DataFrame({'C1': {'A': 4, 'B': 3, 'C': 2, 'D': 1},
                   'C2': {'A': 1, 'B': 2, 'C': 3, 'D': 4}})

print(qnorm.quantile_normalize(df, target=[8, 9, 10, 11]))

With our values now transformed onto the target:

     C1    C2
A  11.0   8.0
B  10.0   9.0
C   9.0  10.0
D   8.0  11.0

How fast is it and what is its memory usage?

How better to measure this than a little benchmark / example? For this example we will consider 100 replicates, each consisting of one million integer values between 0 and 100, which should give us plenty of rank ties.

import numpy as np
import qnorm


test = np.random.randint(0, 100, size=(1_000_000, 100), dtype=np.int32)

qnorm.quantile_normalize(test, ncpus=4)
user@comp:~$ /usr/bin/time --verbose python small_qnorm_script.py |& grep -P "(wall clock|Maximum resident set size)"

Elapsed (wall clock) time (h:mm:ss or m:ss): 0:07.52
Maximum resident set size (kbytes): 2768884

It takes only 7.5 seconds to initialize our table and quantile normalize it. I think that's pretty fast!

The test array we made consists of 100 * 1.000.000 = 100.000.000 single point precision integers, so four bytes each (400.000.000 bytes, 0.4 gigabytes). The memory footprint of our script is 0.27 gigabytes, around 7 times our input. Unfortunately that makes qnorm a bit memory hungry, but that should not be a problem in 99% of the cases. If memory usage is a problem take a look at the low-memory implementation.

Scaling of ncpus

Using more than four cpus generally does not lead to a much bigger speedup.

mini benchmark

Incremental quantile norm

In case you want to quantile normalize excessively large tables, there is a "memory-efficient" implementation. This implementation gets its memory efficiency by calculating the mean "online", which means we calculate it on fractions of the total table and then update the value. The other change is that intermediate results are written to disk. This means that this implementation effectively swaps memory to disk, and thus is not "memory hungy", but "disk hungry". However this incremental method can scale to virtually infinitely large tables (or until you run out of disk space).

Let's say we want to do something crazy like quantile normalize the human genome in 10 basepair bins. That means we will have around 300.000.000 values per sample. File-based qnorm works with both csv/tsv, parquet, and hdf files. For this example we will work with hdf files (make sure to set data_columns=True). Parquet and hdf files also are fast, but csv/tsv files are (very) slow because of the enormous amount of I/O they require.

df = pd.DataFrame(index=range(300_000_000), dtype=int, columns=["sample"+str(col) for col in range(64)])
df[:] = np.random.randint(0, 100, size=df.shape)
df.to_hdf("hg_bins.hdf", key="qnorm", format='table', data_columns=True)

We can now compare the speed and memory of the file-based method vs the "standard" method.

import qnorm

# incremental qnorm (reads hg_bins.hdf from disk and writes output to hg_bins_qnorm.hdf) 
qnorm.incremental_quantile_normalize("hg_bins.hdf", "hg_bins_qnorm.hdf", rowchunksize=500_000, colchunksize=4, ncpus=4)

# standard
df = pd.read_hdf(f"hg_bins.hdf").astype("float32")
df_qnorm = qnorm.quantile_normalize(df, ncpus=4)

mini benchmark file

Our standard method does not come farther that 2^4=16 samples before running out of memory on a 512 gigabyte system! The incremental method has similar timings and even seems to scale better than the standard method for large arrays. And it takes only an hour to normalize 64 samples.

The rowchunksize and colchunksize respectively influence in how large of chunks the output is written to disk and how many columns are being sorted and quantile normalized at the same time. Generally speaking, the larger the better, however the defaults should most of the times be sufficiently fast.

  • note: Both in-memory normalization as incremental normalization should produce identical results, and neither is more correct than the other.

  • note: The incremental implementation requires pandas to be installed (conda/pip install pandas).

  • note: When using hdf files make sure to install (py)tables (conda install pytables or pip install tables).

  • note: When using parquet files make sure to install pyarrow (conda install pyarrow or pip install pyarrow).

  • note: The input format specifies the output format.

  • note: Because of the design of hdf5, there is a limit on the number of columns it can hold. In case you have a lot of columns (500+), it is probably best to use parquet and not hdf5.

Command Line Interface (CLI) example

Qnorm also contains a CLI for converting csv/tsv files. The CLI depends on pandas, but this is an optional dependency of qnorm. To make use of the CLI make sure to install pandas in your current environment as well!

user@comp:~$ qnorm --help

usage: qnorm [-h] [-v] table

Quantile normalize your table

positional arguments:
  table          input csv/tsv file which will be quantile normalized

optional arguments:
  -h, --help     show this help message and exit
  -v, --version  show program's version number and exit

And again the example of Wikipedia:

user@comp:~$ cat table.tsv
        C1      C2      C3
A       5       4       3
B       2       1       4
C       3       4       6
D       4       2       8

user@comp:~$ qnorm table.tsv
        C1      C2      C3
A       5.666666666666666       5.166666666666666       2.0
B       2.0     2.0     3.0
C       3.0     5.166666666666666       4.666666666666666
D       4.666666666666666       3.0     5.666666666666666
  • note: the qnorm cli assumes that the first column and the first row are used as descriptors, and are "ignored" in the quantile normalization process. Lines starting with a hashtag "#" are treated as comments and ignored.

  • note: The CLI requires pandas to be installed (conda/pip install pandas)

Installation

pip

user@comp:~$ pip install qnorm

conda

Installing qnorm from the conda-forge channel can be achieved by adding conda-forge to your channels with:

user@comp:~$ conda config --add channels conda-forge

Once the conda-forge channel has been enabled, qnorm can be installed with:

user@comp:~$ conda install qnorm

local

clone the repository

user@comp:~$ git clone https://github.com/Maarten-vd-Sande/qnorm

And install it

user@comp:~$ cd qnorm
user@comp:~$ pip install .

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

qnorm-0.8.1.tar.gz (16.8 kB view details)

Uploaded Source

Built Distribution

qnorm-0.8.1-py3-none-any.whl (15.1 kB view details)

Uploaded Python 3

File details

Details for the file qnorm-0.8.1.tar.gz.

File metadata

  • Download URL: qnorm-0.8.1.tar.gz
  • Upload date:
  • Size: 16.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.8.1 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.9.7

File hashes

Hashes for qnorm-0.8.1.tar.gz
Algorithm Hash digest
SHA256 61b2f3ef09a9c552a4f3b83dc438cb13f191fa190164361a3a508c4777eed3c7
MD5 ecb8f8a53c18a628e4e904b77c69bd69
BLAKE2b-256 0124be3f1e5bcc80733b99f432efa13a9130ab84c52a52eef323ec30d79ba9b6

See more details on using hashes here.

File details

Details for the file qnorm-0.8.1-py3-none-any.whl.

File metadata

  • Download URL: qnorm-0.8.1-py3-none-any.whl
  • Upload date:
  • Size: 15.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.8.1 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.9.7

File hashes

Hashes for qnorm-0.8.1-py3-none-any.whl
Algorithm Hash digest
SHA256 9d6ce4e82444155922baf06aa89f9f939b54f53844e340bf2c6d9e7ff8821c41
MD5 562fac50213e0c541410688b1e399d4a
BLAKE2b-256 45bc80350513f3abf14f6dd921bf083f241ebc826c94556931c51f05348fd442

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page