Skip to main content

Data analysis of cosmology simulations in parallel.

Project description

nbodykit
========

nbodykit is a software kit for cosmological datasets from
N-body simulations and large scale structure surveys.

Driven by the optimistism regarding the abundance and availability of
large scale computing resources in the future,
the development of nbodykit
distinguishes itself from other similar software packages
([nbodyshop]_, [pynbody]_, [yt]_, [xi]_) by focusing on :

- a **unified** treatment of simulation and observational data sets;
insulating algorithms from data containers;

- reducing wall-clock time by **scaling** to thousands of cores;

- **deployment** and availability on large super computing facilities.

All algorithms are parallel, and run with Message Passing Interface (MPI).

.. todo::

Nbodykit is a set of extension points and a(two) main programs
that wire the plugins together.

We need to write about this in a graceful way. The extension framework
sets the design of nbodykit aside from others.

Build Status
------------

We perform integrated tests for the toplevel executables in a
miniconda environment for Python 2.7 and Python 3.4.

.. image:: https://api.travis-ci.org/bccp/nbodykit.svg
:alt: Build Status
:target: https://travis-ci.org/bccp/nbodykit/
.. image:: https://coveralls.io/repos/github/bccp/nbodykit/badge.svg?branch=develop
:alt: Test Coverage
:target: https://coveralls.io/github/bccp/nbodykit?branch=develop


Top Level Executables
---------------------

The algorithms implemented in nbodykit can be invoked by

- ``bin/nbkit.py``: the main executable that can run a variety of algorithms (see below)

- ``bin/nbkit-batch.py``: runs algorithms in batch-mode, using a task manager to control the iteration over a set of configuration files

Run them with '-h' to see the inline help.

.. code:: bash

nbkit.py --help

For mpirun / mpiexec, it is import to launch nbkit.py via Python:

.. code:: bash

mpirun -n 4 python `which nbkit.py` AlgorithmName [config]


Algorithms
----------

Algorithms are implemented as Plugins.

To obtain a list of algorithms, use

.. code:: bash

nbkit.py --list-algorithms

To see help of a particular algorithm, use

.. code:: bash

nbkit.py AlgorithmName --help

or

.. code:: bash

nbkit.py --list-algorithms AlgorithmName

where AlgorithmName can be one of the following:

- **FFTPower** : Power spectrum (Isotropic, Anisotropic, and multipoles) in a periodic box, via Fast Fourier Transforms

- **BianchiFFTPower** : power spectrum multipoles using FFTs for a data survey with non-trivial geometry, as detailed in [Bianchi]_ et al. 2015

- **FFTCorrelation**: Correlation function (Isotropic, Anisotropic, and multipoles) in a periodic box, via Fast Fourier Transforms

- **PairCountCorrelation**: Correlation function (Isotropic, Anisotropic, and multipoles) via simple pair-counting

- **FOF** : Friend of Friend halos (usually called groups)

- **FOF6D** : 6D Friend of Friend halos (usually called subhalos or galaxies)

- **Describe** : Describes the max and min of a column from any DataSource

- **Subsample** : Create a subsample from a DataSource, and evaluate density (:math:`1 + \delta`) smoothed
at the given scale

- **TraceHalo** : Trace the center of mass and velocity of particles between two different DataSource, joining
by ID.

- **FiberCollisions** : Assign fibers to a galaxy survey assuming a fixed angular collision radius, such that the number of galaxies collided out of the survey is minimized

It is very easy to contribute a new algorithm to nbodykit. We will document this later.

.. todo::

Write about this.

Datasources
-----------

nbodykit insulates algorithms from data containers via 'DataSource' plugins.

The datasource ensures the input to the algorithms are transformed to MPC/h units,
and particles are moved to the redshift space (if requested).

To obtain a list of datasources, use

.. code:: bash

nbkit.py --list-datasources

or to obtain help for a particular DataSource, use

.. code:: bash

nbkit.py --list-datasources DataSourceName

The list of DataSource plugins:

- **FOFGroups** : Friend of Friend catalog generated by the FOF algorithm.

- **HaloLabel** : Halo label (halo id per particle) file by the FOF algorithm.

- **FastPM** : Snapshot files of FastPM

- **PlainText** : Plain text file

- **TPMLabel** : Halo label (halo id per particle) file generated from Martin's TPM.

- **Gadget** : A flavor of Gadget 2 files (experimental)

- **GadgetGroupTab** : A flavor of Gadget 2 FOF catalogue (experimental)

- **TPMSnapshot** : snapshot files from Martin's TPM.

- **Grid** : A special data source representing a simple 3D grid (not particles)

- **Pandas** : Pandas flavored HDF5 and text files.

- **RaDecRedshift** : read (ra, dec, z) from a plaintext file, returning Cartesian coordinates

- **TracerCatalog** : a catalog of tracer objects, with associated `data` and `randoms` catalogs to fully specify the geometry of the survey

It is very easy to contribute a new datasource to nbodykit. We will document this later.

.. todo::

Write about this.

The DataSource plugins implement three reading modes: simple-read, streaming and full-read:

- **simple-read** reads in the entire data set from a single rank, which is then split and scattered to other ranks. This is the minimal
requirement of a DataSource.

- **full-read** reads in the entire dataset; by default it calls simple-read to read.

- **streaming** reads in the data chunk by chunk; by default it does a full-read.

Plugins
-------

nbodykit can be extended by adding plugins. Several plugins are distributed together
with nbodykit main source code, but users can add new plugins as well.

.. todo::

Write about this.

Examples
--------

There are example scripts (which also act as integrated tests) in examples directory.
The supporting data for these scripts can be retrieved from

https://s3-us-west-1.amazonaws.com/nbodykit/nbodykit-data.tar.gz

Check get_data.sh for details.

Dependencies
------------

The software is built on top of existing tools. Please refer to their
documentations:

- [pfft]_ : massively parallel fast fourier transform with pencil domains
- [pfft-python]_ : python binding of pfft
- [pmesh]_ : particle mesh framework in Python
- [kdcount]_ : pair-counting and friend-of-friend clustering with KD-Tree
- [bigfile]_ : A reproducible massively parallel IO library for hierarchical data
- [MP-sort]_ : massively parallel sorting
- [sharedmem]_ : in-node parallelism with fork and copy-on-write.

Some better established dependencies are

- [scipy]_, [numpy]_ : the foundations for Scientific Python.
- [mpi4py]_ : MPI for python
- [h5py]_ : Support for HDF5 files

Optional Dependencies
---------------------

- [pandas]_, [pytables]_ are required to access the PANDAS subset of HDF5 and fast parsing of plain text files.

Build
-----

The software is designed to be installed with the ``pip`` utility like a regular
python package.

Using nbodykit from the source tree is not supported. See 'Development mode' for
details.

The steps listed here is intended for a commodity Linux based cluster
(e.g. a Rocks Cluster [rocksclusters]_) or a Linux based workstation / laptop.
Please note that there are slight changes to the procedure on systems running
a Mac OS X operating system and
Cray super-computers
as explictly noted below in `Special notes for Mac and Cray`_.

Install the main ``nbodykit`` package, as well as the external dependencies
listed above, into the default python installation directory with:

.. code:: sh

git clone http://github.com/bccp/nbodykit
cd nbodykit

# It may take a while to build fftw and pfft.
# Mac and Edison are special, see notes below

pip install -r requirements.txt
pip install -U --force --no-deps .

A different installation directory can be specified via the ``--user`` or ``--root <dir>``
options of the ``pip install`` command.

The pure-python ``nbodykit`` package (without external dependencies) can be installed by
omitting the ``-r requirements.txt`` option, with such an installation only requiring ``numpy``.
The caveat being that the functionality of the package is greatly diminished -- package behavior
in this instance is not tested and considered undefined.


The dependencies of nbodykit are not fully stable, thus we recommend updating
the external dependencies occassionally via the ``-U`` option of ``pip install``.
Also, since nbodykit is
not yet stable enough for versioned releases, ``--force`` ensures the current
sourced version is installed:

.. code:: sh

pip install -U -r requirements.txt
pip install -U --force --no-deps .

To confirm that nbodykit is working, we can type, in a interactive python session:
(please remember to jump to bin/ directory to avoid weird issues about importing in-tree)

.. code:: python

import nbodykit
print(nbodykit)

import kdcount
print(kdcount)

import pmesh
print(pmesh)

Or try the scripts in the bin directory:

.. code:: bash

cd bin/
mpirun -n 4 python-mpi nbkit.py -h

Development Mode
++++++++++++++++

nbodykit can be installed with the development mode (``-e``) of pip

.. code::

pip install -r requirements.txt -e .

In addition to the dependency packages, the 'development' installation
of nbodykit may require a forced update from time to time:

.. code::

pip install -U --force --no-deps -e .

It is sometimes required to manually remove the ``nbodykit`` directory in
``site-packages``, if the above command does not appear to update the installation
as expected.


Special notes for Mac and Cray
------------------------------

Mac Notes
+++++++++

autotools are needed on a Mac

.. code::

sudo port install autoconf automake libtool

On Mac, the `LDSHARED` environment variable must be explicitly set. In bash, the command is

.. code::

export LDSHARED="mpicc -bundle -undefined dynamic_lookup -DOMPI_IMPORTS"; pip install -r requirements.txt .

On recent versions of MacPorts, we also need to tell mpicc to use gcc rather than the default clang
compiler, which doesn't compile fftw correctly due to lack of openmp support.

.. code::

export OMPI_CC=gcc

NERSC Notes
+++++++++++

We maintain a ``daily`` build of nbodykit on NERSC systems that works with the 2.7-anaconda python via the python-mpi-bcast helper. Here is an example job script, that
runs a friend of friend finder (FOF) then measures the halo power spectrum for halos of mass greater than 1e13 (Msun/h).

.. code:: bash

#! /bin/bash

#SBATCH -n 32
#SBATCH -p debug
#SBATCH -t 10:00

set -x
module load python/2.7-anaconda

source /project/projectdirs/m779/nbodykit/activate.sh

srun -n 16 python-mpi $NBKITBIN FOF <<EOF
nmin: 10
datasource:
plugin: FastPM
path: data/fastpm_1.0000
linklength: 0.2
output: output/fof_ll0.200_1.0000.hdf5
calculate_initial_position: True
EOF

srun -n 16 python-mpi $NBKITBIN FFTPower <<EOF

mode: 2d # 1d or 2d
Nmesh: 256 # size of mesh

# here are the input fields
field:
DataSource:
plugin: FOFGroups
path: output/fof_ll0.200_1.0000.hdf5
m0: 2.27e12 # mass of a particle: use OmegaM * 27.75e10 * BoxSize ** 3/ Ntot
rsd: z # direction of RSD, usually use 'z', the default LOS direction.
# select: Rank < 1000 # Limits to the first 1000 halos for abundance matching or
select: LogMass > 13 # limit to the halos with logMass > 13 (LRGs).
output: output/power_2d_fofgroups.dat # output
EOF

Note that we need to provide the
mass of a single particle for FOFGroups. The number is :math:`27.75\times10^{10} \Omega_M (L / N)^3 h^{-1} M_\odot`, where :math:`L` is the boxsize in Mpc/h, and
:math:`N` is the number of particles per side -- :math:`N^3` is the total number of particles.

Detail on NERSC
----------------
If you would like to do development on NERSC (not recommended), then more work is involved.

To use nbodykit on a NERSC Cray system (e.g. [Edison]_, [Cori]_), we need to ensure the python environment
is setup to working efficiently on the computing nodes.

If darshan [darshan]_ or altd are loaded by default, be sure to unload them since they tend to interfere
with Python:

.. code::

module unload darshan
module unload altd

and preferentially, use GNU compilers from PrgEnv-gnu

.. code::

module unload PrgEnv-intel
module unload PrgEnv-cray
module load PrgEnv-gnu

then load the Anaconda [anaconda]_ python distribution,

.. code::

module load python/2.7-anaconda

We will need to set up the fast python start-up on a Cray computer, since
the default start-up scales badly with the number of processes. Our
preferred method is to use [fast-python]_ .

1. Modify the shell profile, and set PYTHONUSERBASE to a unique location.
(e.g. a path you have access on /project) for each machine.

For example, this is excertion from the profile of
a typical user on NERSC (``.bash_profile.ext``),
that has access to ``/project/projectdirs/m779/yfeng1``.

.. code:: bash

if [ "$NERSC_HOST" == "edison" ]; then
export PYTHONUSERBASE=/project/projectdirs/m779/yfeng1/local-edison
fi

if [ "$NERSC_HOST" == "cori" ]; then
export PYTHONUSERBASE=/project/projectdirs/m779/yfeng1/local-cori
fi

export PATH=$PYTHONUSERBASE/bin:$PATH
export LIBRARY_PATH=$PYTHONUSERBASE/lib
export CPATH=$PYTHONUSERBASE/include

2. Install nbodykit to your user base with ``pip install --user``.
Also, create a bundle (tarball) of nbodykit.
Repeat this step if nbodykit (or any dependency) is updated.

.. code:: bash

cd nbodykit;

MPICC=cc pip install --user -r requirements $PWD

# enable python-mpi-bcast (On NERSC)
source /project/projectdirs/m779/python-mpi/activate.sh

# create the bundle
MPICC=cc bundle-pip nbodykit.tar.gz -r requirements.txt $PWD



References
==========

.. [nbodyshop] http://www-hpcc.astro.washington.edu/tools/tools.html

.. [pynbody] https://github.com/pynbody/pynbody

.. [yt] http://yt-project.org/

.. [pfft-python] http://github.com/rainwoodman/pfft-python

.. [pfft] http://github.com/mpip/pfft

.. [pmesh] http://github.com/rainwoodman/pmesh

.. [kdcount] http://github.com/rainwoodman/kdcount

.. [sharedmem] http://github.com/rainwoodman/sharedmem

.. [MP-sort] http://github.com/rainwoodman/MP-sort

.. [h5py] http://github.com/h5py/h5py

.. [numpy] http://github.com/numpy/numpy

.. [scipy] http://github.com/scipy/scipy

.. [pandas] http://pandas.pydata.org/

.. [pytables] http://pandas.pydata.org/

.. [mpi4py] https://bitbucket.org/mpi4py/mpi4py

.. [fast-python] https://github.com/rainwoodman/python-mpi-bcast

.. [bigfile] https://github.com/rainwoodman/bigfile

.. [rocksclusters] http://rocksclusters.org

.. [xi] http://github.com/bareid/xi

.. [edison] https://www.nersc.gov/users/computational-systems/edison/

.. [cori] https://www.nersc.gov/users/computational-systems/cori/

.. [darshan] http://www.mcs.anl.gov/research/projects/darshan/

.. [anaconda] http://docs.continuum.io/anaconda/index

.. [Bianchi] http://arxiv.org/abs/1505.05341

Project details


Release history Release notifications

History Node

0.3.4

History Node

0.3.3

History Node

0.3.2

History Node

0.3.1

History Node

0.3.0

History Node

0.2.9

History Node

0.2.8

History Node

0.2.7

History Node

0.2.6

History Node

0.2.5

History Node

0.2.4

History Node

0.2.3

History Node

0.2.2

History Node

0.2.1

History Node

0.2.0

History Node

0.1.11

History Node

0.1.9

History Node

0.1.8

History Node

0.1.6

History Node

0.1.5

History Node

0.1.4

History Node

0.1.3

History Node

0.1.2

This version
History Node

0.1.1

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Filename, size & hash SHA256 hash help File type Python version Upload date
nbodykit-0.1.1.tar.gz (149.9 kB) Copy SHA256 hash SHA256 Source None Jun 16, 2016

Supported by

Elastic Elastic Search Pingdom Pingdom Monitoring Google Google BigQuery Sentry Sentry Error logging CloudAMQP CloudAMQP RabbitMQ AWS AWS Cloud computing DataDog DataDog Monitoring Fastly Fastly CDN DigiCert DigiCert EV certificate StatusPage StatusPage Status page