Skip to main content

Package to read SWIFT simulation snapshots in MPI.

Project description

An MPI read routine for Swift simulation snapshots

pyread_swift is an MPI read routine for swiftsim snapshots, very similar in style to John Helly's read_eagle code to read EAGLE snapshots.

The package can read swiftsim snapshots both in "collective" (i.e., multiple MPI ranks read from a single file simultaneously) and "distributed" (i.e., each MPI reads an individual snapshot file part in isolation) modes.

Can also read in pure serial, if no MPI libraries are installed.

Installation

Requirements

  • OpenMPI or other MPI library
  • python>=3.8

Recommended modules when working on COSMA7:

  • module load gnu_comp/11.1.0 openmpi/4.1.4 parallel_hdf5/1.12.0 python/3.9.1-C7

Installation from source

Given the need for a parallel HDF5 installation, it is recommended you install pyread_swift within a virtual/conda environment. However you can ofcourse also install directly into your base Python environment if you prefer.

First make sure your pip is up-to-date:

  • python3 -m pip install --upgrade pip

Then you can install the pyread_swift package by typing the following in the root git directory:

  • python3 -m pip install .[mpi]

which will install pyread_swift and any dependencies (omit [mpi] for pure serial version).

MPI installation for collective reading

If you are using pyread_swift to load large snapshots over MPI collectively (i.e., multiple cores read in parallel from the same file), a bit of additional setup is required.

Make sure you have hdf5 installed with parallel compatibility (see here for details).

Then, uninstall any installed versions of mpi4py and h5py:

  • python3 -m pip uninstall mpi4py h5py

and reinstall then from source, with MPI flags:

  • MPICC=mpicc CC=mpicc HDF5_MPI="ON" python3 -m pip install --no-binary=mpi4py mpi4py

  • MPICC=mpicc CC=mpicc HDF5_MPI="ON" python3 -m pip install --no-binary=h5py h5py

If pip struggles to find your HDF5 libraries automatically, e.g., error: libhdf5.so: cannot open shared object file: No such file or directory. You may have to specify the path to the HDF5 installation manually, i.e., HDF5_DIR=/path/to/hdf5/lib (see here for more details).

For our COSMA7 setup, that would be:

HDF5DIR="/cosma/local/parallel-hdf5//gnu_11.1.0_ompi_4.1.4/1.12.0/"

Usage

pyread_swift is build around a primary read wrapper, called SwiftSnapshot. The snapshot particles are loaded into, stored, and manipulated by this object.

Reading follows the same four steps (see also the examples below):

  • Initialize a SwiftSnapshot object pointing to the location of the HDF5 file.

  • Select the spatial region you want to extract the particles from using the select_region() routine.

  • Split the selection over the MPI ranks using the split_selection() routine.

  • Read a selected property of the particles using the read_dataset() routine.

Input parameters to SwiftSnapshot

Input Description Default option
fname Full path to HDF5 snapshot file. If the snapshot is split over multiple files, this can just be one of the file parts -
comm= MPI4PY communicator (if reading in MPI) None
verbose= True for more a more verbose output False
mpi_read_format= How to read the snapshot in MPI mode ("collective" or "distributed")

"collective": Do a collective read of each file, i.e., all ranks read a single file at one. Recommended for single, or few large snapshot file(s). Requires parallel-hdf5 to be installed.

"distributed": Each rank reads its own file part. Recommended for multiple smaller files.
"collective"
max_concur_io= When reading in MPI, how many HDF5 files can be open at once 64

Example usage (No MPI case)

from pyread_swift import SwiftSnapshot

# Set up pyread_swift object pointing at HDF5 snapshot file (or a file part). 
snapshot = "/path/to/snap/part.0.hdf5"
swift = SwiftSnapshot(snapshot)

# Select region to load from.
parttype = 1 # Dark matter
region = [0,100,0,100,0,100] # [xlo,xhi,ylo,yhi,zlo,zhi]
swift.select_region(parttype, *region)

# Divide selection between ranks (needs to be invoked even for non-mpi case).
swift.split_selection()

# Read data.
ids = swift.read_dataset(parttype, "ParticleIDs")

Example usage (MPI case)

from mpi4py import MPI
from pyread_swift import SwiftSnapshot

# MPI communicator.
comm = MPI.COMM_WORLD

# Set up read_swift object pointing at HDF5 snapshot file (or a file part). 
snapshot = "/path/to/snap/part.0.hdf5"
swift = SwiftSnapshot(snapshot, comm=comm)

# Select region to load from.
parttype = 1 # Dark matter
region = [0,100,0,100,0,100] # [xlo,xhi,ylo,yhi,zlo,zhi]
swift.select_region(parttype, *region)

# Divide selection between ranks.
swift.split_selection()

# Read data.
ids = swift.read_dataset(parttype, "ParticleIDs")

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pyread_swift-1.0.0.tar.gz (16.2 kB view details)

Uploaded Source

Built Distribution

pyread_swift-1.0.0-py3-none-any.whl (16.5 kB view details)

Uploaded Python 3

File details

Details for the file pyread_swift-1.0.0.tar.gz.

File metadata

  • Download URL: pyread_swift-1.0.0.tar.gz
  • Upload date:
  • Size: 16.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.12

File hashes

Hashes for pyread_swift-1.0.0.tar.gz
Algorithm Hash digest
SHA256 cf0d1ac91849ad768711c996894b3ec9220f6e124136c84be39d72e5124f9c9d
MD5 847ef0b205e6ac6178fe0e8b1c21cb2e
BLAKE2b-256 a3113986c6e894a54c0fb6684852b9225b7cabe54fa01fefeb38816ef07df293

See more details on using hashes here.

File details

Details for the file pyread_swift-1.0.0-py3-none-any.whl.

File metadata

File hashes

Hashes for pyread_swift-1.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 e53c1919e4a967ed3fe928e68ad50d132a50760095d64a2779e66f4753454f7d
MD5 d4512926847e395b174e038cfcf4541a
BLAKE2b-256 81d9364eda17bb29fa2bccee1668a933151dbb168e82786c1d9ea3ff3e4c2191

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page