offline-processing and pipeline managment for HERA data analysis
Project description
hera_opm
hera_opm
provides a convenient and flexible framework for developing data
analysis pipelines for operating on HERA data. It facilitates "offline
processing", and is portable enough to operate on computer clusters with
batch submission systems or on local machines.
How It Works
The hera_opm
package uses the makeflow
system, which is a part of the
Cooperative Computing Tools
package developed by the
Cooperative Computing Lab. The hera_opm
package
essentially converts a pipeline defined in a configuration file into a format
that can be parsed by makeflow
. This process is also aware of aspects specific
to HERA data, such as the polarization features of the data, in order to build
an appropriate software pipeline. Once the makeflow
instructions file has been
generated, the makeflow
program itself is used to execute the steps in the
pipeline.
There are generally 5 steps required to "build a pipeline":
- Write task scripts that will be executed by
makeflow
for a given stage in the pipeline. These scripts should generally be as atomic as possible, and perform only a single logical component of a pipeline (though it may in turn call several supporting scripts or commands). - Write a configuration file which defines the order of tasks to be completed. This configuration file defines the logical flow of the pipeline, as well as prerequisites for each task. It also allows for defining compute and memory requirements, for systems that support resource management.
- Use the provided
build_makeflow_from_config.py
script to build amakeflow
instruction file that specifies the pipeline tasks applied to the data files. - Use the provided
makeflow_nrao.sh
ormakeflow_local.sh
to execute the pipeline in either the NRAO batch scheduler environment, or on a local machine, respectively. - (Optional) Use the provided
clean_up_makeflow.py
to clean up the work directory for makeflow. This will remove the wrapper scripts and output files, and generate a single log file for all jobs in the makeflow.
Installation
To install the hera_opm
package, simply:
pip install .
As mentioned above, hera_opm
uses makeflow
as the backing pipeline management
software. As such, makeflow
must be installed. To install makeflow
in your
home directory:
git clone https://github.com/cooperative-computing-lab/cctools.git
cd cctools
./configure --prefix=${HOME}/cctools
make clean
make install
export PATH=${PATH}:${HOME}/cctools/bin
For convenience, it is helpful to add the export
statement to your .bashrc
file, so that the makeflow
commands are always on your PATH
.
Dependencies
When installing the package, setuptools will attempt to download and install any missing dependencies. If you prefer to manage your own python environment (through conda or pip or some other manager), you can install them yourself.
Required
- toml >= 0.9.4
Optional
Generating an lstbin
pipeline (instead of analysis
) requires that hera_cal
be installed. The main package and tests can be run without this requirement.
Task Scripts and Config Files
For documentation on building task scripts, see the task scipts docs page. For documentation on config files, see the config file docs page.
Testing
hera_opm
uses pytest
as its testing framework. To run the test suite, do:
pytest
from the root repo directory. This may require running pip install .[test]
to
install testing dependencies.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.