Skip to main content

Project with code needed to filter, trim and slim ntuples produced by AP

Project description

Description

This project is used to:

  • Filter, slim, trim the trees from a given AP production
  • Rename branches
  • Download the outputs

This is done using configurations in a YAML file and through DIRAC jobs.

Check this for installation instructions and for instructions on how to setup an environment to use this project.

Submitting jobs

Check latest version of virtual environment

All the jobs below require code that lives in a virtual environment, there should be multiple versions of this environment and the latest one should be obtained by running:

dirac-dms-user-lfns -w dcheck.tar -b /lhcb/user/${LXNAME:0:1}/$LXNAME/run3/venv

currently, the latest is 023. Unless you have made your own tarballs, LXNAME=acampove.

Submit jobs

To run the filtering, after properly installing the project, as shown here do:

# Local will create a local sandbox, use wms to send to the grid

# For data, there are about 11K ROOT files in the input, 11K jobs should do one file per job, with -t, only first job will be done
job_filter -n data_test_job -p rd_ap_2024             -s       data -c /home/acampove/Packages/config_files/post_ap/v1.yaml -j 11000 -e 023 -u acampove -m local -t

# For real jobs, 1K jobs should be enough
job_filter -n data_job      -p rd_ap_2024             -s       data -c /home/acampove/Packages/config_files/post_ap/v1.yaml -j 1000 -e 023 -u acampove -m local

# For MC using noPID samples, there are only 44 input ROOT files, therefore at most 44 jobs are possible
job_filter -n mc_job        -p -btoxll_mva_2024_nopid -s simulation -c /home/acampove/Packages/config_files/post_ap/v1.yaml -j   44 -e 023 -u acampove -m local

where the options mean:

  -h, --help            show this help message and exit
  -n NAME  --name NAME  Name of job, needed for dirac naming and to name output
  -p PROD, --prod PROD  Name of production, e.g. rd_ap_2024, this shoudl be the same as in the config section.
  -s SAMP, --samp SAMP  Sample nickname found in the config section `samples`
  -c CONF, --conf CONF  Path to config file, which should be a YAML file and a few examples are linked below.
  -j NJOB, --njob NJOB  Number of grid jobs, this will depend on the number of files, for data typically 11K, and 1000 jobs would suffice
  -e VENV, --venv VENV  Index of virtual environment, e.g. 023
  -u USER, --user USER  User associated to venv, currently acampove should be the only choice, but if you author your own virtual environment and upload it, then this should be your user name
  -m {local,wms}, --mode {local,wms} Run locally (for tests) or in the grid
  -t       --test       If used, will send only one job

Regarding the name, the output will go to a directory in EOS named JOBNAME_SAMPLENAME, e.g. test_001_data if -n test_001 is used on the data sample. Some config files can be found here

Downloading ntuples

A test would look like:

run3_download_ntuples -j flt_004 -n 3 [-d $PWD/files]

where:

-j: Is the name of the job, which has to coincide with the directory name, where the ntuples are in EOS, e.g. /eos/lhcb/grid/user/lhcb/user/a/acampove/flt_004.
-n: Number of ntuples to download, if not pased, will download everything.
-d: Directory where output ntuples will go, if not passed, directory pointed by DOWNLOAD_NTUPPATH will be used.

A real download would look like:

run3_download_ntuples -j flt_001 -m 40

Where -m denotes the number of threads used to download, -j the name of the job.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

post_ap-0.1.3.tar.gz (24.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

post_ap-0.1.3-py3-none-any.whl (28.0 kB view details)

Uploaded Python 3

File details

Details for the file post_ap-0.1.3.tar.gz.

File metadata

  • Download URL: post_ap-0.1.3.tar.gz
  • Upload date:
  • Size: 24.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/5.1.1 CPython/3.12.7

File hashes

Hashes for post_ap-0.1.3.tar.gz
Algorithm Hash digest
SHA256 23803d007ae14c5f79ea1afa23d9cfb9d50db5755c20ce2e9390b7f4b95f3bdf
MD5 764ebc302ca75d409781c8d6be7d8762
BLAKE2b-256 e0a25b59ec5405c98adf45c832e7b95c390507e3181b7adc1e08de8c0684b165

See more details on using hashes here.

Provenance

The following attestation bundles were made for post_ap-0.1.3.tar.gz:

Publisher: publish.yaml on acampove/post_ap

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file post_ap-0.1.3-py3-none-any.whl.

File metadata

  • Download URL: post_ap-0.1.3-py3-none-any.whl
  • Upload date:
  • Size: 28.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/5.1.1 CPython/3.12.7

File hashes

Hashes for post_ap-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 8bf6f71ef2d2b4b4c293157c3e86e0200b06d88d020de88673a89c8cd1a32adc
MD5 71818e4fa02375cc3fbb9e72f321a36c
BLAKE2b-256 1c5f880c9e245437740b770434182f75fc9cb2254a59db89e5f9c929cdb2068d

See more details on using hashes here.

Provenance

The following attestation bundles were made for post_ap-0.1.3-py3-none-any.whl:

Publisher: publish.yaml on acampove/post_ap

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page