Skip to main content

Calculate 2D and 3d acceptance model for creating maps with gammapy (IACT analysis)

Project description

Description

This package create background model (or acceptance model) to be used for IACT analysis with gammapy BAccMod is licensed under the GNU Lesser General Public License (LGPL) v3.0.

Installation

pip install BAccMod

Dependencies :

  • numpy
  • scipy
  • astropy
  • iminuit >= 2.0
  • gammapy >= 1.1
  • regions >= 0.7

Example of use

Basic use

User interface is through two main functions available in any of the modelisation classes. One function allows for the creation of a background IRF model or collection of models which can be used in later steps. The second function applies a model, either created dynamically or created before, to observations to associate a new background IRF to each one.

Example

Below is an example of the creation of a single background model with radial symmetry using directly the observations of interest, and the application of this model to all the observations. This is not the adviced usage of the package.

from gammapy.maps import MapAxis
from gammapy.data import DataStore
from regions import CircleSkyRegion
import astropy.units as u
from astropy.coordinates import SkyCoord
from baccmod import RadialAcceptanceMapCreator

# The observations to use for creating the acceptance model
data_store = DataStore.from_dir("$GAMMAPY_DATA/hess-dl3-dr1")
obs_collection = data_store.get_observations([23523, 23526, 23559, 23592])

# The exclusion regions to apply during acceptance model calculation
exclude_regions = [CircleSkyRegion(center=SkyCoord.from_name('Crab'),
                                   radius=0.2 * u.deg), ]

# Define the binning of the model
e_min, e_max = 0.1 * u.TeV, 10. * u.TeV
size_fov = 2.5 * u.deg
offset_axis_acceptance = MapAxis.from_bounds(0. * u.deg, size_fov, nbin=6, name='offset')
energy_axis_acceptance = MapAxis.from_energy_bounds(e_min, e_max, nbin=6, name='energy')

First step is to create an instance of one of the model creation class. Here the RadialAcceptanceMapCreator to which we provide minimal inputs : the axis binning and exclusion region to remove the Crab Nebula.

acceptance_model_creator = RadialAcceptanceMapCreator(energy_axis_acceptance,
                                                      offset_axis_acceptance,
                                                      exclude_regions=exclude_regions)
acceptance_model = acceptance_model_creator.create_model(obs_collection)

You can then check the acceptance model by plotting it using

acceptance_model.peek()

To use it with gammapy, you could first save it on a FITS file

hdu_acceptance = acceptance_model.to_table_hdu()
hdu_acceptance.writeto('acceptance.fits', overwrite=True)

It's then possible to load the acceptance model in the current gammapy DataStore with this code. You would need then to recreate your gammapy Observations object in order than the acceptance model is taken into account for the analysis.

data_store.hdu_table.remove_rows(data_store.hdu_table['HDU_TYPE']=='bkg')
for obs_id in np.unique(data_store.hdu_table['OBS_ID']):
    data_store.hdu_table.add_row({'OBS_ID': obs_id, 
                                 'HDU_TYPE': 'bkg',
                                 "HDU_CLASS": "bkg_2d",
                                 "FILE_DIR": "",
                                 "FILE_NAME": 'acceptance.fits',
                                 "HDU_NAME": "BACKGROUND",
                                 "SIZE": hdu_acceptance.size})
data_store.hdu_table = data_store.hdu_table.copy()

obs_collection = data_store.get_observations([23523, 23526, 23559, 23592])

data_store.hdu_table

[!WARNING]
Due to some change in gammapy 1.3, it's strongly suggested to use only use models that have been created with the same version of gammapy used

Available model

All models have an identical interface. You just need to change the class used to change the model created.

There are two model currently available :

  • A 2D model with hypothesis of a radial symmetry of the background across the FoV. This is the class RadialAcceptanceMapCreator.
    from baccmod import RadialAcceptanceMapCreator
    acceptance_model_creator = RadialAcceptanceMapCreator(energy_axis_acceptance,
                                                          offset_axis_acceptance,
                                                          exclude_regions=exclude_regions)    
    
  • A 3D model with a regular grid describing the FoV. This is the class Grid3DAcceptanceMapCreator. This class support the use of either a direct event stacking or the fitting of spatial models to create the IRF.
    from baccmod import Grid3DAcceptanceMapCreator
    acceptance_model_creator = Grid3DAcceptanceMapCreator(energy_axis_acceptance,
                                                          offset_axis_acceptance,
                                                          exclude_regions=exclude_regions,
                                                          method='stack' or 'fit' (default is 'stack'))
    

Each class accepts a set of optionnal parameters described in its docstring. The majority are common and allow to activate or control options such as splitting of the data in azimuth or in zenith bins, and interpolation. Some are described later in this document.

Model creation

Once the acceptance_model_creator is initated, the method create_acceptance_model can be used to create and output a background model. While the stteings of the zenith binning or interpolation are defined during the class object creation, the decision to use one or the other is done at the time of the call :

model = acceptance_model_creator.create_acceptance_model(# observations used to create the models
                                                         off_observations,
                                                         # Activates the zenith binning
                                                         zenith_binning = False
                                                )

A model created with zenith binning will be valid for application with the zenith interpolation later.

Observation-wise IRF application

To associate a background IRF to a set of observations, the user should use the function create_acceptance_map_per_observation. It can either, and in this order of priority, apply a provided model, create a model from off observations and apply it, or create a model fron the target observation and apply it.

# acceptance_map will be a dictionnary of obs_id to BackgroundIRF
acceptance_models = acceptance_model_creator.create_acceptance_map_per_observation(
  observations, # Target observations
  zenith_binning = False, # Activates the zenith binning (ignored if interpolating)
  zenith_interpolation = False, # Activates the zenith interpolation
  runwise_normalisation = True, # If True, the IRF for the target observations is renormalised to the observed counts
  off_observations = None, # observations to use to create a model if none is provided. If None, `observations`is used.
  base_model = None # model to apply if provided
)

Logging setup

The module uses logging to output various information to the user. Standard logging functionalities are supported.

In addition, some helper functions are implemented and BAccMod uses a logging level MOREINFO with value 15 between the logging INFO and DEBUG levels.

Logging functions can be used as follow :

from baccmod.logging import MOREINFO, set_log_level

set_log_level(level= MOREINFO, # Supports both int and string representation, e.g.
                               # 15, MOREINFO and 'MOREINFO' would be equivalent
              module= 'baccmod' # default 'baccmod', 'baccmod.' is automatically 
                                # added if missing so users can provide a 
                                # submodule directly
              )

# Example usage : 
set_log_level(logging.INFO)
set_log_level('WARNING')
set_log_level(20, 'grid3d_acceptance_map_creator')
# equivalent to
set_log_level(level=logging.INFO,
              module='baccmod.grid3d_acceptance_map_creator')

The set_log_level function has some simple usages show below but takes a number of optionnal parameters for more advance configuration. Its signature is :

setup_logging_output(output_file=None, use_terminal=True,
                     handlers=None, module='baccmod',
                     fmode='w', term_level=logging.NOTSET,
                     log_format='%(levelname)s:%(name)s: %(message)s')

Some usages :

from baccmod.logging import set_log_level, setup_logging_output

# Setup the output of logging to the provided file. Keep output in terminal.
setup_logging_output(output_file='/path/to/log_file.log') 

# Setup the output of logging to the provided file. No output in terminal.
setup_logging_output(output_file='/path/to/log_file.log', use_terminal=False)

# Setup a file output with a global log level of INFO,
# but only output WARNING and above levels to the terminal
set_log_level('INFO')
setup_logging_output(output_file='/path/to/log_file.log',
                     use_terminal=True,
                     term_level='WARNING')

module can be used to define the logging for a submodule only.

fmode defines the writing strategy for the log file (default overwrites).

log_format can be used to change the format of the messages.

handlers (advanced users) allow to directly provide a list of Handlers instead of the one created by the output_file and use_terminal options.

Telescope position

The observations should contain the telescope position in order to have the algorithm working. If the information is missing in the DL3, you could either add it or it possible to add it directly to the observation as shown in the example below. This snippet of code may require some adaptation depending on the version of gammapy you are using.

from astropy.coordinates import EarthLocation

# Your telescope position in an EarthLocation object
loc = EarthLocation.of_site('Roque de los Muchachos')

# Add telescope position to observations
for i in obs_collection:
    obs_collection[i].obs_info['GEOLON'] = loc.lon.value
    obs_collection[i].obs_info['GEOLAT'] = loc.lat.value
    obs_collection[i].obs_info['GEOALT'] = loc.height.value
    obs_collection[i]._location = loc

Zenith binned model

It's also possible to create model binned per cos zenith. For this in the methods create_acceptance_model or create_acceptance_map_per_observation set the option zenith_binning at True. The minimum width of zenith bin can be controlled at the creation of the object with the parameter initial_cos_zenith_binning. You can chose the cos_zenith_binning_method: min_livetime or min_n_observation to give a condition on minimum livetime or number of observations for each bin. The algorithm will then automatically rebin to larger bin in order to have in each bin at least cos_zenith_binning_parameter_value livetime (in seconds) or observation per bin. If you add _per_wobble to the method name, wobbles will be identified and the binning will require the condition to be fulfilled for each identified wobble. (with the observation ID as index). Set verbose to True to get the binning result and a plot of the binned livetime.

Example :

acceptance_model_creator = RadialAcceptanceMapCreator(energy_axis_acceptance,
                                                      offset_axis_acceptance,
                                                      exclude_regions=exclude_regions,
                                                      oversample_map=10,
                                                      cos_zenith_binning_method='min_livetime_per_wobble',
                                                      cos_zenith_binning_parameter_value=3600,
                                                      initial_cos_zenith_binning=0.01)
acceptance_models = acceptance_model_creator.create_acceptance_map_per_observation(obs_collection,
                                                                                   zenith_binning=True)

Zenith interpolated model

It's also possible to create model interpolated between the binned model per cos zenith. For this use the method create_acceptance_map_per_observation but with the option zenith_interpolation set at True. All the parameters controlling the cos zenith binning remain active.

acceptance_model_creator = RadialAcceptanceMapCreator(energy_axis_acceptance,
                                                      offset_axis_acceptance,
                                                      exclude_regions=exclude_regions,
                                                      oversample_map=10,
                                                      min_run_per_cos_zenith_bin=3,
                                                      initial_cos_zenith_binning=0.01,
                                                      interpolation_zenith_type = 'linear' or 'log')
acceptance_models = acceptance_model_creator.create_acceptance_map_per_observation(obs_collection,
                                                                                   zenith_binning=True, # ignored
                                                                                   zenith_interpolation=True)

Using OFF runs for background model

It is possible to create a model from OFF runs and to apply in ON runs you want to analyse. The exclusions regions should cover potential sources both in the ON and OFF runs at the same time, or the model can be created beforehand. The OFF runs don't need to be spatially connected.

acceptance_models = acceptance_model_creator.create_acceptance_map_per_observation(obs_collection,
                                                                                   off_observations=obs_collection_off,
                                                                                   zenith_binning=True,
                                                                                   zenith_interpolation=True)

Store background model for later application

It could be in some case useful to precompute a model and to apply it on data later. The example below cover the case where you want to use a model per run using zenith interpolation and OFF runs.

However, it's possible to use this functionality without OFF runs or zenith interpolation. In the last case you just need to provide a model in a gammapy format.

There are at this stage no define file format for storing the intermediate results, we suggest to store the BackgroundCollectionZenith object created directly using pickle.

Creating the model

The example below creates the BackgroundCollectionZenith object containing the zenith binned model. The obs_collection provided could either be your ON runs if you want to compute background directly from the data or OFF runs if you want to use other data for the background model.

acceptance_model_creator = RadialAcceptanceMapCreator(energy_axis_acceptance,
                                                      offset_axis_acceptance,
                                                      exclude_regions=exclude_regions,
                                                      oversample_map=10,
                                                      min_run_per_cos_zenith_bin=3,
                                                      initial_cos_zenith_binning=0.01)
base_model = acceptance_model_creator.create_acceptance_model(obs_collection,
                                                              zenith_binning=True)

Storing and loading a model with pickle

The BackgroundCollectionZenith object containing the models could then be store using pickle.

import pickle
with open('my_bkg_model.pck', mode='wb') as f:
    pickle.dump(base_model, f)

It is then possible to retrieve it later using pickle.

import pickle
with open('my_bkg_model.pck', mode='rb') as f:
    base_model = pickle.load(f)

Applying a model in memory

When you have a precomputed model in memory, it is possible to apply it directly on a given set of runs by using the base_mode parameter.

acceptance_model_creator = RadialAcceptanceMapCreator(energy_axis_acceptance,
                                                      offset_axis_acceptance,
                                                      exclude_regions=exclude_regions)
acceptance_models = acceptance_model_creator.create_acceptance_map_per_observation(obs_collection,
                                                                                   base_model=base_model,
                                                                                   zenith_binning=True,
                                                                                   zenith_interpolation=True)

Compute background model with a higher time resolution than the observation run

If the background evolve quickly, like at high zenith angle, you could compute in the case of zenith binned or interpolated background the model at a smaller time scale than the observation run. The background model for the run will then correspond to the average of all the model computed for each part of the run. It should improve accuracy of the model at the expanse of a larger compute time. For this you need to set use_mini_irf_computation = True and you could control the time resolution used for computation with the parameter mini_irf_time_resolution.

acceptance_model_creator = RadialAcceptanceMapCreator(energy_axis_acceptance,
                                                      offset_axis_acceptance,
                                                      exclude_regions=exclude_regions,
                                                      oversample_map=10,
                                                      min_run_per_cos_zenith_bin=3,
                                                      initial_cos_zenith_binning=0.01,
                                                      use_mini_irf_computation = True,
                                                      mini_irf_time_resolution = 1. * u.min)
acceptance_models = acceptance_model_creator.create_acceptance_map_per_observation(obs_collection,
                                                                                   zenith_binning=True,
                                                                                   zenith_interpolation=True)

Compute background model with a non regular energy binning

For all models it's possible to compute model on a non regular energy binning. The output model will then be interpolated to the energy_axis provided. The main advantage of this feature is too allow for a finner binning at low energy where they are a high statistics available and coarser one at high energy where less events statistics is available. This energy binning should be passed through the energy_axis_computation argument when creating object. An example :

energy_axis_computation = MapAxis.from_energy_edges((list(np.geomspace(0.05, 0.5, 6)) + list(np.geomspace(0.5, 5, 3)[1:])) * u.TeV, name='energy')
acceptance_model_creator = RadialAcceptanceMapCreator(energy_axis_acceptance,
                                                      offset_axis_acceptance,
                                                      exclude_regions=exclude_regions,
                                                      oversample_map=10)
acceptance_models = acceptance_model_creator.create_acceptance_map_per_observation(obs_collection)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

baccmod-0.4.1.tar.gz (54.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

baccmod-0.4.1-py3-none-any.whl (56.3 kB view details)

Uploaded Python 3

File details

Details for the file baccmod-0.4.1.tar.gz.

File metadata

  • Download URL: baccmod-0.4.1.tar.gz
  • Upload date:
  • Size: 54.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.3

File hashes

Hashes for baccmod-0.4.1.tar.gz
Algorithm Hash digest
SHA256 13f894b55fb65df89ba5e2a2e4d2f82a55c5afcfef37b78077bbd9d6561970e3
MD5 d2f38fced2b70374e37ed9de12c5f385
BLAKE2b-256 2135b18568f5f5b586d75af411b832c24d3b4d8568c6f0926ce5e062df5ec7fd

See more details on using hashes here.

File details

Details for the file baccmod-0.4.1-py3-none-any.whl.

File metadata

  • Download URL: baccmod-0.4.1-py3-none-any.whl
  • Upload date:
  • Size: 56.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.14.3

File hashes

Hashes for baccmod-0.4.1-py3-none-any.whl
Algorithm Hash digest
SHA256 1f2d500f38330c0251645f47fe74aa5585f56d2c80e4adec370e73773295e744
MD5 615e3b839a7fe7edc27345fd618d3227
BLAKE2b-256 1a870b9d2f09adb6d2a56586b6c8dcd857848ecffb2618f281a858df77ec47ed

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page