Skip to main content

Library to analyse, plot, and export data stored in HDF5 files.

Project description

CLS H5Analysis

This is a library to analyse, plot, and export HDF5 data. The package is meant to provide a framework to load data into jupyter and enable data interaction.

Installation

Install the package from PyPi with the pip package manager. This is the recommended way to obtain a copy for your local machine and will install all required dependencies.

    $ pip install h5analysis

You will also need Jupyter Notebook together with python 3 on your local machine.

In case that certain widgets aren't rendered properly, make sure to enable the appropriate jupyter extensions

    $ jupyter nbextension enable --py widgetsnbextension

Running

Launch your local jupyter installation with

    $ jupyter notebook

Examples

Load the required module

Before you start, you will need to import the required h5analysis package, and enable bokeh plotting.

## Setup necessarry inputs
from h5analysis.LoadData import *
from h5analysis.config import h5Config
from bokeh.io import show, output_notebook
output_notebook(hide_banner=True)

All data loaders require a proper configuration variable which is beamline/data format specific. An example configuration for the CLS REIXS beamline may look like this:

config = h5Config()

config.key("SCAN_{scan:03d}",'scan')
config.sca_folder('Data')

config.sca('Mono Energy','Data/beam')
config.sca('Mesh Current','Data/i0')
config.sca('Sample Current','Data/tey')
config.sca('TEY','Data/tey','Data/i0')
config.sca("MCP Energy", 'Data/mcpMCA_scale')
config.sca("SDD Energy", 'Data/sddMCA_scale')
config.sca("XEOL Energy", 'Data/xeolMCA_scale')

config.mca('SDD','Data/sddMCA','Data/sddMCA_scale',None)
config.mca('MCP','Data/mcpMCA','Data/mcpMCA_scale',None)
config.mca('XEOL','Data/xeolMCA','Data/xeolMCA_scale',None)

config.stack('mcpSTACK','Data/mcp_a_img','Data/mcpMCA_scale',None,None)

1d plots

sca = Load1d()
sca.load(config,'FileName.h5','x_stream','y_stream',1,2,3,4)  # Loads multiple scans individually
sca.add(config,'FileName.h5','x_stream','y_stream',1,2,3,4)  # Adds multiple scans
sca.subtract(config,'FileName.h5','x_stream','y_stream',1,2,3,4,norm=False) # Subtracts scans from the first scan
sca.xlim(lower_lim,upper_lim) # Sets the horizontal axis plot region
sca.ylim(lower_lim,upper_lim) # Sets the vertical axis plot region
sca.plot_legend("pos string as per bokeh") # Determines a specific legend position
sca.vline(position) # Draws a vertical line
sca.hline(position) # Draws a horizontal line
sca.label(pos_x,pos_y,'Text') # Adds a label to the plot
sca.plot() # Plots the defined object
sca.exporter() # Exports the data by calling an exporter widget
  1. Create "Loader" object

  2. Enter the file name of the scan to analyse ('FileName.h5')

  3. Options for x_stream quantities include:

  • All quantities contained in the sca folder(s) specified in the config
  • All SCA specified in the config
  1. Options for y_stream quantities include:
  • All quantities contained in the sca folder(s) specified in the config
  • All SCA specified in the config
  • All MCA specified in the config with applied ROI
  • All STACK specified in the config with two applied ROIs
  1. List all scans to analyse (comma-separated)

  2. Set optional flags. Options include:

  • norm (Normalizes to [0,1])
  • xcoffset (Defines a constant shift in the x-stream)
  • xoffset (Takes a list of tuples and defines a polynomial fit of the x-stream)
  • ycoffset (Defines a constant shift in the y-stream)
  • yoffset (Takes a list of tuples and defines a polynomial fit of the y-stream) e.g. offset = [(100,102),(110,112),(120,121)]
  • grid_x (Takes a list with three arguments to apply 1d interpolation gridding) e.g. grid_x = [Start Energy, Stop Energy, Delta]
  • savgol (Takes a list with two or three arguments to apply data smoothing and derivatives) e.g. savgol = [Window length, Polynomial order, deriavtive] as specified in the scipy Savitzky-Golay filter
  • binsize (int, allows to perform data binning to improve Signal-to-Noise)
  • legend_items (dict={scan_number:"name"}, overwrites generic legend names; works for the load method)
  • legend_item (str, overwrites generic legend name in the add/subtract method)

2d Images

Note: Can only load one scan at a time!

General loader for MCA detector data

load2d = Load2d()
load2d.load(config,'Filename.h5','x_stream','detector',1)
load2d.plot()
load2d.exporter()
  1. Create "Loader" object

  2. Enter the file name of the scan to analyse ('FileName.h5')

  3. Options for x_stream quantities include:

  • All quantities contained in the sca folder(s) specified in the config
  • All SCA specified in the config
  1. Options for detector quantities include:
  • All MCA specified in the config
  • All STACK specified in the config with applied ROI
  1. Select scan to analyse (comma-separated)

  2. Set optional flags. Options include:

  • norm (Normalizes to [0,1])
  • xcoffset (Defines a constant shift in the x-stream)
  • xoffset (Takes a list of tuples and defines a polynomial fit of the x-stream)
  • ycoffset (Defines a constant shift in the y-stream)
  • yoffset (Takes a list of tuples and defines a polynomial fit of the y-stream) e.g. offset = [(100,102),(110,112),(120,121)]
  • grid_x (Takes a list with three arguments to apply 1d interpolation gridding) e.g. grid_x = [Start Energy, Stop Energy, Delta]
  • norm_by (Normalizes to specified stream)

2d histogram

mesh = LoadMesh()
mesh.load(config,'Filename.h5','x_stream','y_stream','z_stream',24)
mesh.plot()
mesh.exporter()
  1. Create "Loader" object

  2. Enter the file name of the scan to analyse ('FileName.h5')

  3. Options for x_stream quantities include:

  • All quantities contained in the sca folder(s) specified in the config
  • All SCA specified in the config
  • All MCA specified in the config with ROI specified
  • All STACK specified in the config with two ROIs specified
  1. Options for y_stream quantities include:
  • All quantities contained in the sca folder(s) specified in the config
  • All SCA specified in the config
  • All MCA specified in the config with ROI specified
  • All STACK specified in the config with two ROIs specified
  1. Options for z_stream quantities include:
  • All quantities contained in the sca folder(s) specified in the config
  • All SCA specified in the config
  • All MCA specified in the config with ROI specified
  • All STACK specified in the config with two ROIs specified
  1. Specify scan to analyse

  2. Set optional flags. Options include:

  • norm (Normalizes to [0,1])
  • xcoffset (Defines a constant shift in the x-stream)
  • xoffset (Takes a list of tuples and defines a polynomial fit of the x-stream)
  • ycoffset (Defines a constant shift in the y-stream)
  • yoffset (Takes a list of tuples and defines a polynomial fit of the y-stream) e.g. offset = [(100,102),(110,112),(120,121)]

3d Images

Note: Can only load one scan at a time!

load3d = Load2d()
load3d.load(config,'Filename.h5','ind_stream','stack',1)
load3d.plot()
load3d.export()
  1. Create "Loader" object

  2. Enter the file name of the scan to analyse ('FileName.h5')

  3. Enter SCA name for independent stream, corresponding to length of first axis.

  4. Options for stack quantities include:

  • All STACK specified in the config
  1. Select scan to analyse

  2. Set optional flags. Options include:

  • norm (Normalizes to [0,1])
  • xcoffset (Defines a constant shift in the x-stream)
  • xoffset (Takes a list of tuples and defines a polynomial fit of the x-stream)
  • ycoffset (Defines a constant shift in the y-stream)
  • yoffset (Takes a list of tuples and defines a polynomial fit of the y-stream) e.g. offset = [(100,102),(110,112),(120,121)]
  • grid_x (Takes a list with three arguments to apply 1d interpolation gridding) e.g. grid_x = [Start Energy, Stop Energy, Delta]
  • norm_by (Normalizes to specified stream)

Meta Data

bl = LoadBeamline()
bl.load(config,'Filename.h5','path to variable')
bl.plot()
  1. Create "Loader" object

  2. Enter the file name of the scan to analyse ('FileName.h5')

  3. Options for path to variable quantities include:

  • All directory paths within the specified h5 file

Spreadsheet

df = getSpreadsheet(config,'Filename.h5', average = False,columns=None)
  1. Create "Loader" object

  2. Enter the file name of the scan to analyse ('FileName.h5')

3a. Specify average boolean

  • If False, return all data (even if 1d array)
  • If True, return average of 1d array where applicable

3b. Options for columns quantities include:

  • Custom dictionary with column headers and quantities, see example below:
columns = dict()

columns['Command'] = 'command'
columns['Sample Stage (ssh)'] = 'Endstation/Motors/ssh'
columns['Sample Stage (ssv)'] = 'Endstation/Motors/ssv'
columns['Sample Stage (ssd)'] = 'Endstation/Motors/ssd'
columns['Spectrometer (XES dist)'] = 'Endstation/Motors/spd'
columns['Spectrometer (XES angl)'] = 'Endstation/Motors/spa'
columns['Flux 4-Jaw (mm)'] = 'Beamline/Apertures/4-Jaw_2/horz_gap'
columns['Mono Grating'] = '/Beamline/Monochromator/grating'
columns['Mono Mirror'] = '/Beamline/Monochromator/mirror'
columns['Polarization'] = 'Beamline/Source/EPU/Polarization'
#columns['Comment'] = 'command'
columns['Status'] = 'status'

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

h5analysis-0.0.16.tar.gz (44.0 kB view details)

Uploaded Source

Built Distribution

h5analysis-0.0.16-py3-none-any.whl (50.7 kB view details)

Uploaded Python 3

File details

Details for the file h5analysis-0.0.16.tar.gz.

File metadata

  • Download URL: h5analysis-0.0.16.tar.gz
  • Upload date:
  • Size: 44.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/4.0.2 CPython/3.11.7

File hashes

Hashes for h5analysis-0.0.16.tar.gz
Algorithm Hash digest
SHA256 cd454294c943a7ab37bed0c8e9e245a800bce988408c09b86890ae18d64a9587
MD5 f655fab3ca10f53781e2ddb41454725d
BLAKE2b-256 445f8848ac73ad7fa5204b1102d94e9272fc9d9c9e24042b1d31c2a126ae0a0f

See more details on using hashes here.

File details

Details for the file h5analysis-0.0.16-py3-none-any.whl.

File metadata

  • Download URL: h5analysis-0.0.16-py3-none-any.whl
  • Upload date:
  • Size: 50.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/4.0.2 CPython/3.11.7

File hashes

Hashes for h5analysis-0.0.16-py3-none-any.whl
Algorithm Hash digest
SHA256 bef2de6ce5dab82a5487a61e08dae033702902174dde58269b2b6ae183189d16
MD5 91247d37a474c3f80728118973488689
BLAKE2b-256 cd464ae970d204f9591795a3a815d12da2d0ca14eebf6bc01c3d4adf7b43b2e2

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page