Skip to main content

Python package for BCDI data analysis

Project description

Welcome to the gwaihir package

The goal of the module is to provide a way to process and analyse BCDI data either via jupyter notebook or the command line.

It requires the python packages bcdi and pynx

The idea is to allow command line use of the package for fast analysis but also to use a GUI (for jupyter notebooks) to allow new users to understand each step during the process but also to allow experienced users to finelly tune their parameters and to offer a non "black-box" approach to the data analysis

GUI

The use of python objects for the GUI and for the datasets could allow one to easily share the data analysis between one another.

This means that:

  • A Dataset object can be reopened via the GUI,
  • The GUI presents several tabs, describing the data analysis (e.g. the parameters used during the analysis) so that one group could directly have a better understanding of the data history, these parameters would all be saved as attributes of the data object
  • Guidelines for the analysis are be provided in the GUI, in a README tab
  • These ideas apply the concept of thorondor, a similar package but for XANES data analysis (still in development)
  • All the parameters related to the different beamlines will be in the code. It will only require a single widget to change the beamline, while keeping the same output architecture and output format, making it much easier to share and compare the results. Most of the code will be common to all the beamlines. Python power!
  • Current hardcoded parameters (e.g. spec file location, ...) will not require the user to open scripts but just to edit a widget in notebook,
  • One GUI for all the data, one GUI to rule them all

Current workflow for BCDI data analysis

We can define separate steps in the processing and analysis of BCDI data:

  • preprocessing: masking, cropping, ... of the diffraction intensity (BCDI package)
  • phase retrieval: semi-automated, computes the amplitude and phase from the diffractogram intensity, depends on the input parameters, via jupyter notebook or terminal (PyNX package)
  • postprocessing:
    • solution selection: semi-automated, using different criteria such as the free Log-likelihood (PyNX package) or the standard deviation of the reconstructed object's amplitude, via jupyter notebook or terminal (@Clatlan)
    • decomposition into modes of the best solutions to create one final object, via jupyter notebook or terminal (PyNX package)
    • orthogonalisation of the data into the crystal frame for object comparison (BCDI package)
    • data correction: compute the q vector at the centre of mass from the detector angles (BCDI package)
    • computation of the displacement and strain of the object from its phase and q vector (BCDI package)
  • data analysis:
    • visualisation via paraview, depends on isosurface, usually selected as the lower foot of the curve in the amplitude histogram
    • facet retrieval, determine the size and normals to the facets on the isosurface from the normals to isosurface's points, depends on input parameters (FacetAnalyzer package in Paraview)
    • facet analysis, rotate the orientation matrix of the object so that the facet normals correspond to the normals to the crystallographic planes (e.g. 100 in crystal frame with y perpendicular to the interface becomes 111 ...) (@DupraxM and RichardMI)
    • plotting of mean strain and displacement values for each facet via a jupyter notebook

How to use the terminal scripts for bcdi analysis

There are a few scripts that are meant to directly retrieve the paraview file (.vti) from the 3D intensity collected with the detector. The steps described above are automatic until facet retrieval. If you want to use the scripts for BCDI at ID01 or SIXS, you need to:

  • Have cloned and installed the repository
  • Respect the following architecture for the folders, these folder you need to create by hand:
    • ./Temperature/Condition/ParticleName/ScanNumber/...
    • e.g.: ./350/Helium/III_B2/
  • Make sure that the spec files, template image file, data folders, rocking angle, etc are exact in the scripts.
  • You must launch all the scripts from the ./ folder, the following directories will be created.
  • In the ./ folder, launch quick_process_ID01.sh with the first argument the path up to the Sxxx folder and as second argument the scan number, e.g. quick_process_ID01.sh 350/Helium/III_B2/ 603

In the scan folder, different subfodlers with be automatically generated, keping the same example:

  • /data : contains the .nxs data (for SIXS, otherwise empty)
  • /pynxraw:
    • README_preprocess.md: Preprocessing output logs
    • S603_data_before_masking_stack.npz: 3D detector data before masking
    • S603_maskpynx_align-q-y_norm_200_360_324_1_1_1.npz: 3D detector mask
    • S603_pynx_align-q-y_norm_200_360_324_1_1_1.npz: 3D detector intensity after masking
    • S603_pynx_align-q-y_norm_200_360_324_1_1_1.cxi: hdf5 file containing metadata and detector data
    • finalmask_S603_align-q-y_norm_200_360_324_1_1_1.png: image, sum of mask in each direction
    • finalsum_S603_align-q-y_norm_200_360_324_1_1_1.png: image, sum of detector data in each direction
    • PhasingNotebook.ipynb: notebook for phase retrieval, output in /reconstruction folder
    • pynx_run.txt contains the input parameter for phase retrieval if performed via terminal
    • reconstructions/:
      • modes.h5: Output of decomposition into modes via jupyter notebook
      • *.cxi: single solutions from phase retrieval
    • all/:
      • modes_job.h5: Output of decomposition into modes via terminal
      • README_modes.md: Logs of modes decomposition via terminal
      • README_pynx.md: Logs of phase retrieval via terminal
      • *.cxi: single solutions from phase retrieval
  • /postprocessing:
    • CompareFacetsEvolution.ipynb: notebook for facet analysis
    • /corrections: contains the rocking curve, and the detector image at the maximum of the rocking curve
      • central_slice.png: Image, detector image at peak_pos
      • rocking_curve.png: Image, rocking curve and its interpolation
      • README_correct_angles.md: Logs of data correction
      • correct_detector_data.npz: Contains interpolated rocking curve, FWHH, peak_position, detector bragg angles, ...
  • /result_crystal_frame:
    • README_strain.md: Logs of strain computation
    • S603_ampdispstrain_mode_avg3_apodize_blackman_crystalframe.npz: contains the orthogonalised data with amplitude, phase, displacement and strain
    • S603_amp-disp-strain_mode_avg3_apodize_blackman_crystalframe.vti: contains the orthogonalised data with amplitude, phase, displacement and strain, to be viewed via paraview

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gwaihir-0.0.1.tar.gz (205.3 kB view details)

Uploaded Source

File details

Details for the file gwaihir-0.0.1.tar.gz.

File metadata

  • Download URL: gwaihir-0.0.1.tar.gz
  • Upload date:
  • Size: 205.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.6.4 pkginfo/1.7.1 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.62.2 CPython/3.9.7

File hashes

Hashes for gwaihir-0.0.1.tar.gz
Algorithm Hash digest
SHA256 29bccf6dcded1993d6eb38567fefa9302ee11dd661fa69c75282eaadf3aad53d
MD5 19194300a60dd64cc1a70b9bdb7030dd
BLAKE2b-256 cd28321d76bc9c3848ee5120ed67fc39d557eeb85bc9b84cde69c78317dc78f8

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page