Transformer-based models to fast-simulate the LHCb ECAL detector
Project description
Transformer-based models to fast-simulate the LHCb ECAL detector
Transformer
The Transformer architecture is freely inspired by Vaswani et al. [arXiv:1706.03762] and Dosovitskiy et al. [arXiv:2010.11929].
Discriminator
The Discriminator is implemented through the Deep Sets model proposed by Zaheer et al. [arXiv:1703.06114] and its architecture is freely inspired by what developed by the ATLAS Collaboration for flavor tagging [ATL-PHYS-PUB-2020-014].
Credits
Transformer implementation freely inspired by the TensorFlow tutorial Neural machine translation with a Transformer and Keras.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
calotron-0.0.12.tar.gz
(41.2 kB
view hashes)
Built Distribution
calotron-0.0.12-py3-none-any.whl
(68.2 kB
view hashes)
Close
Hashes for calotron-0.0.12-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 39ce91a354b9fbf8042f6b06cbff25c86644fa735d0eb37d045ff6baaa839e56 |
|
MD5 | d6f59acc53e2122d543662117d7a1b58 |
|
BLAKE2b-256 | 8747163ab4b4f7699f1e6a31c7b919f3a2a5018cbbd592bba9787b2727e6857b |