Fast Learning of Atomistic Rare Events
Project description
NOTE: This is the latest release 1.3.3 which includes significant changes compared to the previous version 0.2.4. Please check the updated tutorials and documentations from the links below.
FLARE: Fast Learning of Atomistic Rare Events
FLARE is an open-source Python package for creating fast and accurate interatomic potentials.
Major Features
Note:
We implement Sparse GP, all the kernels and descriptors in C++ with Python interface.
We implement Full GP, Mapped GP, RBCM, Squared Exponential kernel and 2+3-body descriptors in Python.
Please do NOT mix them.
Documentations and Tutorials
Documentation of the code can be accessed here: https://mir-group.github.io/flare
Applications using FLARE and gallery
Google Colab Tutorials
FLARE (ACE descriptors + sparse GP). The tutorial shows how to run flare with ACE and SGP on energy and force data, demoing "offline" training on the MD17 dataset and "online" on-the-fly training of a simple aluminum force field. All the trainings use yaml files for configuration.
FLARE (LAMMPS active learning)
This tutorial demonstrates new functionality for running active learning all within LAMMPS, with LAMMPS running the dynamics to allow arbitrarily complex molecular dynamics workflows while maintaining a simple interface. This also demonstrates how to use the C++ API directly from Python through pybind11
. Finally, there's a simple demonstration of phonon calculations with FLARE using phonopy
.
FLARE (ACE descriptors + sparse GP) with LAMMPS. The tutorial shows how to compile LAMMPS with FLARE pair style and uncertainty compute code, and use LAMMPS for Bayesian active learning and uncertainty-aware molecular dynamics.
Compute thermal conductivity from FLARE and Boltzmann transport equations. The tutorial shows how to use FLARE (LAMMPS) potential to compute lattice thermal conductivity from Boltzmann transport equation method, with Phono3py for force constants calculations and Phoebe for thermal conductivities.
Using your own customized descriptors with FLARE. The tutorial shows how to attach your own descriptors with FLARE sparse GP model and do training and testing.
All the tutorials take a few minutes to run on a normal desktop computer or laptop (excluding installation time).
Installation
Pip installation
Please check the installation guide here. This will take a few minutes on a normal desktop computer or laptop.
Developer's installation guide
For developers, please check the installation guide.
Compiling LAMMPS
See documentation on compiling LAMMPS with FLARE
Trouble shooting
If you have problem compiling and installing the code, please check the FAQs to see if your problem is covered. Otherwise, please open an issue or contact us.
System requirements
Software dependencies
- GCC 9
- Python 3
- pip>=20
MKL is recommended but not required. All other software dependencies are taken care of by pip.
The code is built and tested with Github Actions using the GCC 9 compiler. (You can find a summary of recent builds here.) Other C++ compilers may work, but we can't guarantee this.
Operating systems
flare++ is tested on a Linux operating system (Ubuntu 20.04.3), but should also be compatible with Mac and Windows operating systems. If you run into issues running the code on Mac or Windows, please post to the issue board.
Hardware requirements
There are no non-standard hardware requirements to download the software and train simple models—the introductory tutorial can be run on a single cpu. To train large models (10k+ sparse environments), we recommend using a compute node with at least 100GB of RAM.
Tests
We recommend running unit tests to confirm that FLARE is running properly on your machine. We have implemented our tests using the pytest suite. You can call pytest
from the command line in the tests directory.
Instructions (either DFT package will suffice):
pip install pytest
cd tests
pytest
References
If you use FLARE++ including B2 descriptors, NormalizedDotProduct kernel and Sparse GP, please cite the following paper:
[1] Vandermause, J., Xie, Y., Lim, J.S., Owen, C.J. and Kozinsky, B., 2021. Active learning of reactive Bayesian force fields: Application to heterogeneous hydrogen-platinum catalysis dynamics. Nature Communications 13.1 (2022): 5183. https://www.nature.com/articles/s41467-022-32294-0
If you use FLARE active learning workflow, full Gaussian process or 2-body/3-body kernel in your research, please cite the following paper:
[2] Vandermause, J., Torrisi, S. B., Batzner, S., Xie, Y., Sun, L., Kolpak, A. M. & Kozinsky, B. On-the-fly active learning of interpretable Bayesian force fields for atomistic rare events. npj Comput Mater 6, 20 (2020). https://doi.org/10.1038/s41524-020-0283-z
If you use FLARE LAMMPS pair style or MGP (mapped Gaussian process), please cite the following paper:
[3] Xie, Y., Vandermause, J., Sun, L. et al. Bayesian force fields from active learning for simulation of inter-dimensional transformation of stanene. npj Comput Mater 7, 40 (2021). https://doi.org/10.1038/s41524-021-00510-y
If you use FLARE PyLAMMPS for training, please cite the following paper:
[4] Xie, Y., Vandermause, J., Ramakers, S., Protik, N.H., Johansson, A. and Kozinsky, B., 2022. Uncertainty-aware molecular dynamics from Bayesian active learning: Phase Transformations and Thermal Transport in SiC. npj Comput. Mater. 9(1), 36 (2023).
If you use FLARE LAMMPS Kokkos pair style with GPU acceleration, please cite the following paper:
[5] Johansson, A., Xie, Y., Owen, C.J., Soo, J., Sun, L., Vandermause, J. and Kozinsky, B., 2022. Micron-scale heterogeneous catalysis with Bayesian force fields from first principles and active learning. arXiv preprint arXiv:2204.12573.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file mir_flare-1.4.0.tar.gz
.
File metadata
- Download URL: mir_flare-1.4.0.tar.gz
- Upload date:
- Size: 1.4 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.0 CPython/3.10.14
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 99d3710110f2647e697896df32f604537294c9ff20a84bad6db2bc8bc02b23f7 |
|
MD5 | ec15a62a5727bb6971b59c1f1aaf1ef3 |
|
BLAKE2b-256 | 1fb4b6665ae4d47819b3fff56611321ef2800dff0bf6a88aabfe6f86e9537105 |