A framework for structural shape optimization based on automatic differentiation (AD) and the adjoint method, enabled by JAX
Project description
JaxSSO
A differentiable finite element analysis (FEA) solver for structural optimization, enabled by JAX.
Developed by Gaoyuan Wu @ Princeton.
Features
- Automatic differentiation (AD): an easy and accurate way for gradient evaluation. The implementation of AD avoids deriving derivatives manually or trauncation errors from numerical differentiation.
- Acclerated linear algebra (XLA) and just-in-time compilation: these features in JAX boost the gradient evaluation
- Hardware acceleration: run on GPUs and TPUs for faster experience
- Support beam-column elements and MITC-4 quadrilateral shell elements
- Shape optimization, size optimization and topology optimization
- Seamless integration with machine learning (ML) libraries
Usage
Installation
Install it with pip: pip install JaxSSO
Dependencies
JaxSSO is written in Python and requires:
- numpy >= 1.22.0.
- JAX: "JAX is Autograd and XLA, brought together for high-performance machine learning research." Please refer to this link for the installation of JAX.
- scipy.
Optional:
- Nlopt: Nlopt is a library for nonlinear optimization. It has Python interface, which is implemented herein. Refer to this link for the installation of Nlopt. Alternatively, you can use
pip install nlopt, please refer to nlopt-python. - Flax: neural network library based on JAX. JAXSSO can be integrated with
flax, please seeExamples/Neural_Network_Topo_Shape.ipynb - Optax: optimization library based on JAX, can be used to train neural networks.
Quickstart
The project provides you with interactive examples with Google Colab for quick start. No installation locally is required.
- Shape optimization of grid shell: geometry from Favilli et al. 2024
Cite us
Please star, share our project with others and/or cite us if you find our work interesting and helpful.
We have a new manuscript under review.
Our previous work can be seen in this paper. Cite our previous work using:
@article{wu2023framework,
title={A framework for structural shape optimization based on automatic differentiation, the adjoint method and accelerated linear algebra},
author={Wu, Gaoyuan},
journal={Structural and Multidisciplinary Optimization},
volume={66},
url = {https://doi.org/10.1007/s00158-023-03601-0},
doi = {10.1007/s00158-023-03601-0},
pages={151},
year={2023},
publisher={Springer}
}
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file jaxsso-1.0.1.tar.gz.
File metadata
- Download URL: jaxsso-1.0.1.tar.gz
- Upload date:
- Size: 27.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.6
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
856835e9b886790c891854b77390e662b0fd95cc21525f64320230c3c923a67d
|
|
| MD5 |
81b043b843144b047b8ad2431037d616
|
|
| BLAKE2b-256 |
9977f36a3e898efd0538573e9cd554d2f8e519fe1897f801283367edabcfe1d2
|
File details
Details for the file jaxsso-1.0.1-py3-none-any.whl.
File metadata
- Download URL: jaxsso-1.0.1-py3-none-any.whl
- Upload date:
- Size: 29.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.6
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
09a779e8830659f5bb83efee602e7c83d52d938b4d2eee1e64cb0098cad1453b
|
|
| MD5 |
44f1e601e2d82434dd1e33144eb6ade3
|
|
| BLAKE2b-256 |
dd920edfbedbb6064ce2f239256c0cf487d8f677ff5955ef2b006128a11908bc
|