A Practical Method for Constructing Equivariant Multilayer Perceptrons for Arbitrary Matrix Groups
Project description
A Practical Method for Constructing Equivariant Multilayer Perceptrons for Arbitrary Matrix Groups
| Paper | |
EMLP is a jax library for the automated construction of equivariant layers in deep learning. You can read the documentation here.
WARNING: Our library (and paper) have not yet been released, and may have sharp edges, bugs, and may be subject to breaking changes. Use at your own caution. But if you notice things behaving unexpectedly or get frustrated, send me an email so I can make the library better.
We provide a type system for representations. With the operators ρᵤ⊗ρᵥ, ρᵤ⊕ρᵥ, ρ* implemented as *
,+
and .T
build up different representations. The basic building blocks for representations are the base vector representation V
and tensor representations T(p,q) = V**p*V.T**q
.
For any given matrix group and representation formed in our type system, you can get the equivariant basis with rep.equivariant_basis()
or a matrix which projects to that subspace with rep.equivariant_projector()
.
For example to find all O(1,3) (Lorentz) equivariant linear maps from from a 4-Vector Xᶜ to a rank (2,1) tensor Mᵇᵈₐ, you can run
from emlp.reps import V,T
from emlp.groups import *
G = O13()
Q = (T(1,0)>>T(1,2))(G).equivariant_basis()
or how about equivariant maps from one Rubik's cube to another?
G = RubiksCube()
Q = (V(G)>>V(G)).equivariant_basis()
Using +
and *
you can put together composite representations (where multiple representations are concatenated together). For example lets find all equivariant linear maps from 5 node features and 2 edge features to 3 global invariants and 1 edge feature of a graph of size n=5:
G=S(5)
repin = 10*T(1)+5*T(2)
repout = 3*T(0)+T(2)
Q = (repin(G)>>repout(G)).equivariant_basis()
From the examples above, there are many different ways of writing a representation like 10*T(1)+5*T(2)
which are all equivalent.
10*T(1)+5*T(2)
= 10*V+5*V**2
= 5*V*(2+V)
You can even mix and match representations from different groups. For example with the cyclic group ℤ₃, the permutation group 𝕊₄, and the orthogonal group O(3)
rep = 2*V(Z(3))*V(S(4))+V(O(3))**2
Q = (rep>>rep).equivariant_basis()
Outside of these tensor representations, our type system works with any finite dimensional linear representation and you can even build your own bespoke representations following the instructions here.
You can visualize these equivariant bases with vis(repin,repout)
, such as with the three examples above
Checkout our documentation to see how to use our system and some worked examples.
Installation instructions
To install as a package, run
pip install emlp
To run the scripts you will instead need to clone the repo and install it locally which you can do with
git clone https://github.com/mfinzi/equivariant-MLP.git
cd equivariant-MLP
pip install -e .
Experimental Results from Paper
Assuming you have installed the repo locally, you can run the experiments we described in the paper.
To train the regression models on one of the Inertia
, O5Synthetic
, or ParticleInteraction
datasets found in emlp.datasets.py
you can run the script experiments/train_regression.py
with command line arguments specifying the dataset, network, and symmetry group. For example to train EMLP
with SO(3)
equivariance on the Inertia
dataset, you can run
python experiments/train_regression.py --dataset Inertia --network EMLP --group "SO(3)"
or to train the MLP baseline you can run
python experiments/train_regression.py --dataset Inertia --network MLP
Other command line arguments such as --aug=True
for data augmentation or --ch=512
for number of hidden units and others are available, and you can browse the options and their defaults with python experiments/train_regression.py -h
. If no group is specified, EMLP will automatically choose the one matched to the dataset, but you can also go crazy with any of the other groups implemented in groups.py
provided the dimensions match the data (e.g. for the 3D inertia dataset you could do --group=
"Z(3)"
or "DkeR3(3)"
but not "Sp(2)"
or "SU(5)"
).
For the dynamical systems modeling experiments you can use the scripts
experiments/neuralode.py
to train (equivariant) Neural ODEs and experiments/hnn.py
to train (equivariant) Hamiltonian Neural Networks.
For the dynamical system task, the Neural ODE and HNN models have special names. EMLPode
and MLPode
for the Neural ODEs in neuralode.py
and EMLPH
and MLPH
for the HNNs in hnn.py
. For example,
python experiments/neuralode.py --network EMLPode --group="O2eR3()"
or
python experiments/hnn.py --network EMLPH --group="DkeR3(6)"
These models are trained to fit a double spring dynamical system. 30s rollouts of the dataset, along with rollout error on these trajectories, and conservation of angular momentum are shown below.
If you find our work helpful, please cite it with
@article{finzi2021emlp,
title={A Practical Method for Constructing Equivariant Multilayer Perceptrons for Arbitrary Matrix Groups},
author={Finzi, Marc and Welling, Max and Wilson, Andrew Gordon},
journal={Arxiv},
year={2021}
}
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file emlp-0.9.1.tar.gz
.
File metadata
- Download URL: emlp-0.9.1.tar.gz
- Upload date:
- Size: 42.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.4.1 importlib_metadata/3.10.0 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.55.1 CPython/3.8.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | fc927b3e5acba0e0fb475ca83031a3e0c63163e39db997c264813f6682e2e161 |
|
MD5 | e4f98fe81fd38ccc4e730fbf03f6d6bd |
|
BLAKE2b-256 | 37e36e0b7c1c7723e3ff8172967b7de64f5568b8827ddee8ea5ceac5c3803d6b |
File details
Details for the file emlp-0.9.1-py3-none-any.whl
.
File metadata
- Download URL: emlp-0.9.1-py3-none-any.whl
- Upload date:
- Size: 49.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.4.1 importlib_metadata/3.10.0 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.55.1 CPython/3.8.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | c0066b88780dc88b98ed3f47025e0302f91e3d0b5325c927a4088e6c334046d1 |
|
MD5 | 128e0328be42aa6e553c8d82f1c32e2d |
|
BLAKE2b-256 | 50c189e3f23ff689ea4c90958649e69b32fc2c9f615c74bd42f9f1d352990be6 |