Multiple dispatch in JAX via custom interpreters.
Project description
Quax
Uses JAX's nonstandard interpretation to perform multiple dispatch on custom array-ish objects, like:
- LoRA weight matrices
- symbolic zeros
- arrays with named dimensions
- structured (e.g. tridiagonal) matrices
- sparse arrays
- etc!
This works via a custom JAX transform. This means that it works even with existing programs, that were not written to accept such array-ish objects: just wrap the program in the quaxify
transform. A typical use-case is applying LoRA to fine-tune arbitrary models.
Implementations for LoRA and symbolic zeros are both already built-in to Quax. You can also create your own types and rules very easily; see the examples library for demo implementations for named arrays, sparse arrays, tridiagonal matrices, and PRNG keys.
Installation
pip install quax
Example: LoRA
import equinox as eqx
import jax.random as jr
import quax
# Start off with any JAX program: here, the forward pass through a linear layer.
key1, key2, key3 = jr.split(jr.PRNGKey(0), 3)
linear = eqx.nn.Linear(10, 12, key=key1)
vector = jr.normal(key2, (10,))
# Make some of the inputs be an array-ish object. This function finds all
# `eqx.nn.Linear` layers, and wraps their weights in `LoraArray`s.
lora_linear = quax.lora.loraify(linear, rank=2, key=key3)
# For this simple model, we could also do it manually.
lora_weight = quax.lora.LoraArray(linear.weight, rank=2, key=key3)
lora_linear = eqx.tree_at(lambda l: l.weight, linear, lora_weight)
# Wrap your function call in quaxify. This transform calls your original function,
# whilst looking up any multiple dispatch rules registered for any custom array-ish
# objects.
out = quax.quaxify(lora_linear)(vector)
Work in progress!
This library is a work in progress! Right now it should support enough to run LoRA on common models. However, some operations (e.g. jax.lax.cond_p
) are not yet supported. If you attempt to use these then an error will be thrown whilst tracing your program.
If you find yourself hitting any of these, then go ahead and open an issue, and/or a pull request!
See also: other libraries in the JAX ecosystem
Equinox: neural networks.
jaxtyping: type annotations for shape/dtype of arrays.
Optax: first-order gradient (SGD, Adam, ...) optimisers.
Diffrax: numerical differential equation solvers.
Optimistix: root finding, minimisation, fixed points, and least squares.
Lineax: linear solvers.
BlackJAX: probabilistic+Bayesian sampling.
Orbax: checkpointing (async/multi-host/multi-device).
sympy2jax: SymPy<->JAX conversion; train symbolic expressions via gradient descent.
Eqxvision: computer vision models.
Levanter: scalable+reliable training of foundation models (e.g. LLMs).
PySR: symbolic regression. (Non-JAX honourable mention!)
Acknowledgements
Significantly inspired by https://github.com/davisyoshida/qax, https://github.com/stanford-crfm/levanter, and jax.experimental.sparse
.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file quax-0.0.2.tar.gz
.
File metadata
- Download URL: quax-0.0.2.tar.gz
- Upload date:
- Size: 16.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | b08f51444cb573f0945ef21d270237da32d3c230e34ff3d2c99a14af50864c95 |
|
MD5 | 58733ed7f90bba61d3aef6a2be53add6 |
|
BLAKE2b-256 | 34b0e0a7bad7951f7839a62e1abf9982d7437beca69396e5a3f84c9c3dbb57bf |
File details
Details for the file quax-0.0.2-py3-none-any.whl
.
File metadata
- Download URL: quax-0.0.2-py3-none-any.whl
- Upload date:
- Size: 21.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 3add670a3804bb1f3df73047bfb13af67fcc4fea8889e460a33b0c4b788c4866 |
|
MD5 | 6b24828384560ac393c98f10c026c7ce |
|
BLAKE2b-256 | 86537e630ffdfeef5ad6acd1d1ce6b1c243ec79ee20c03b2d600d45e6bd8e48e |