Hardware accelerated, batchable and differentiable optimizers in JAX.
Project description
JAXopt
Status | Installation | Documentation | Examples | Cite us
Hardware accelerated, batchable and differentiable optimizers in JAX.
- Hardware accelerated: our implementations run on GPU and TPU, in addition to CPU.
- Batchable: multiple instances of the same optimization problem can be automatically vectorized using JAX's vmap.
- Differentiable: optimization problem solutions can be differentiated with respect to their inputs either implicitly or via autodiff of unrolled algorithm iterations.
Status
JAXopt is no longer maintained nor developed. Alternatives may be found on the JAX website. Some of its features (like losses, projections, lbfgs optimizer) have been ported into optax. We are sincerely grateful for all the community contributions the project has garnered over the years.
Installation
To install the latest release of JAXopt, use the following command:
$ pip install jaxopt
To install the development version, use the following command instead:
$ pip install git+https://github.com/google/jaxopt
Alternatively, it can be installed from sources with the following command:
$ python setup.py install
Cite us
Our implicit differentiation framework is described in this paper. To cite it:
@article{jaxopt_implicit_diff,
title={Efficient and Modular Implicit Differentiation},
author={Blondel, Mathieu and Berthet, Quentin and Cuturi, Marco and Frostig, Roy
and Hoyer, Stephan and Llinares-L{\'o}pez, Felipe and Pedregosa, Fabian
and Vert, Jean-Philippe},
journal={arXiv preprint arXiv:2105.15183},
year={2021}
}
Disclaimer
JAXopt was an open source project maintained by a dedicated team in Google Research. It is not an official Google product.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file jaxopt-0.8.5.tar.gz.
File metadata
- Download URL: jaxopt-0.8.5.tar.gz
- Upload date:
- Size: 121.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
2790bd68ef132b216c083a8bc7a2704eceb35a92c0fc0a1e652e79dfb1e9e9ab
|
|
| MD5 |
50bc6406051fa4ba977e8675b0c355e2
|
|
| BLAKE2b-256 |
3adaff7d7fbd13b8ed5e8458e80308d075fc649062b9f8676d3fc56f2dc99a82
|
File details
Details for the file jaxopt-0.8.5-py3-none-any.whl.
File metadata
- Download URL: jaxopt-0.8.5-py3-none-any.whl
- Upload date:
- Size: 172.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.13.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ff221d1a86908ec759eb1e219ee1d12bf208a70707e961bf7401076fe7cf4d5e
|
|
| MD5 |
92b157b105615f13e731e144380c24b8
|
|
| BLAKE2b-256 |
45d855e0901103c93d57bab3b932294c216f0cbd49054187ce29f8f13808d530
|