Skip to main content

A framework for machine learning on Apple silicon.

Project description

MLX

Quickstart | Installation | Documentation | Examples

CircleCI

MLX is an array framework for machine learning on Apple silicon, brought to you by Apple machine learning research.

Some key features of MLX include:

  • Familiar APIs: MLX has a Python API that closely follows NumPy. MLX also has fully featured C++, C, and Swift APIs, which closely mirror the Python API. MLX has higher-level packages like mlx.nn and mlx.optimizers with APIs that closely follow PyTorch to simplify building more complex models.

  • Composable function transformations: MLX supports composable function transformations for automatic differentiation, automatic vectorization, and computation graph optimization.

  • Lazy computation: Computations in MLX are lazy. Arrays are only materialized when needed.

  • Dynamic graph construction: Computation graphs in MLX are constructed dynamically. Changing the shapes of function arguments does not trigger slow compilations, and debugging is simple and intuitive.

  • Multi-device: Operations can run on any of the supported devices (currently the CPU and the GPU).

  • Unified memory: A notable difference from MLX and other frameworks is the unified memory model. Arrays in MLX live in shared memory. Operations on MLX arrays can be performed on any of the supported device types without transferring data.

MLX is designed by machine learning researchers for machine learning researchers. The framework is intended to be user-friendly, but still efficient to train and deploy models. The design of the framework itself is also conceptually simple. We intend to make it easy for researchers to extend and improve MLX with the goal of quickly exploring new ideas.

The design of MLX is inspired by frameworks like NumPy, PyTorch, Jax, and ArrayFire.

Examples

The MLX examples repo has a variety of examples, including:

Quickstart

See the quick start guide in the documentation.

Installation

MLX is available on PyPI. To install MLX on macOS, run:

pip install mlx

To install the CUDA backend on Linux, run:

pip install mlx[cuda]

To install a CPU-only Linux package, run:

pip install mlx[cpu]

Checkout the documentation for more information on building the C++ and Python APIs from source.

Contributing

Check out the contribution guidelines for more information on contributing to MLX. See the docs for more information on building from source, and running tests.

We are grateful for all of our contributors. If you contribute to MLX and wish to be acknowledged, please add your name to the list in your pull request.

Citing MLX

The MLX software suite was initially developed with equal contribution by Awni Hannun, Jagrit Digani, Angelos Katharopoulos, and Ronan Collobert. If you find MLX useful in your research and wish to cite it, please use the following BibTex entry:

@software{mlx2023,
  author = {Awni Hannun and Jagrit Digani and Angelos Katharopoulos and Ronan Collobert},
  title = {{MLX}: Efficient and flexible machine learning on Apple silicon},
  url = {https://github.com/ml-explore},
  version = {0.0},
  year = {2023},
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

mlx_cuda_12-0.30.6-py3-none-manylinux_2_35_x86_64.whl (69.2 MB view details)

Uploaded Python 3manylinux: glibc 2.35+ x86-64

mlx_cuda_12-0.30.6-py3-none-manylinux_2_35_aarch64.whl (66.9 MB view details)

Uploaded Python 3manylinux: glibc 2.35+ ARM64

File details

Details for the file mlx_cuda_12-0.30.6-py3-none-manylinux_2_35_x86_64.whl.

File metadata

File hashes

Hashes for mlx_cuda_12-0.30.6-py3-none-manylinux_2_35_x86_64.whl
Algorithm Hash digest
SHA256 3970b7b4e9ba423bba90c44ceafb43584485c654aff04aee85f4f13c1d56baa4
MD5 565b00b6bf3cf4faba6fb1d0a9f92c10
BLAKE2b-256 d74bce4d25f882c32064ccea6bd876249a19a659b3592bca4afa5da245093b8e

See more details on using hashes here.

Provenance

The following attestation bundles were made for mlx_cuda_12-0.30.6-py3-none-manylinux_2_35_x86_64.whl:

Publisher: release.yml on ml-explore/mlx

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file mlx_cuda_12-0.30.6-py3-none-manylinux_2_35_aarch64.whl.

File metadata

File hashes

Hashes for mlx_cuda_12-0.30.6-py3-none-manylinux_2_35_aarch64.whl
Algorithm Hash digest
SHA256 9a9a9666287f93a84a7b33e6200b6b441c84598c18b9f8f4ff1e1d482af83e39
MD5 54337dc3ddcbd2edc4b252099005d9a4
BLAKE2b-256 00d2855d6d7a20ec12cf25559bdb53d9dd1dd3811548905e804a36f49892adfc

See more details on using hashes here.

Provenance

The following attestation bundles were made for mlx_cuda_12-0.30.6-py3-none-manylinux_2_35_aarch64.whl:

Publisher: release.yml on ml-explore/mlx

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page