MLX: An array framework for Apple silicon
mlx
Go to file
Cyril Zakka, MD e080290ba4
Added eye/identity ops (#119)
`eye` and `identity` C++ and Python ops
2023-12-11 12:38:17 -08:00
.circleci Nits (#38) 2023-12-06 13:32:41 -08:00
.github Pr template (#99) 2023-12-09 09:36:56 -08:00
benchmarks Added activation functions: leaky_relu relu6 softplus elu celu logsigmoid (#108) 2023-12-10 16:31:38 -08:00
cmake jagrit's commit files 2023-11-29 10:52:08 -08:00
docs Added eye/identity ops (#119) 2023-12-11 12:38:17 -08:00
examples Add isort pre-commit and run (#68) 2023-12-08 11:31:47 -08:00
mlx Added eye/identity ops (#119) 2023-12-11 12:38:17 -08:00
python Added eye/identity ops (#119) 2023-12-11 12:38:17 -08:00
tests Added eye/identity ops (#119) 2023-12-11 12:38:17 -08:00
.clang-format awni's commit files 2023-11-29 10:30:41 -08:00
.gitignore add nn module for sigmoid activation (#111) 2023-12-10 07:00:39 -08:00
.pre-commit-config.yaml Use compiled black as the pre-commit formatter (#94) 2023-12-09 07:06:46 -08:00
ACKNOWLEDGMENTS.md Add metal cpp license to acknowledgements (#3) 2023-11-30 11:33:23 -08:00
CMakeLists.txt Update cmake to detect and throw warnings if not on a arm system (#81) 2023-12-08 11:03:25 -08:00
CODE_OF_CONDUCT.md contribution and code of conduct (#1) 2023-11-29 12:54:28 -08:00
CONTRIBUTING.md Readme (#2) 2023-11-29 16:23:42 -08:00
LICENSE copyright + ack 2023-11-30 11:12:53 -08:00
MANIFEST.in awni's commit files 2023-11-29 10:30:41 -08:00
mlx.pc.in awni's commit files 2023-11-29 10:30:41 -08:00
pyproject.toml jagrit's commit files 2023-11-29 10:52:08 -08:00
README.md Update README.md (#121) 2023-12-10 14:47:37 -08:00
setup.py Add isort pre-commit and run (#68) 2023-12-08 11:31:47 -08:00

MLX

Quickstart | Installation | Documentation | Examples

CircleCI

MLX is an array framework for machine learning on Apple silicon, brought to you by Apple machine learning research.

Some key features of MLX include:

  • Familiar APIs: MLX has a Python API that closely follows NumPy. MLX also has a fully featured C++ API, which closely mirrors the Python API. MLX has higher-level packages like mlx.nn and mlx.optimizers with APIs that closely follow PyTorch to simplify building more complex models.

  • Composable function transformations: MLX supports composable function transformations for automatic differentiation, automatic vectorization, and computation graph optimization.

  • Lazy computation: Computations in MLX are lazy. Arrays are only materialized when needed.

  • Dynamic graph construction: Computation graphs in MLX are constructed dynamically. Changing the shapes of function arguments does not trigger slow compilations, and debugging is simple and intuitive.

  • Multi-device: Operations can run on any of the supported devices (currently the CPU and the GPU).

  • Unified memory: A notable difference from MLX and other frameworks is the unified memory model. Arrays in MLX live in shared memory. Operations on MLX arrays can be performed on any of the supported device types without transferring data.

MLX is designed by machine learning researchers for machine learning researchers. The framework is intended to be user-friendly, but still efficient to train and deploy models. The design of the framework itself is also conceptually simple. We intend to make it easy for researchers to extend and improve MLX with the goal of quickly exploring new ideas.

The design of MLX is inspired by frameworks like NumPy, PyTorch, Jax, and ArrayFire.

Examples

The MLX examples repo has a variety of examples, including:

Quickstart

See the quick start guide in the documentation.

Installation

MLX is available on PyPI. To install the Python API, run:

pip install mlx

Checkout the documentation for more information on building the C++ and Python APIs from source.

Contributing

Check out the contribution guidelines for more information on contributing to MLX.