Examples in the MLX framework
mlx
Go to file
Anchen 362e88a744
feat: move lora into mlx-lm (#337)
* feat: Add lora and qlora training to mlx-lm


---------

Co-authored-by: Awni Hannun <awni@apple.com>
2024-01-23 08:44:37 -08:00
bert Change tuple type definitions to use Tuple (#308) 2024-01-12 11:15:09 -08:00
cifar Updated CIFAR-10 ResNet example to use BatchNorm instead of LayerNorm (#257) 2024-01-12 05:43:11 -08:00
gcn Support Hugging Face models (#215) 2024-01-03 15:13:26 -08:00
llms feat: move lora into mlx-lm (#337) 2024-01-23 08:44:37 -08:00
lora feat(lora): add de-quantized support for fuse.py (#351) 2024-01-22 17:32:24 -08:00
mnist Refactor activation function and loss calculation (#325) 2024-01-16 13:42:56 -08:00
normalizing_flow Fix import order of normalizing_flow (#326) 2024-01-16 08:45:55 -08:00
speechcommands Use pip for mlx data with speech commands (#307) 2024-01-12 11:06:33 -08:00
stable_diffusion two minor fixes (#335) 2024-01-18 14:18:13 -08:00
t5 add speculative decoding example for llama (#149) 2023-12-28 15:20:43 -08:00
transformer_lm Add llms subdir + update README (#145) 2023-12-20 10:22:25 -08:00
whisper Added lora support for Phi-2 (#302) 2024-01-12 13:45:30 -08:00
.gitignore Align CLI args and some smaller fixes (#167) 2023-12-22 14:34:32 -08:00
.pre-commit-config.yaml Feat: Bump isort version (#350) 2024-01-21 06:35:15 -08:00
ACKNOWLEDGMENTS.md Add PLaMo-13B model as an LLM example (#303) 2024-01-23 07:17:24 -08:00
CODE_OF_CONDUCT.md contribution + code of conduct 2023-11-29 12:31:18 -08:00
CONTRIBUTING.md Update CONTRIBUTING.md 2023-12-09 08:02:34 +09:00
LICENSE consistent copyright 2023-11-30 11:11:04 -08:00
README.md Update README.md (#251) 2024-01-07 11:35:39 -08:00

MLX Examples

This repo contains a variety of standalone examples using the MLX framework.

The MNIST example is a good starting point to learn how to use MLX.

Some more useful examples are listed below.

Text Models

Image Models

Audio Models

Other Models

  • Semi-supervised learning on graph-structured data with GCN.

Hugging Face

Note: You can now directly download a few converted checkpoints from the MLX Community organization on Hugging Face. We encourage you to join the community and contribute new models.

Contributing

We are grateful for all of our contributors. If you contribute to MLX Examples and wish to be acknowledged, please add your name to the list in your pull request.

Citing MLX Examples

The MLX software suite was initially developed with equal contribution by Awni Hannun, Jagrit Digani, Angelos Katharopoulos, and Ronan Collobert. If you find MLX Examples useful in your research and wish to cite it, please use the following BibTex entry:

@software{mlx2023,
  author = {Awni Hannun and Jagrit Digani and Angelos Katharopoulos and Ronan Collobert},
  title = {{MLX}: Efficient and flexible machine learning on Apple silicon},
  url = {https://github.com/ml-explore},
  version = {0.0},
  year = {2023},
}