mlx-examples/llms/mlx_lm
Ivan Fioravanti 4576946151
Add checkpoints directory for adapter weights (#431)
* Add checkpoints directory for adapter weights

The code was modified to create a checkpoints directory if it doesn't exist yet. Adapter weights are now saved to this checkpoints directory during the training iterations.
Corrected indentation of Save adapter weights code because it was part of "if eval"

* Fixing a blank added by mistake
2024-02-12 10:50:05 -08:00
..
models fix(mlx-m): lazy load hf_olmo (#424) 2024-02-08 09:02:43 -08:00
tuner Add checkpoints directory for adapter weights (#431) 2024-02-12 10:50:05 -08:00
__init__.py Mlx llm package (#301) 2024-01-12 10:25:56 -08:00
convert.py feat: move lora into mlx-lm (#337) 2024-01-23 08:44:37 -08:00
fuse.py feat(mlx-lm): add de-quant for fuse.py (#365) 2024-01-25 18:59:32 -08:00
generate.py fix the chinese character generation as same as PR #321 (#342) 2024-01-23 12:44:23 -08:00
LORA.md feat: move lora into mlx-lm (#337) 2024-01-23 08:44:37 -08:00
lora.py Olmo in MLX LM (#415) 2024-02-05 21:13:49 -08:00
py.typed Add py.typed to support PEP-561 (type-hinting) (#389) 2024-01-30 21:17:38 -08:00
README.md feat: move lora into mlx-lm (#337) 2024-01-23 08:44:37 -08:00
requirements.txt Update a few examples to use compile (#420) 2024-02-08 13:00:41 -08:00
UPLOAD.md Mlx llm package (#301) 2024-01-12 10:25:56 -08:00
utils.py chore(mlx-lm): add model weight index in save_weights (#413) 2024-02-06 05:32:15 -08:00

Generate Text with MLX and 🤗 Hugging Face

This an example of large language model text generation that can pull models from the Hugging Face Hub.

For more information on this example, see the README in the parent directory.

This package also supports fine tuning with LoRA or QLoRA. For more information see the LoRA documentation.