mlx-examples/llms/mlx_lm/models
Anchen 30be4c4734
refactor(qwen): moving qwen into mlx-lm (#312)
* refactor(qwen): moving qwen into mlx-lm

* chore: update doc

* chore: fix type hint

* add qwen model support in convert

* chore: fix doc

* chore: only load model in quantize_model

* chore: make the convert script only copy tokenizer files instead of load it and save

* chore: update docstring

* chore: remove unnecessary try catch

* chore: clean up for tokenizer and update  transformers 4.37

* nits in README

---------

Co-authored-by: Awni Hannun <awni@apple.com>
2024-01-22 15:00:07 -08:00
..
__init__.py Mlx llm package (#301) 2024-01-12 10:25:56 -08:00
base.py Mlx llm package (#301) 2024-01-12 10:25:56 -08:00
llama.py two minor fixes (#335) 2024-01-18 14:18:13 -08:00
mixtral.py feat(mlx_lm): add mixtral support in mlx_lm (#318) 2024-01-15 07:18:14 -08:00
phi2.py chore(mlx-lm): update phi2 model args to sync with hf config format. (#311) 2024-01-13 07:51:45 -08:00
qwen.py refactor(qwen): moving qwen into mlx-lm (#312) 2024-01-22 15:00:07 -08:00