This commit is contained in:
Awni Hannun 2024-12-03 19:20:02 -08:00
parent feb7f10888
commit 02649b30e5

View File

@ -77,9 +77,7 @@ to see how to use the API in more detail.
The `mlx-lm` package also comes with functionality to quantize and optionally The `mlx-lm` package also comes with functionality to quantize and optionally
upload models to the Hugging Face Hub. upload models to the Hugging Face Hub.
You can convert models in the [Hugging Face You can convert models using the Python API:
Space](https://huggingface.co/spaces/mlx-community/mlx-my-repo) or using the
Python API:
```python ```python
from mlx_lm import convert from mlx_lm import convert
@ -165,6 +163,10 @@ mlx_lm.convert \
--upload-repo mlx-community/my-4bit-mistral --upload-repo mlx-community/my-4bit-mistral
``` ```
Models can also be converted and quantized directly in the
[mlx-my-repo]]https://huggingface.co/spaces/mlx-community/mlx-my-repo) Hugging
Face Space.
### Long Prompts and Generations ### Long Prompts and Generations
`mlx-lm` has some tools to scale efficiently to long prompts and generations: `mlx-lm` has some tools to scale efficiently to long prompts and generations: