mirror of
https://github.com/ml-explore/mlx-examples.git
synced 2025-06-26 02:33:23 +08:00
46 lines
1.2 KiB
Markdown
46 lines
1.2 KiB
Markdown
![]() |
# Deepseek Coder
|
||
|
|
||
|
Deepseek Coder is a family of code generating language models based on the
|
||
|
Llama architecture.[^1] The models were trained from scratch on a corpus of 2T
|
||
|
tokens, with a composition of 87% code and 13% natural language containing both
|
||
|
English and Chinese.
|
||
|
|
||
|
### Setup
|
||
|
|
||
|
Install the dependencies:
|
||
|
|
||
|
```
|
||
|
pip install -r requirements.txt
|
||
|
```
|
||
|
|
||
|
Next, download and convert the model.
|
||
|
|
||
|
```sh
|
||
|
python convert.py --hf-path <path_to_huggingface_model>
|
||
|
```
|
||
|
|
||
|
To generate a 4-bit quantized model, use `-q`. For a full list of options run:
|
||
|
|
||
|
```
|
||
|
python convert.py --help
|
||
|
```
|
||
|
|
||
|
The converter downloads the model from Hugging Face. The default model is
|
||
|
`deepseek-ai/deepseek-coder-6.7b-instruct`. Check out the [Hugging Face
|
||
|
page](https://huggingface.co/deepseek-ai) to see a list of available models.
|
||
|
|
||
|
By default, the conversion script will save the converted `weights.npz`,
|
||
|
tokenizer, and `config.json` in the `mlx_model` directory.
|
||
|
|
||
|
### Run
|
||
|
|
||
|
Once you've converted the weights, you can interact with the Deepseek coder
|
||
|
model:
|
||
|
|
||
|
```
|
||
|
python deepseek_coder.py --prompt "write a quick sort algorithm in python."
|
||
|
```
|
||
|
|
||
|
[^1]: For more information [blog post](https://deepseekcoder.github.io/) by
|
||
|
DeepSeek AI
|