|
||
---|---|---|
.. | ||
convert.py | ||
llama.py | ||
README.md | ||
requirements.txt | ||
sample_prompt.txt |
Llama
An example of generating text with Llama (1 or 2) using MLX.
Llama is a set of open source language models from Meta AI Research12 ranging from 7B to 70B parameters. This example also supports Llama Chat and Code Llama.
Setup
Install the dependencies:
pip install -r requirements.txt
Next, download and convert the model. If you do not have access to the model weights you will need to request access from Meta:
Alternatively, you can also download a select converted checkpoints from the mlx-llama community organisation on Hugging Face and skip the conversion step.
Convert the weights with:
python convert.py --model_path <path_to_torch_model>
The conversion script will save the converted weights in the same location.
Run
Once you've converted the weights to MLX format, you can interact with the LlaMA model:
python llama.py <path_to_model> <path_to_tokenizer.model> "hello"
Run python llama.py --help
for more details.
-
For Llama v1 refer to the arXiv paper and blog post for more details. ↩︎