mirror of
https://github.com/ml-explore/mlx-examples.git
synced 2025-09-01 04:14:38 +08:00
Example reading directly from gguf file (#222)
* Draft of tiny llama from gguf * Transpose all * No transposition with new layout * Read config from gguf * Create tokenizer from gguf * move gguf and update to be similar to hf_llm * change model to HF style + updates to REAMDE * nits in REAMDE * nit readme * only use mlx for metadata * fix eos/bos tokenizer * fix tokenization * quantization runs * 8-bit works * tokenizer fix * bump mlx version --------- Co-authored-by: Juarez Bochi <juarez.bochi@grammarly.com> Co-authored-by: Awni Hannun <awni@apple.com>
This commit is contained in:
4
llms/gguf_llm/requirements.txt
Normal file
4
llms/gguf_llm/requirements.txt
Normal file
@@ -0,0 +1,4 @@
|
||||
mlx>=0.0.11
|
||||
numpy
|
||||
protobuf==3.20.0
|
||||
sentencepiece
|
Reference in New Issue
Block a user