.. |
examples
|
rm temp argument (#1267)
|
2025-02-09 11:39:11 -08:00 |
models
|
Add plamo-2-1b model (#1283)
|
2025-02-24 19:24:43 -08:00 |
tuner
|
Fix num layers in fine tune (#1294)
|
2025-02-20 13:32:01 -08:00 |
__init__.py
|
Fix detokenizer space match for quote (#1072)
|
2024-10-27 15:06:07 -07:00 |
_version.py
|
Fix logits processor bugs with spec dec (#1291)
|
2025-02-20 15:55:55 -08:00 |
cache_prompt.py
|
Fix prompt cache for models without chat template (#1250)
|
2025-02-06 11:10:58 -08:00 |
chat.py
|
fix encoding with special tokens + chat template (#1189)
|
2025-01-03 10:50:59 -08:00 |
convert.py
|
Mixed quant recipes (#1300)
|
2025-02-26 11:32:36 -08:00 |
evaluate.py
|
Use max tokens from options in mlx_lm evaluate (#1302)
|
2025-02-26 15:46:16 -08:00 |
fuse.py
|
Adding full finetuning (#903)
|
2024-09-29 17:12:47 -07:00 |
generate.py
|
Add IBM granite model (#1265)
|
2025-02-08 15:46:15 -08:00 |
gguf.py
|
Fix export to gguf (#993)
|
2024-09-20 13:33:45 -07:00 |
LORA.md
|
Completion only fine-tuning of instruction models with collections of HF datasets (#1103)
|
2025-02-09 20:12:34 -08:00 |
lora.py
|
Fix num layers in fine tune (#1294)
|
2025-02-20 13:32:01 -08:00 |
MANAGE.md
|
Add model management functionality for local caches (#736)
|
2024-05-03 12:20:13 -07:00 |
manage.py
|
fix manage for new transformers (#1304)
|
2025-02-26 15:44:57 -08:00 |
MERGE.md
|
Create executables for generate, lora, server, merge, convert (#682)
|
2024-04-16 16:08:49 -07:00 |
merge.py
|
Create executables for generate, lora, server, merge, convert (#682)
|
2024-04-16 16:08:49 -07:00 |
py.typed
|
Add py.typed to support PEP-561 (type-hinting) (#389)
|
2024-01-30 21:17:38 -08:00 |
README.md
|
feat: move lora into mlx-lm (#337)
|
2024-01-23 08:44:37 -08:00 |
requirements.txt
|
deepseek v3 model with pipeline parallelism (#1191)
|
2025-01-09 15:55:53 -08:00 |
sample_utils.py
|
batched min p and fix spec gen sampling (#1222)
|
2025-01-27 15:40:31 -08:00 |
SERVER.md
|
Fix object property value in mlx_lm.server chat completions response to match OpenAI spec (#1119)
|
2024-11-24 16:37:37 -08:00 |
server.py
|
chore(mlx-lm): support text type content in messages (#1225)
|
2025-01-27 17:13:50 -08:00 |
tokenizer_utils.py
|
Completion only fine-tuning of instruction models with collections of HF datasets (#1103)
|
2025-02-09 20:12:34 -08:00 |
UPLOAD.md
|
Mlx llm package (#301)
|
2024-01-12 10:25:56 -08:00 |
utils.py
|
Mixed quant recipes (#1300)
|
2025-02-26 11:32:36 -08:00 |