.. |
examples
|
Generation refactor: part 2 (#1099)
|
2024-11-23 11:47:06 -08:00 |
models
|
Length masking for batch inputs (#1173)
|
2024-12-18 19:43:52 -08:00 |
tuner
|
Add support for cohere2 (#1157)
|
2024-12-16 08:01:03 -08:00 |
__init__.py
|
Fix detokenizer space match for quote (#1072)
|
2024-10-27 15:06:07 -07:00 |
_version.py
|
Bpe stream without space (#1154)
|
2024-12-12 13:13:50 -08:00 |
awq.py
|
Add learned AWQ quantization
|
2024-12-19 19:13:22 -08:00 |
cache_prompt.py
|
Allow prompt callback to generate_step (#1133)
|
2024-12-03 16:17:14 -08:00 |
chat.py
|
Fix max_tokens (#1148)
|
2024-12-10 11:26:04 -08:00 |
convert.py
|
override dtype with quant (#1062)
|
2024-10-22 09:56:45 -07:00 |
evaluate.py
|
chore: update evaluate.py (#1159)
|
2024-12-15 06:06:29 -08:00 |
fuse.py
|
Adding full finetuning (#903)
|
2024-09-29 17:12:47 -07:00 |
generate.py
|
Fix no template prompt + top_k sampling (#1166)
|
2024-12-18 18:46:50 -08:00 |
gguf.py
|
Fix export to gguf (#993)
|
2024-09-20 13:33:45 -07:00 |
LORA.md
|
LoRA: update tools datasets docs (#1063)
|
2024-10-22 12:19:11 -07:00 |
lora.py
|
LoRA: Support HuggingFace dataset via data parameter (#996)
|
2024-09-30 07:36:21 -07:00 |
MANAGE.md
|
Add model management functionality for local caches (#736)
|
2024-05-03 12:20:13 -07:00 |
manage.py
|
Add model management functionality for local caches (#736)
|
2024-05-03 12:20:13 -07:00 |
MERGE.md
|
Create executables for generate, lora, server, merge, convert (#682)
|
2024-04-16 16:08:49 -07:00 |
merge.py
|
Create executables for generate, lora, server, merge, convert (#682)
|
2024-04-16 16:08:49 -07:00 |
py.typed
|
Add py.typed to support PEP-561 (type-hinting) (#389)
|
2024-01-30 21:17:38 -08:00 |
README.md
|
feat: move lora into mlx-lm (#337)
|
2024-01-23 08:44:37 -08:00 |
requirements.txt
|
Wire models in MLX LM (#1069)
|
2024-10-31 08:17:14 -07:00 |
sample_utils.py
|
Fix no template prompt + top_k sampling (#1166)
|
2024-12-18 18:46:50 -08:00 |
SERVER.md
|
Fix object property value in mlx_lm.server chat completions response to match OpenAI spec (#1119)
|
2024-11-24 16:37:37 -08:00 |
server.py
|
[mlx-lm] Use top p in server (#1144)
|
2024-12-12 11:12:21 -08:00 |
tokenizer_utils.py
|
Fix decoding manually added tokens (#1164)
|
2024-12-17 09:54:29 -08:00 |
UPLOAD.md
|
Mlx llm package (#301)
|
2024-01-12 10:25:56 -08:00 |
utils.py
|
Add support for cohere2 (#1157)
|
2024-12-16 08:01:03 -08:00 |