Goekdeniz-Guelmez
d653371e3d
nits
2025-02-22 02:12:02 +01:00
Goekdeniz-Guelmez
235348c211
generation speed improvement in training too from 3 t/s to 15 t/s
2025-02-22 02:03:01 +01:00
Goekdeniz-Guelmez
79de353530
nits
2025-02-22 01:05:58 +01:00
Goekdeniz-Guelmez
c51b0a2715
fix
2025-02-22 00:21:47 +01:00
Goekdeniz-Guelmez
710bc1490e
training mode working too got from 2 toks/sec to 30 toks/sec with raw 1.5B model
2025-02-21 22:42:15 +01:00
Goekdeniz-Guelmez
6086137131
Huge speed improvement in validation mode.
2025-02-21 22:08:49 +01:00
Goekdeniz-Guelmez
2f20107d9b
little faster generation + prints ot a examplke generatino in validation mode, more optimization in trianing function
2025-02-21 16:02:27 +01:00
Awni Hannun
3d793ecf68
Fix logits processor bugs with spec dec ( #1291 )
...
* Fix logits processor bugs with spec dec
* bump patch
2025-02-20 15:55:55 -08:00
Awni Hannun
85669451d0
Fix num layers in fine tune ( #1294 )
2025-02-20 13:32:01 -08:00
Goekdeniz-Guelmez
541f0be937
fix generation cutoff in evaluation
2025-02-17 14:39:38 +01:00
Gökdeniz Gülmez
1eea135a20
Merge branch 'ml-explore:main' into adding-GRPO-training
2025-02-17 14:25:03 +01:00
Goekdeniz-Guelmez
6a6bd53e43
removing print and switching some variables in the math
2025-02-15 15:38:51 +01:00
Goekdeniz-Guelmez
5ec4790656
removing comments + adding temperature + reward weighting
2025-02-15 15:29:22 +01:00
Goekdeniz-Guelmez
baeb9f117f
reduncancy fix + nits
2025-02-14 09:09:59 +01:00
Matthias Neumayer
96bf37008e
Update README.md to include how to set temperature ( #1280 )
...
* Update README.md to include how to set temperature
* nits
---------
Co-authored-by: Awni Hannun <awni@apple.com>
2025-02-13 19:32:56 -08:00
Awni Hannun
7b07b14e67
add logits processor to spec gen ( #1260 )
2025-02-13 19:19:53 -08:00
Goekdeniz-Guelmez
65a49dda0e
nits
2025-02-13 21:46:30 +01:00
Goekdeniz-Guelmez
8179b99436
quick prompting fix
2025-02-12 19:24:35 +01:00
Goekdeniz-Guelmez
a7273f6a56
small fix
2025-02-12 18:30:12 +01:00
Gökdeniz Gülmez
3823154014
Merge branch 'ml-explore:main' into adding-GRPO-training
2025-02-12 11:10:10 +01:00
Goekdeniz-Guelmez
e33d9d509b
updates
2025-02-12 11:07:53 +01:00
Goekdeniz-Guelmez
c42e858d7e
Merge branch 'adding-GRPO-training' of https://github.com/Goekdeniz-Guelmez/mlx-examples into adding-GRPO-training
2025-02-12 08:57:33 +01:00
Goekdeniz-Guelmez
5aeefc8c47
update new iterade batches function + nits
2025-02-12 08:57:26 +01:00
Awni Hannun
ec30dc3538
hunyuan finetune ( #1270 )
2025-02-11 16:49:35 -08:00
Awni Hannun
42413c5d85
fix lora timings after validation ( #1278 )
2025-02-11 16:48:55 -08:00
Awni Hannun
f8cbf159e0
fix sharding for more even number of layers ( #1276 )
2025-02-11 16:26:59 -08:00
Awni Hannun
e879ea70e1
fix generation evaluations ( #1277 )
2025-02-11 16:10:30 -08:00
Matt Clayton
3d677f0870
Add "from_draft" to GenerationResponse ( #1272 )
...
* Add from_draft field in GenerationResponse
* Cleanup
* Re-work for minimal changes, add test
* Fix comment
2025-02-11 15:41:02 -08:00
Goekdeniz-Guelmez
978deab589
small fix
2025-02-11 17:48:42 +01:00
Goekdeniz-Guelmez
35ecc17042
fix
2025-02-11 17:07:08 +01:00
Goekdeniz-Guelmez
e80bf95182
fix
2025-02-11 09:26:43 +01:00
Goekdeniz-Guelmez
e96afe9e9f
updates
2025-02-11 09:09:28 +01:00
Goekdeniz-Guelmez
88ca747e9e
nits
2025-02-10 19:46:19 +01:00
Goekdeniz-Guelmez
b7bc811507
nits
2025-02-10 19:45:19 +01:00
Goekdeniz-Guelmez
e5aa2c3b5d
nits
2025-02-10 17:51:14 +01:00
Goekdeniz-Guelmez
f88e897019
removing helper functions
2025-02-10 16:07:28 +01:00
Goekdeniz-Guelmez
d9da35f458
nits
2025-02-10 10:52:32 +01:00
Gökdeniz Gülmez
0dac286539
Merge branch 'main' into adding-GRPO-training
2025-02-10 10:43:22 +01:00
Chime Ogbuji
5865899c81
Completion only fine-tuning of instruction models with collections of HF datasets ( #1103 )
...
- Optional completion only fine-tuning with `--mask-prompt`
- Collections of Hugging Face datasets
---------
Co-authored-by: Awni Hannun <awni@apple.com>
2025-02-09 20:12:34 -08:00
Sri Harsha Pamu
1ced1b00ca
rm temp argument ( #1267 )
2025-02-09 11:39:11 -08:00
Goekdeniz-Guelmez
00712522ba
rebase loss calculation
2025-02-09 17:13:05 +01:00
Goekdeniz-Guelmez
a527cdb39b
fix: prevent gradients from flowing through the reference model's logits
2025-02-09 17:02:58 +01:00
Goekdeniz-Guelmez
54179901b5
fix
2025-02-09 15:41:47 +01:00
Goekdeniz-Guelmez
39e9469059
freeze ref model
2025-02-09 15:30:51 +01:00
Goekdeniz-Guelmez
9ba6146a76
fix
2025-02-09 14:32:50 +01:00
Awni Hannun
1503bd4f55
support hunyuan 7b ( #1263 )
2025-02-08 15:46:47 -08:00
Awni Hannun
31611b62d7
Add IBM granite model ( #1265 )
...
* add granite
* add thinking option
2025-02-08 15:46:15 -08:00
Awni Hannun
6120a5f376
Faster DSv2/3 expert score computation ( #1257 )
...
* fix deepseek sharding (#1242 )
* compile and use put along axis in deep seek routing function
2025-02-07 10:24:57 -08:00
Awni Hannun
52c41b5b5a
Fix prompt cache for models without chat template ( #1250 )
...
* fix deepseek sharding (#1242 )
* fix prompt cache with no chat template
2025-02-06 11:10:58 -08:00
Gökdeniz Gülmez
94dcd0f63e
Merge branch 'ml-explore:main' into adding-GRPO-training
2025-02-06 08:15:58 +01:00
Goekdeniz-Guelmez
bcfa55d882
updates
2025-02-05 15:02:12 +01:00
Goekdeniz-Guelmez
0a19522ec4
updates
2025-02-05 14:38:09 +01:00
Goekdeniz-Guelmez
35a2d99cf9
smoll fix
2025-02-05 11:30:21 +01:00
Goekdeniz-Guelmez
a33cad84b4
udpates
2025-02-05 09:48:00 +01:00
Goekdeniz-Guelmez
d84ad0cf86
fix testing
2025-02-05 08:53:30 +01:00
Goekdeniz-Guelmez
2a8e6f6e44
udpate
2025-02-05 08:47:03 +01:00
Goekdeniz-Guelmez
0a09a93454
fix cache handling
2025-02-05 08:44:06 +01:00
Pedro Cuenca
e2e5478da5
READMEs: fix typo in link, minor update. ( #1246 )
2025-02-04 11:52:32 -08:00
Goekdeniz-Guelmez
7b0141455e
better create_dataset
2025-02-04 10:43:00 +01:00
Goekdeniz-Guelmez
bd1a42ec2f
adding args into dataset handling
2025-02-04 10:22:34 +01:00
Goekdeniz-Guelmez
7173840283
first succesfull training run
2025-02-04 09:18:45 +01:00
Awni Hannun
21d0ab6e8a
fix deepseek sharding ( #1242 )
2025-02-03 16:59:50 -08:00
Gökdeniz Gülmez
0989c073b0
Optimizations for mamba1 ( #1213 )
...
* added mx.einsum() operations: before: 41.293 tokens-per-sec, after: 57.822 tokens-per-sec
* Fused Operations in delta, B, C = ... :. Before: 57.822 tokens-per-sec, after: 83.890 tokens-per-sec
* Pre-computing A_log. After: 83.890 tokens-per-sec, before: 85.848 tokens-per-sec
* Update MambaBlock, Batched Input Processing, Improved Cache Handling, Pre-computed Constants, Cleaner State Management, Explicit Return Values:. Before: 82.442 tokens-per-sec, after: 129.130 tokens-per-sec.
* cleaning up and adding apple copyright to helium modelfile
* update Copyright to this year
* nits + even faster
---------
Co-authored-by: Awni Hannun <awni.hannun@gmail.com>
2025-02-03 13:36:08 -08:00
Goekdeniz-Guelmez
ca32424043
updates
2025-02-03 21:57:26 +01:00
Goekdeniz-Guelmez
54e295ea80
fix name funcs
2025-02-03 19:56:11 +01:00
Goekdeniz-Guelmez
06f9c29c94
print func name
2025-02-03 19:47:40 +01:00
Goekdeniz-Guelmez
40bca770ae
fixes
2025-02-03 19:43:49 +01:00
Goekdeniz-Guelmez
05d921b788
optims
2025-02-03 19:37:05 +01:00
Awni Hannun
d9924d08d1
Fix no validation in lora ( #1241 )
2025-02-03 09:55:24 -08:00
Goekdeniz-Guelmez
1d9e4802f0
first working prototype, will try training out at home
2025-02-03 12:05:29 +01:00
Goekdeniz-Guelmez
23d75cd7ad
starting fist training test run
2025-02-03 10:08:28 +01:00
Goekdeniz-Guelmez
41ff5364d7
Merge branch 'adding-GRPO-training' of https://github.com/Goekdeniz-Guelmez/mlx-examples into adding-GRPO-training
2025-02-03 09:19:00 +01:00
Goekdeniz-Guelmez
a3ed632422
dataset wrapper done
2025-02-03 09:13:17 +01:00
Gökdeniz Gülmez
734d6f4a69
Merge branch 'ml-explore:main' into adding-GRPO-training
2025-02-03 09:07:20 +01:00
Goekdeniz-Guelmez
d034ca369e
adding function for R1
2025-02-03 08:26:42 +01:00
Awni Hannun
9c2ef38d4d
only download local shard ( #1240 )
2025-02-02 13:58:44 -08:00
Goekdeniz-Guelmez
243c9621d9
update lora.py
2025-01-31 21:10:44 +01:00
Goekdeniz-Guelmez
a57d553fc1
update
2025-01-31 16:57:43 +01:00
Goekdeniz-Guelmez
80bcf68956
grpo_trainer shoudl be done
2025-01-31 16:54:18 +01:00
Goekdeniz-Guelmez
6c58aa995c
updates
2025-01-31 16:27:31 +01:00
Goekdeniz-Guelmez
93370ff1c3
updates ans fixing the KL div lines
2025-01-30 23:55:40 +01:00
Gökdeniz Gülmez
b1e573d6e8
Merge branch 'ml-explore:main' into adding-GRPO-training
2025-01-29 15:07:52 +01:00
Goekdeniz-Guelmez
5e0ae83487
initial commit, gn
2025-01-29 00:19:07 +01:00
Awni Hannun
e8afb59de4
better overflow correction ( #1229 )
2025-01-28 14:37:30 -08:00
Anchen
7a83077cd7
chore(mlx-lm): support text type content in messages ( #1225 )
...
* chore(mlx-lm): support text type content
* chore: optimize the messagef content processing
* nits + format
---------
Co-authored-by: Awni Hannun <awni@apple.com>
2025-01-27 17:13:50 -08:00
Awni Hannun
f44a52e2dc
batched min p and fix spec gen sampling ( #1222 )
2025-01-27 15:40:31 -08:00
Gökdeniz Gülmez
77faa14ba4
adding support for kyutai's helium ( #1208 )
...
* initial commit
* adding helium into training
* Update ACKNOWLEDGMENTS.md
* nits
* nits
* fixes / nits
---------
Co-authored-by: Awni Hannun <awni@apple.com>
2025-01-26 07:19:07 -08:00
Awni Hannun
9a3ddc3e65
some fixes for pipeline parallel deep seek r1 ( #1216 )
2025-01-21 19:40:29 -08:00
Victor Nogueira
df1406735b
Fix dataset variable name, in datasets.py
( #1212 )
2025-01-21 14:12:43 -08:00
Jarrett
07f88f8057
fix(lora): add back store_true default args ( #1205 )
2025-01-16 11:15:42 -08:00
Awni Hannun
50f0a7f6d9
add internlm3 ( #1206 )
2025-01-15 14:55:41 -08:00
Ivan Fioravanti
6ae6c72c2e
reduction moved to CPU in case of distributed training ( #1200 )
2025-01-14 17:20:42 -08:00
Awni Hannun
c117af83b8
fix gpt bigcode ( #1204 )
2025-01-13 10:22:32 -08:00
Chime Ogbuji
0228c46434
Custom local dataset features ( #1085 )
...
* Generalize prompt_feature and completion_feature for use in local datasets to facilitate compatibility with many other training dataset formats.
* Persist configured prompt/completion key
* rebase + nits
---------
Co-authored-by: Awni Hannun <awni@apple.com>
2025-01-13 10:01:18 -08:00
Prince Canuma
bf2da36fc6
Fix Cohere2: mask shape error (long context) ( #1202 )
...
* fix mask shape error (long context)
* Update llms/mlx_lm/models/cohere2.py
Co-authored-by: Awni Hannun <awni.hannun@gmail.com>
* revert layer_idx
* black formatting
* Update cohere2.py
* format
---------
Co-authored-by: Awni Hannun <awni.hannun@gmail.com>
Co-authored-by: Awni Hannun <awni@apple.com>
2025-01-12 12:58:08 -08:00
Xingjun.Wang
514502da22
Support snapshot_download for ModelScope ( #1194 )
...
* add MLX_USE_MODELSCOPE env
* update
* update snapshot_download
* update
* remove modelscope dependency and add import check
* update
* nits
* fix
---------
Co-authored-by: wangxingjun778 <jason@U-C7X6TX5G-2239.local>
Co-authored-by: Awni Hannun <awni@apple.com>
2025-01-10 15:29:34 -08:00
Awni Hannun
93c5cfd781
Add a speculative decoding generator ( #1155 )
...
* add a speculative decoding generator
* fix
* fixes
* optional kwarg pop
2025-01-10 15:27:08 -08:00
Awni Hannun
5cae0a60e6
deepseek v3 model with pipeline parallelism ( #1191 )
...
* deepseekv3
* use upload_large_file instead of deprecated multi comit
* add pipeline generation and example
* comment
* get fp16 working
* use mlx==0.22
2025-01-09 15:55:53 -08:00
Jarrett
40b88eff48
fix(lora): config yaml & arg default merge bug ( #1196 )
2025-01-09 11:33:54 -08:00
Pedro Cuenca
b8f0cacfa8
Use upload_large_folder ( #1193 )
2025-01-07 09:18:31 -08:00