Commit Graph

  • 69a8f11f7b
    Merge branch 'ml-explore:main' into adding-dpo-training Gökdeniz Gülmez 2025-01-22 14:18:24 +0100
  • 9a3ddc3e65
    some fixes for pipeline parallel deep seek r1 (#1216) Awni Hannun 2025-01-21 19:40:29 -0800
  • a4b716e65d small optimization Goekdeniz-Guelmez 2025-01-22 00:15:02 +0100
  • df1406735b
    Fix dataset variable name, in datasets.py (#1212) Victor Nogueira 2025-01-21 23:12:43 +0100
  • 12e9f34524 removing unnessesairy lines and cleaning up Goekdeniz-Guelmez 2025-01-21 23:06:40 +0100
  • c13de475f6 removing custom RMSNorm class Goekdeniz-Guelmez 2025-01-21 22:52:45 +0100
  • a6a92cb91f codestral inference exxtually works now Goekdeniz-Guelmez 2025-01-21 21:01:39 +0100
  • 5a6ada2df0 getting reall closer: Goekdeniz-Guelmez 2025-01-21 20:44:51 +0100
  • eb432f4b7d inference with the origional mamba2 model woirks but still not with codestral. working: rokyang/mamba2-130m-hf rokyang/mamba2-370m-hf rokyang/mamba2-780m-hf rokyang/mamba2-1.3b-hf rokyang/mamba2-2.7b-hf Goekdeniz-Guelmez 2025-01-21 19:37:38 +0100
  • be4bc7a090
    Merge branch 'ml-explore:main' into adding-support-for-mamba2 Gökdeniz Gülmez 2025-01-21 10:57:21 +0100
  • 1c0e33615a stereo musicgen model output properly encoded/decoded and output as stereo wav David Van de Ven 2025-01-20 15:41:40 -0800
  • 679d317af0 some fixes for pipeline parallel deep seek r1 Awni Hannun 2025-01-20 15:27:53 -0800
  • 310d7ddf61 cli: align directory handling with codebase style Cavit Erginsoy 2025-01-20 22:30:46 +0000
  • 2dff75fba5 cli: align directory handling with codebase style Cavit Erginsoy 2025-01-20 22:29:09 +0000
  • cde50ec846 cli: align directory handling with codebase style Cavit Erginsoy 2025-01-20 22:27:15 +0000
  • de4f28f3d7 Remove redundant type=str from audio argument since argparse handles strings by default Cavit Erginsoy 2025-01-20 22:07:17 +0000
  • 6850fce80e whisper: add directory transcription support Cavit Erginsoy 2025-01-20 21:59:11 +0000
  • e96c17d061 inference works Goekdeniz-Guelmez 2025-01-20 19:50:08 +0100
  • db514f24c8 update Goekdeniz-Guelmez 2025-01-20 19:44:05 +0100
  • dfd51f16d6 Update MambaBlock, Batched Input Processing, Improved Cache Handling, Pre-computed Constants, Cleaner State Management, Explicit Return Values:. Before: 82.442 tokens-per-sec, after: 129.130 tokens-per-sec. Goekdeniz-Guelmez 2025-01-20 18:59:16 +0100
  • db582e4f9e Pre-computing A_log. After: 83.890 tokens-per-sec, before: 85.848 tokens-per-sec Goekdeniz-Guelmez 2025-01-20 18:42:39 +0100
  • 9494a275ac Fused Operations in delta, B, C = ... :. Before: 57.822 tokens-per-sec, after: 83.890 tokens-per-sec Goekdeniz-Guelmez 2025-01-20 18:39:22 +0100
  • e43ac7c90e added mx.einsum() operations: before: 41.293 tokens-per-sec, after: 57.822 tokens-per-sec Goekdeniz-Guelmez 2025-01-20 18:37:58 +0100
  • 531ac96481 fixing cache Goekdeniz-Guelmez 2025-01-20 18:26:21 +0100
  • ff0372103f
    Fix dataset variable name, in datasets.py Victor Nogueira 2025-01-20 11:36:20 +0100
  • 5c89d1f6a6 Add instruct tuning support to LoRA training Anupam Mediratta 2025-01-20 11:42:13 +0530
  • 61cd25362c nits Goekdeniz-Guelmez 2025-01-19 13:46:20 +0100
  • 363bde634e fixes Goekdeniz-Guelmez 2025-01-19 13:45:33 +0100
  • ea0d11cd2f update Goekdeniz-Guelmez 2025-01-19 02:05:43 +0100
  • 2a5b315f60 update ACKNOWLEDGMENTS.md Goekdeniz-Guelmez 2025-01-19 02:05:06 +0100
  • 424cb854e9 nits Goekdeniz-Guelmez 2025-01-19 02:03:50 +0100
  • 9ede9db19b nits Goekdeniz-Guelmez 2025-01-19 02:03:31 +0100
  • fa80d081f2 finish Goekdeniz-Guelmez 2025-01-19 01:58:29 +0100
  • 7d279b51ef remerge with dpo Goekdeniz-Guelmez 2025-01-19 01:14:08 +0100
  • 477000ec9d removing unneeded functions Goekdeniz-Guelmez 2025-01-19 01:13:17 +0100
  • a9b7609118 initial commit Goekdeniz-Guelmez 2025-01-19 01:09:43 +0100
  • 51fd621fdb nits Goekdeniz-Guelmez 2025-01-19 01:03:07 +0100
  • 040f7c38ac update ACKNOWLEDGMENTS.md Goekdeniz-Guelmez 2025-01-19 00:53:41 +0100
  • 06a9f5d106 update lora_config.yaml Goekdeniz-Guelmez 2025-01-19 00:53:41 +0100
  • 1b4e19675d update LORA.md Goekdeniz-Guelmez 2025-01-19 00:48:45 +0100
  • 582f979dfd fixing reference model loading and freezing Goekdeniz-Guelmez 2025-01-19 00:41:27 +0100
  • 1ff788821c initial commit Goekdeniz-Guelmez 2025-01-19 00:19:36 +0100
  • 9b54443d8e
    Merge branch 'ml-explore:main' into adding-support-for-helium Gökdeniz Gülmez 2025-01-18 22:58:19 +0100
  • 1512f1a2ad nits Goekdeniz-Guelmez 2025-01-18 21:01:05 +0100
  • 62842d218d nits Goekdeniz-Guelmez 2025-01-18 20:59:02 +0100
  • 342fa4af66 Update ACKNOWLEDGMENTS.md Goekdeniz-Guelmez 2025-01-18 20:56:57 +0100
  • 84055470ab adding helium into training Goekdeniz-Guelmez 2025-01-18 20:52:00 +0100
  • 0e9fd16b26 initial commit Goekdeniz-Guelmez 2025-01-18 20:35:25 +0100
  • 07f88f8057
    fix(lora): add back store_true default args (#1205) Jarrett 2025-01-16 12:15:42 -0700
  • dd9f26e604 Merge branch 'main' into pr/1048 apolinário 2025-01-16 16:20:05 +0100
  • 50f0a7f6d9
    add internlm3 (#1206) Awni Hannun 2025-01-15 14:55:41 -0800
  • eec093ad01 add internlm3 Awni Hannun 2025-01-15 08:22:19 -0800
  • 6ae6c72c2e
    reduction moved to CPU in case of distributed training (#1200) Ivan Fioravanti 2025-01-15 02:20:42 +0100
  • 5278143c1b Fix test Angelos Katharopoulos 2025-01-14 17:14:44 -0800
  • 2a9ed5d2c9
    fix(lora): add back store_true default args jjaareet 2025-01-14 13:12:12 -0700
  • b9eff0d744 Improve printing for FLUX distributed training flux-dist-improv Angelos Katharopoulos 2025-01-13 22:32:35 -0800
  • 2e08e8b96c moving all distributed ops to cpu ivanfioravanti 2025-01-13 23:06:58 +0000
  • dd4957f3da adding correct initialisation of dt, A and D Goekdeniz-Guelmez 2025-01-13 21:28:43 +0100
  • 5509ef8e52
    Merge branch 'ml-explore:main' into adding-support-for-mamba2 Gökdeniz Gülmez 2025-01-13 20:16:04 +0100
  • c117af83b8
    fix gpt bigcode (#1204) Awni Hannun 2025-01-13 10:22:32 -0800
  • 0228c46434
    Custom local dataset features (#1085) Chime Ogbuji 2025-01-13 13:01:18 -0500
  • 7499720b09 rebase + nits Awni Hannun 2025-01-13 09:43:23 -0800
  • 40438b1371 Persist configured prompt/completion key Chime Ogbuji 2024-11-02 19:02:47 -0400
  • db9898d104 Generalize prompt_feature and completion_feature for use in local datasets to facilitate compatibility with many other training dataset formats. Chime Ogbuji 2024-11-02 18:05:27 -0400
  • 471a01fdcf fix gpt bigcode Awni Hannun 2025-01-13 07:40:28 -0800
  • bf2da36fc6
    Fix Cohere2: mask shape error (long context) (#1202) Prince Canuma 2025-01-12 21:58:08 +0100
  • 89c49d6e23 format Awni Hannun 2025-01-12 11:02:51 -0800
  • 0e908bddff
    Update cohere2.py Awni Hannun 2025-01-12 10:58:18 -0800
  • d7638e029c black formatting Prince Canuma 2025-01-12 15:27:10 +0100
  • 9d7f38abb8 revert layer_idx Prince Canuma 2025-01-12 15:25:02 +0100
  • ad93729dce
    Update llms/mlx_lm/models/cohere2.py Prince Canuma 2025-01-12 15:23:19 +0100
  • 1107364c3a fix mask shape error (long context) Prince Canuma 2025-01-11 21:56:00 +0100
  • ff1719afc3 reduction moved to CPU in case of distributed training ivanfioravanti 2025-01-11 00:32:54 +0100
  • 514502da22
    Support snapshot_download for ModelScope (#1194) Xingjun.Wang 2025-01-11 07:29:34 +0800
  • 93c5cfd781
    Add a speculative decoding generator (#1155) Awni Hannun 2025-01-10 15:27:08 -0800
  • a84bb0589e fix Awni Hannun 2025-01-10 15:26:10 -0800
  • d20181e692 nits Awni Hannun 2025-01-10 15:23:27 -0800
  • c781df114a update wangxingjun778 2025-01-11 03:13:28 +0800
  • 5303c2bafd remove modelscope dependency and add import check wangxingjun778 2025-01-11 03:08:38 +0800
  • a63aabcd8e ok Merge branch 'main' of github.com:wangxingjun778/mlx-examples into add_modelscope wangxingjun778 2025-01-11 01:41:57 +0800
  • 28eebfe5bf Utilze a specific model version from HuggingFace during LoRA fine-tuning Sindhu Satish 2025-01-09 21:01:30 -0800
  • a8da484c8b optional kwarg pop Awni Hannun 2024-12-13 21:03:30 -0800
  • 8c0b4ee7f3 fixes Awni Hannun 2024-12-13 20:21:34 -0800
  • e34ecb79b4 fix Awni Hannun 2024-12-13 18:44:56 -0800
  • f01bc5fb17 add a speculative decoding generator Awni Hannun 2024-12-13 17:57:44 -0800
  • 5cae0a60e6
    deepseek v3 model with pipeline parallelism (#1191) Awni Hannun 2025-01-09 15:55:53 -0800
  • 26e2d97e27 comment Awni Hannun 2025-01-09 15:30:16 -0800
  • 5e8f88d079 clang format Awni Hannun 2025-01-09 15:21:17 -0800
  • 2797c438bb some cleanup Awni Hannun 2025-01-09 12:21:31 -0800
  • 40b88eff48
    fix(lora): config yaml & arg default merge bug (#1196) Jarrett 2025-01-09 12:33:54 -0700
  • 761b2c9886 export and run llama in C++ Awni Hannun 2025-01-08 16:17:45 -0800
  • 28f6275852
    fix(lora): config yaml & arg default merge bug jjaareet 2025-01-08 12:26:54 -0700
  • 2d32c6d29d ok Merge branch 'add_modelscope' of github.com:wangxingjun778/mlx-examples into add_modelscope xingjun.wang 2025-01-08 02:58:07 +0800
  • bb87aec13e update xingjun.wang 2025-01-08 02:56:09 +0800
  • 6bb7163013 ok Merge branch 'main' of github.com:ml-explore/mlx-examples into add_modelscope xingjun.wang 2025-01-08 02:46:56 +0800
  • 5c472e7721
    Merge branch 'ml-explore:main' into add_modelscope Xingjun.Wang 2025-01-08 02:44:43 +0800
  • b8f0cacfa8
    Use upload_large_folder (#1193) Pedro Cuenca 2025-01-07 18:18:31 +0100
  • fe45a37286 Use upload_large_folder Pedro Cuenca 2025-01-07 11:32:16 +0100
  • d6bb6e105e use mlx==0.22 Awni Hannun 2025-01-06 13:28:22 -0800
  • 46d53ce110 get fp16 working Awni Hannun 2025-01-06 13:23:53 -0800