Commit Graph

  • ed67d295f2
    Merge 5e8ab40d7d into 4b2a0df237 Huss 2025-06-20 16:28:13 +1200
  • 5e8ab40d7d
    Merge branch 'ml-explore:main' into main Huss 2025-06-19 15:45:31 +0300
  • 231d008065
    Merge f1b5aabd26 into 4b2a0df237 Huss 2025-06-11 18:01:05 +0300
  • c79c611ea7
    Merge 282304a87d into 4b2a0df237 Pranav 2025-06-10 21:20:21 -0400
  • 4b2a0df237
    adding wwdc25 samples (#1370) main Shashank 2025-06-10 10:23:25 -0700
  • 4aa906fbdd adding wwdc25 samples shashankprasanna 2025-06-10 09:58:35 -0700
  • f1b5aabd26
    Update ACKNOWLEDGMENTS.md Huss 2025-05-25 17:44:49 +0300
  • 33c6ff9d8f
    Merge dd9f26e604 into 977cd30242 apolinário 2025-05-05 09:43:58 -0400
  • 977cd30242
    Update lora README.md (#1365) Denrei Keith 2025-05-01 21:00:14 +0800
  • 4eb573b7d5
    Update lora README.md Denrei Keith 2025-05-01 18:05:53 +0800
  • 4c9f9f9be7
    Made llama and mistral files mypy compatible (#1359) Param Thakkar 2025-04-24 02:53:46 +0530
  • d5443aa13a revert some stuff Awni Hannun 2025-04-23 14:18:43 -0700
  • 3992ca5554 fix circle Awni Hannun 2025-04-23 14:07:24 -0700
  • b8fee34f24 Fixes paramthakkar123 2025-04-09 23:13:43 +0530
  • 8f8f9b6991 Added fixes to speculative-decoding paramthakkar123 2025-04-08 23:26:35 +0530
  • 4304f5aaf5 Added more fixes paramthakkar123 2025-04-07 08:59:07 +0530
  • 298178d669 reformatted paramthakkar123 2025-04-04 07:36:47 +0530
  • d7cab9d5f5 Made mypy compatible paramthakkar123 2025-04-04 07:34:43 +0530
  • 17c737695d doc: 1、中文说明 wujiale 2025-03-31 13:38:20 +0800
  • c52cc748f8
    Distributed FLUX (#1325) Angelos Katharopoulos 2025-03-24 22:16:48 -0700
  • d06c4cde92 Comments Angelos Katharopoulos 2025-03-24 22:15:37 -0700
  • c4d08de8b3 Update README Angelos Katharopoulos 2025-03-22 17:44:20 -0700
  • c109d9b596 Fix the seed for data parallel Angelos Katharopoulos 2025-03-22 16:50:28 -0700
  • a1e259607e Fix data parallel generation Angelos Katharopoulos 2025-03-22 16:43:40 -0700
  • 208856520d Update for nn.layers.distributed Angelos Katharopoulos 2025-03-22 16:35:32 -0700
  • 02b007f19c Reduce the number of communications Angelos Katharopoulos 2025-03-05 02:51:05 -0800
  • 7fbd1619eb Initial working distributed flux Angelos Katharopoulos 2025-03-03 23:17:30 -0800
  • c243370044
    remove mlx lm (#1353) Awni Hannun 2025-03-18 18:47:55 -0700
  • 0ac7c0330a remove mlx lm Awni Hannun 2025-03-18 13:00:28 -0700
  • 7ca05d2e51
    LoRa/README.md should be --hf-path instead of --hf-repo (#1350) Tingzhen 2025-03-16 23:02:52 -0400
  • d9e1d9c0ef
    mlx-lm move notice (#1346) Awni Hannun 2025-03-16 15:14:28 -0700
  • 3aaca78d25 LoRa/README.md should be --hf-path instead of --hf-repo du tingzhen 2025-03-16 15:16:37 -0400
  • 8680bb614e
    Update mixed_2_6 and mixed_3_6 quant predicates to explicitly pass high_bits parameter sealad886 2025-03-16 08:31:47 +0000
  • 4b3870b89e remove mlx lm tests Awni Hannun 2025-03-14 07:28:49 -0700
  • 1de4c8695a removings Goekdeniz-Guelmez 2025-03-14 08:49:20 +0100
  • 31581d8df0 mlx-lm move notice Awni Hannun 2025-03-13 20:49:10 -0700
  • 2fce02acd8
    Add support for Gemma3 (#1336) Prince Canuma 2025-03-13 16:14:25 +0100
  • f124c1e373 fix sliding window mask Awni Hannun 2025-03-13 07:46:16 -0700
  • 1dd038cbbc
    fix: remove eos_token_ids to prevent early stop B1ACK917 2025-03-13 19:00:19 +0900
  • 4c60cb8ef9 Fix prompt cache issue in server.py. Honestly I don't understand why this fixes it, but prompt_cache seems to become functional with this change. Jeonghyun Lee 2025-03-13 17:24:58 +0900
  • 3e5baf583b
    Make sure to use UTF-8 when loading tokenizer.json (#1340) Mirko Nasato 2025-03-13 02:17:14 +0000
  • 10adfa76bf multible ssd step frunctions Goekdeniz-Guelmez 2025-03-12 17:33:53 +0100
  • f6ea019f0d
    Merge branch 'ml-explore:main' into adding-GRPO-training Gökdeniz Gülmez 2025-03-12 16:53:36 +0100
  • 4a9d75ae4f
    Merge branch 'ml-explore:main' into adding-dpo-training Gökdeniz Gülmez 2025-03-12 16:53:23 +0100
  • 64ba244e55
    Merge branch 'ml-explore:main' into adding-orpo-training Gökdeniz Gülmez 2025-03-12 16:53:10 +0100
  • c83a85350b
    add: mlx_lm.bench README B1ACK917 2025-03-12 23:58:39 +0900
  • d8e6996254 Make sure to use UTF-8 when loading tokenizer.json Mirko Nasato 2025-03-12 14:24:32 +0000
  • f5cd999774
    update: pre-commit format hook B1ACK917 2025-03-12 23:10:58 +0900
  • 0cff3b781b
    Update llms/mlx_lm/models/gemma3_text.py Prince Canuma 2025-03-12 15:01:13 +0100
  • bbadd1d5af
    Update llms/mlx_lm/models/gemma3_text.py Prince Canuma 2025-03-12 15:00:49 +0100
  • 92b06ddcad
    Update llms/mlx_lm/models/gemma3_text.py Prince Canuma 2025-03-12 15:00:26 +0100
  • f034a7b72a
    Update llms/mlx_lm/models/gemma3_text.py Prince Canuma 2025-03-12 15:00:17 +0100
  • 2d13d9b60f
    Update llms/mlx_lm/models/gemma3_text.py Prince Canuma 2025-03-12 15:00:06 +0100
  • 9eb59b7250
    Update llms/mlx_lm/models/gemma3_text.py Prince Canuma 2025-03-12 14:59:56 +0100
  • dc090f26df
    Update llms/mlx_lm/models/gemma3_text.py Prince Canuma 2025-03-12 14:59:45 +0100
  • 3f70efb486 fix typo Goekdeniz-Guelmez 2025-03-12 14:49:21 +0100
  • 346d9641d7
    add: mlx_lm.bench for more convenient benchmark (#126) B1ACK917 2025-03-12 22:48:41 +0900
  • a608ae99bc update lora.md and lora_config.yaml Goekdeniz-Guelmez 2025-03-12 14:46:19 +0100
  • 0e28fdb345
    Merge branch 'ml-explore:main' into adding-reporting-to-wandb Gökdeniz Gülmez 2025-03-12 14:35:39 +0100
  • 04537fa346 adding wandb reporting to lora.py Goekdeniz-Guelmez 2025-03-12 14:31:23 +0100
  • 57175b7b95 initial commit Goekdeniz-Guelmez 2025-03-12 11:55:09 +0100
  • 822546dc91 formatting Prince Canuma 2025-03-12 10:35:12 +0100
  • 3c15130f39 add tests Prince Canuma 2025-03-12 10:34:46 +0100
  • 37cd4a382b revert Prince Canuma 2025-03-12 10:30:02 +0100
  • 2d30f6787a revert is sliding pattern Prince Canuma 2025-03-12 09:48:14 +0100
  • 645b666890 revert rmsnorm Prince Canuma 2025-03-12 09:41:42 +0100
  • 0e57d38f47 fix model loading Prince Canuma 2025-03-12 09:37:17 +0100
  • 8fd3f5a131 add support for gemma3 Prince Canuma 2025-03-12 09:12:23 +0100
  • 4c3df00162
    make_sampler creates sampler chain with all sampling parameters (#1330) Neil Mehta 2025-03-11 16:37:35 -0400
  • 21da261028 Use mx.allclose Neil Mehta 2025-03-11 10:24:26 -0400
  • 7300b3d871 udpate LORA.md Goekdeniz-Guelmez 2025-03-11 10:56:37 +0100
  • d9d77f952d update LORA.md Goekdeniz-Guelmez 2025-03-11 10:52:38 +0100
  • 9fd6a5b6d0 adding PPO like clipping adapted from trl Goekdeniz-Guelmez 2025-03-11 09:08:38 +0100
  • 06ff47012f match pytoch imeplentation for loss calculation Goekdeniz-Guelmez 2025-03-11 09:00:21 +0100
  • 64a0b0cddb
    Merge branch 'main' into adding-support-for-mamba2 Gökdeniz Gülmez 2025-03-10 19:41:27 +0100
  • e6031a78e4 Remove unnecessary mx.where Neil Mehta 2025-03-10 11:27:14 -0400
  • a81e8bcc2d Add yarn option for qwen2 Awni Hannun 2025-03-10 07:11:29 -0700
  • 7ee76a32a4 Add memory estimation tool for MLX language models Cavit Erginsoy 2025-03-10 02:59:09 +0000
  • f1961f1b79 fix batch size Goekdeniz-Guelmez 2025-03-09 00:26:41 +0100
  • e88f0fad4b clean up Goekdeniz-Guelmez 2025-03-09 00:18:33 +0100
  • 0bc2a881ad generation should be fixed now Goekdeniz-Guelmez 2025-03-09 00:16:40 +0100
  • 46d6146102
    Merge branch 'ml-explore:main' into adding-GRPO-training Gökdeniz Gülmez 2025-03-08 22:41:10 +0100
  • 4a4775557c udpate Goekdeniz-Guelmez 2025-03-08 22:38:23 +0100
  • 73cc094681 fix optimizer Goekdeniz-Guelmez 2025-03-08 22:36:20 +0100
  • 956da0ddc7 Create sampler chain Neil Mehta 2025-03-08 14:08:55 -0500
  • 932b7c0510 top_k and min_p refactor Neil Mehta 2025-03-08 10:12:28 -0500
  • 58e912966a top_p refactor Neil Mehta 2025-03-08 08:55:49 -0500
  • 700c3ef5cc
    Merge branch 'main' into adding-orpo-training Gökdeniz Gülmez 2025-03-08 10:16:40 +0100
  • 3aaf2d6c9f
    Merge branch 'main' into adding-dpo-training Gökdeniz Gülmez 2025-03-08 10:07:48 +0100
  • 292979d447
    Merge branch 'ml-explore:main' into main chaihahaha 2025-03-08 12:06:04 +0800
  • d2e02b3aae
    fix mixed quant option (#1326) Awni Hannun 2025-03-07 08:35:48 -0800
  • c72811da57 fix mixed quant option Awni Hannun 2025-03-07 06:55:35 -0800
  • 595f5da146
    remove lm head if unused (#1324) Awni Hannun 2025-03-06 15:35:47 -0800
  • 877d2a345b
    Change DEFAULT_SEED to None for stochastic generation by default (#1323) cavit99 2025-03-06 14:49:35 +0000
  • bc2fcec230
    Update llms/mlx_lm/generate.py Awni Hannun 2025-03-06 06:45:59 -0800
  • 421b0219a9
    Update llms/mlx_lm/chat.py Awni Hannun 2025-03-06 06:45:52 -0800
  • 717e415ad4 remove lm head if unused Awni Hannun 2025-03-06 06:18:46 -0800
  • 5a4252f290 Change DEFAULT_SEED to None for stochastic generation by default Cavit Erginsoy 2025-03-05 23:24:15 +0000
  • 32d10036de
    fix flaky test (#1322) Awni Hannun 2025-03-05 14:00:09 -0800
  • e150621095
    Adding multiple optimizers to mlx lm (#1315) Gökdeniz Gülmez 2025-03-05 22:54:54 +0100