Commit Graph

  • d7d70487eb fix sliding window Prince Canuma 2024-12-14 17:06:57 +0100
  • 0337646b4e add to tuner Prince Canuma 2024-12-14 16:39:55 +0100
  • 2f443cc6d7 add tests Prince Canuma 2024-12-14 16:39:46 +0100
  • 52595dafae fix tests and sliding window attention Prince Canuma 2024-12-14 16:39:22 +0100
  • 5d8b36ce7c revert to act_fn to silu Prince Canuma 2024-12-14 16:22:00 +0100
  • d7c64c621a add support for cohere2 Prince Canuma 2024-12-14 16:15:24 +0100
  • e5ff577dab add support for cohere2 Prince Canuma 2024-12-14 16:03:45 +0100
  • dff4e52910 adding the modelnames in the LORA.md file and removing unused functions from mamba2.py Goekdeniz-Guelmez 2024-12-12 22:52:00 +0100
  • 9f2ea5892e
    Bpe stream without space (#1154) Awni Hannun 2024-12-12 13:13:50 -0800
  • 88693bf162 version bump Awni Hannun 2024-12-12 12:33:56 -0800
  • e2ba4ceaab bpe streaming detokenization without space Awni Hannun 2024-12-12 12:33:36 -0800
  • a883e39f41 optimizing the code for faster inference but still generates giberish Goekdeniz-Guelmez 2024-12-12 21:08:33 +0100
  • 2ba0e36683
    [mlx-lm] Use top p in server (#1144) Awni Hannun 2024-12-12 11:12:21 -0800
  • 19abf3dcaa
    Replace unicode errors instead of raising exception (#1146) Angelos Katharopoulos 2024-12-12 11:10:41 -0800
  • 62391bec30 Fix bug Angelos Katharopoulos 2024-12-12 11:08:23 -0800
  • 2024181b7c couple other fixes Awni Hannun 2024-12-12 10:51:38 -0800
  • 2277033a24 use top p in server Awni Hannun 2024-12-08 20:16:44 -0800
  • 6ef5ca4ce5 Change the last_segment logic in tokenizer utils Angelos Katharopoulos 2024-12-12 10:37:29 -0800
  • 06af3c9b0e
    Add finish_reason in GenerationResponse (#1153) madroid 2024-12-13 02:37:40 +0800
  • 77b42b7c8b
    fix llava (#1149) Awni Hannun 2024-12-12 10:37:26 -0800
  • f67631679d Add finish_reason in GenerationResponse madroid 2024-12-11 20:56:24 +0800
  • 1fd2bfa5f5 Fix for stream_generate() takes 3 positional arguments but 4 were given Ivan Fioravanti 2024-12-11 06:17:33 +0100
  • aac5116347 fix llava Awni Hannun 2024-12-10 12:57:17 -0800
  • c1d9ec329c
    Merge branch 'ml-explore:main' into adding-support-for-mamba2 Gökdeniz Gülmez 2024-12-10 20:43:11 +0100
  • 135c5818c1
    Fix max_tokens (#1148) Alex Barron 2024-12-10 11:26:04 -0800
  • e294f06acc
    Fix max_tokens Alex Barron 2024-12-10 11:00:49 -0800
  • 184d3d3267 clean up Goekdeniz-Guelmez 2024-12-10 18:20:13 +0100
  • 80e88b4f4d nits Goekdeniz-Guelmez 2024-12-10 18:18:59 +0100
  • b10afe3662 nits Goekdeniz-Guelmez 2024-12-10 18:15:12 +0100
  • 9f8a6a3509 inference on codestral works but is giberish Goekdeniz-Guelmez 2024-12-10 17:34:44 +0100
  • ddad2105ef
    Merge branch 'main' into adding-support-for-mamba2 Gökdeniz Gülmez 2024-12-10 14:32:44 +0100
  • b444acfd69 Replace unicode errors instead of raising exception Angelos Katharopoulos 2024-12-09 09:16:51 -0800
  • 12083c4b7e
    Support for multiple EOS tokens (#1141) madroid 2024-12-10 00:53:58 +0800
  • 9b53599e6c nits Awni Hannun 2024-12-09 08:51:22 -0800
  • 5687d5b99b
    Adds EXAONE architecture. (#1145) n8programs 2024-12-09 10:58:25 -0500
  • 9f3531c814 clean up and fix rope Awni Hannun 2024-12-09 07:44:55 -0800
  • d90c6af11a clean up and fix rope Awni Hannun 2024-12-09 07:44:32 -0800
  • b2b16500fb format Awni Hannun 2024-12-09 06:50:16 -0800
  • aa3defcfb0 nits + format Awni Hannun 2024-12-09 06:46:28 -0800
  • 893b3f085e
    Change Flux default max_shift to 1.15 to match the official one (#1137) hehua2008 2024-12-09 15:29:48 +0800
  • ed91bbc4dc
    Fix final message at end of flux training (#1143) Peter Sibley 2024-12-09 02:01:53 -0500
  • 25f423eef8 Remove the message with the missing filename Angelos Katharopoulos 2024-12-08 22:58:26 -0800
  • 1fa3a10e3d Revert "fix missing --final-adapter-file crash in dreambooth.py" Angelos Katharopoulos 2024-12-08 22:56:55 -0800
  • 1fd6aae871
    Fix flux training with batch size (#1135) hehua2008 2024-12-09 14:09:04 +0800
  • 68538a6e1d
    Adds EXAONE architecture. n8programs 2024-12-09 00:52:05 -0500
  • ceb0ae3416 Revert "Fix shapes mismatch error in FluxPipeline.training_loss when batch-size >= 2" Angelos Katharopoulos 2024-12-08 21:36:16 -0800
  • d50ad3ec27 Fix batched add_noise Angelos Katharopoulos 2024-12-08 21:35:16 -0800
  • 2211b27388
    Mixed Quantizations (#1132) Alex Barron 2024-12-08 14:21:50 -0800
  • 4345c7c8bb count bias too Alex Barron 2024-12-08 14:05:12 -0800
  • 46109e4141 more concise bpw Alex Barron 2024-12-08 12:44:29 -0800
  • cd8cf28c39
    mlx_lm.evaluate (#1140) Alex Barron 2024-12-08 12:20:10 -0800
  • facaf01b8d Remove model_config & add eos_token_id madroid 2024-12-08 12:08:09 +0800
  • 1cfb005647 Change _eos_token_ids type from list to set madroid 2024-12-08 11:59:12 +0800
  • f00ca80fdb fix missing --final-adapter-file crash in dreambooth.py Peter Sibley 2024-12-06 22:23:55 -0500
  • f8379fb3ef Support for multiple EOS tokens madroid 2024-12-07 11:19:20 +0800
  • 74e5b48e8f comment Alex Barron 2024-12-06 17:49:08 -0800
  • 47cbc418e2 relative import Alex Barron 2024-12-06 16:33:25 -0800
  • b1a4ba0827 create output dir Alex Barron 2024-12-06 16:31:34 -0800
  • c3e51fcf02 typo Alex Barron 2024-12-06 16:26:26 -0800
  • 80590f5ec2 add lm eval version Alex Barron 2024-12-06 16:21:20 -0800
  • dc56226bf8 only write top level results Alex Barron 2024-12-06 15:54:29 -0800
  • 04412c2edd Add evaluation script Alex Barron 2024-12-06 15:52:55 -0800
  • 76fac6eee0 Add a fix for special added tokens Billel Mokeddem 2024-12-05 22:56:49 +0400
  • 76ba80316f Chang Flux's default max_shift to 1.15 to match the official one hehua2008 2024-12-05 19:30:55 +0800
  • 3b5cd401d8 add bits per weight Alex Barron 2024-12-04 08:31:40 -0800
  • 67c68452bb Fix shapes mismatch error in FluxPipeline.training_loss when batch-size >= 2 hehua2008 2024-12-04 22:37:04 +0800
  • 64ceb62674 load q4_k_m inefficiently load-gguf Alex Barron 2024-12-03 19:54:57 -0800
  • 1727959a27
    Add mentions of MLX-my-repo. (#1129) vb 2024-12-04 04:21:39 +0100
  • b365d22785 move Awni Hannun 2024-12-03 19:21:06 -0800
  • 02649b30e5 move Awni Hannun 2024-12-03 19:20:02 -0800
  • feb7f10888 simplify Awni Hannun 2024-12-03 17:10:41 -0800
  • 1963df8565
    Allow prompt callback to generate_step (#1133) Awni Hannun 2024-12-03 16:17:14 -0800
  • 2c8d49dfcc bump version Awni Hannun 2024-12-03 07:01:46 -0800
  • 3c63be8c55 comments Awni Hannun 2024-12-03 06:28:21 -0800
  • 0ca162cfb2
    Fix data_iter in prepare_dataset from speechcommands example (#1113) sakares saengkaew 2024-12-03 14:56:07 +0700
  • 43d023948f
    add reset() Sakares Saengkaew 2024-12-03 15:53:00 +0800
  • eb9277f574
    Allow loading from diffusers ckpt (#1117) Angelos Katharopoulos 2024-12-02 13:15:50 -0800
  • 2a9294a5f0
    Fix bug in FluxSampler.timesteps method (#1131) hehua2008 2024-12-03 05:15:19 +0800
  • 21a05d14d2 nit Awni Hannun 2024-12-02 12:19:53 -0800
  • 1036798dac allow prompt callback and use in cache_prompt Awni Hannun 2024-12-02 12:16:46 -0800
  • 8801beb66f
    Add olmo2 (#1128) Awni Hannun 2024-12-02 11:42:58 -0800
  • 5828703a5a comment Alex Barron 2024-12-02 11:11:23 -0800
  • 80e5c37bb9 saving/loading mixed quantizations Alex Barron 2024-12-02 10:19:39 -0800
  • 4c55e92523 mlx_lm.server --model mlx-community/Qwen2.5-Coder-14B-Instruct-8bit --trust-remote-code --port 8722 chenguangjian.jk 2024-12-02 16:49:11 +0800
  • aef79523a2
    Merge branch 'ml-explore:main' into main chenguangjian 2024-12-02 16:42:30 +0800
  • d459cabea4 Fix bug in FluxSampler.timesteps method hehua2008 2024-12-01 13:07:48 +0800
  • 7c59bfeff2 Add mentions of MLX-my-repo. Vaibhav Srivastav 2024-11-29 14:51:23 +0100
  • fde4b4dc42 add olmo2 Awni Hannun 2024-11-27 12:15:45 -0800
  • b0ef3a4c9e add olmo2 Awni Hannun 2024-11-27 11:50:09 -0800
  • cefe793ae0
    Accept mx.array type for prompt argument for stream_generate (#1125) Neil Mehta 2024-11-26 19:51:55 -0500
  • 83bc764e78 Fix formatting Neil Mehta 2024-11-26 16:46:08 -0500
  • 152d6b1e1e Accept mx.array type for prompt argument for stream_generate Neil Mehta 2024-11-26 13:53:03 -0500
  • cfc29c29f4
    Put prompt processing in same stream (#1122) Awni Hannun 2024-11-25 09:47:00 -0800
  • 63e582a058 patch Awni Hannun 2024-11-25 08:11:59 -0800
  • a5e173802e
    docs: update stream_generate return type annotation (#1121) madroid 2024-11-26 00:10:14 +0800
  • 3586c876aa put prompt processing in same stream Awni Hannun 2024-11-25 08:06:12 -0800
  • 5cd7531c22 docs: update stream_generate return type annotation madroid 2024-11-25 14:31:35 +0800
  • adaab81029
    Allow converting models from local directories (#1118) Remixer Dec 2024-11-25 04:41:06 +0400
  • 0ffdb6dd20
    Fix object property value in mlx_lm.server chat completions response to match OpenAI spec (#1119) Kevin Conner 2024-11-24 16:37:37 -0800
  • ec494a97ec Fix object property value in mlx_lm.server chat completions response to match OpenAI spec Kevin Conner 2024-11-24 14:19:12 -0800