Commit Graph

  • 418d9a5511
    Feature: QDoRA (#891) Zai Thottakath 2024-09-30 10:01:11 -0500
  • 7526da7754 some simplifications and fixes Awni Hannun 2024-09-30 07:46:45 -0700
  • aa1c8abdc6
    LoRA: Support HuggingFace dataset via data parameter (#996) madroid 2024-09-30 22:36:21 +0800
  • 7ecec3b7ea feat: QDoRA with tests and a small bug fix for recalculation of self.m Zai Thottakath 2024-07-15 17:42:23 -0500
  • 3b1c70ea2d nits and doc Awni Hannun 2024-09-29 19:21:53 -0700
  • 50e5ca81a8
    Adding full finetuning (#903) Gökdeniz Gülmez 2024-09-30 02:12:47 +0200
  • 636cc0c375 some consistency changes Awni Hannun 2024-09-29 13:21:20 -0700
  • 39e5152ed8
    refactor of repetition_penalty and logits_bias to use logits_processor Nathan Ranchin 2024-09-29 18:01:37 +0200
  • cdb2d29d4b Server: support function calling madroid 2024-09-29 15:46:22 +0800
  • 6d46cbb4a6 update data argument help madroid 2024-09-29 14:44:12 +0800
  • a83ca489a4 fix pre-commit lint madroid 2024-09-29 14:22:38 +0800
  • 32a0bc903a handle load hf dataset error madroid 2024-09-29 14:21:24 +0800
  • c265cb08f0 resolve merge Awni Hannun 2024-09-28 15:03:42 -0700
  • 1c5eed70e3 nits Awni Hannun 2024-09-28 14:23:11 -0700
  • 09fd091ea6
    Merge branch 'main' into adding-full-finetuning Gökdeniz Gülmez 2024-09-28 19:54:37 +0200
  • 7ec2021bb9
    LoRA: support tools(function calling) format datasets (#995) madroid 2024-09-29 01:41:36 +0800
  • e80d1f4b20 nit in readme Awni Hannun 2024-09-28 10:38:00 -0700
  • bdad0cd4a6 nit in readme Awni Hannun 2024-09-28 10:36:54 -0700
  • f01dbe89e8 nit in readme Awni Hannun 2024-09-28 10:25:56 -0700
  • ace2bb5890
    Add logits_processor option to generate_step function (#983) nathan 2024-09-28 19:08:49 +0200
  • 83aaf0c98b add back logit bias + test Awni Hannun 2024-09-28 09:55:43 -0700
  • c8216caa61 test Awni Hannun 2024-09-28 08:32:09 -0700
  • 824f7fda58 nits + test Awni Hannun 2024-09-28 08:28:37 -0700
  • 50e4665c1d fix the variable name Nathan Ranchin 2024-09-28 16:20:26 +0200
  • 917cdb4ae3 remove the logit_bias argument from generate_step method Nathan Ranchin 2024-09-28 15:20:00 +0200
  • 3d27dbf304 Rename the tokens variable for clarity Nathan Ranchin 2024-09-10 22:10:28 +0200
  • 92ba4b2bf5 concatenation correction Nathan Ranchin 2024-09-10 22:05:03 +0200
  • 67a9325603 Add logits_processor option for the generation as in huggingface transformers library Nathan Ranchin 2024-09-10 21:24:48 +0200
  • d812516d3d
    Add /v1/models endpoint to mlx_lm.server (#984) jamesm131 2024-09-29 00:21:11 +1000
  • 92e30b71a0 doc Awni Hannun 2024-09-28 07:16:45 -0700
  • b96a64c343 simplify test Awni Hannun 2024-09-28 07:11:22 -0700
  • 76710f61af
    Adding support for mamba (#940) Gökdeniz Gülmez 2024-09-28 16:02:53 +0200
  • 1738a06c33 nits Awni Hannun 2024-09-28 06:36:47 -0700
  • a10f20654a nits Awni Hannun 2024-09-28 06:35:27 -0700
  • 03d3d19e6a fix test Awni Hannun 2024-09-27 17:49:48 -0700
  • ae9cc8c862 some speedups and cleanups Awni Hannun 2024-09-27 17:37:34 -0700
  • 97257511c4 modified mlx_lm.server.py to support prompt caching. nath1295 2024-09-27 17:49:51 +0100
  • 7e98499ee3 Better support for rotating KV-cache and add stop word list as an argument for generate and stream_generate nath1295 2024-09-27 13:43:45 +0100
  • d6cd23de16
    Update README.md mark 2024-09-26 17:21:03 +0100
  • bb433e2b78
    Update generate.py mark 2024-09-26 16:59:01 +0100
  • e5c98f4715
    Update utils.py mark 2024-09-26 16:58:02 +0100
  • 1b6440f4ae
    Merge branch 'ml-explore:main' into adding-support-for-mamba Gökdeniz Gülmez 2024-09-26 17:56:02 +0200
  • 45865cbaee
    Merge branch 'ml-explore:main' into adding-full-finetuning Gökdeniz Gülmez 2024-09-26 17:55:52 +0200
  • dafda90980
    Update README.md mark 2024-09-26 16:52:46 +0100
  • 26cb8f444c
    Update README.md mark 2024-09-26 16:52:37 +0100
  • 03332d863b fix spelling errors madroid 2024-09-26 11:41:42 +0800
  • e776c970f7
    Fix llava model when using text-only prompt (#998) Cheng 2024-09-25 23:19:41 +0900
  • 5731995652 Fix llava model when using text-only prompt Cheng 2024-09-25 15:31:43 +0900
  • fe0de13a9d
    Merge branch 'ml-explore:main' into adding-support-for-mamba Gökdeniz Gülmez 2024-09-24 14:51:03 +0200
  • 3857a77093
    Merge branch 'ml-explore:main' into adding-full-finetuning Gökdeniz Gülmez 2024-09-24 14:50:54 +0200
  • 9bb2dd62f3
    Encodec (#991) Awni Hannun 2024-09-23 11:39:25 -0700
  • 4111473f9d nits Awni Hannun 2024-09-23 11:38:00 -0700
  • b101182853 update tests to check hf_cache for models jamesm131 2024-09-22 20:11:23 +1000
  • e17a7dc443 Check hf_cache for mlx models jamesm131 2024-09-22 20:10:56 +1000
  • e4eb4e29b7 Add test for new 'models' server endpoint jamesm131 2024-09-12 20:09:36 +1000
  • a599190a8b Add 'models' endpoint to server jamesm131 2024-09-12 20:08:24 +1000
  • eb016aea9c
    Merge branch 'ml-explore:main' into adding-full-finetuning Gökdeniz Gülmez 2024-09-21 23:59:50 +0200
  • d5ff17c407
    Merge branch 'ml-explore:main' into adding-support-for-mamba Gökdeniz Gülmez 2024-09-21 23:59:40 +0200
  • dcad8339e1 LoRA: split small functions madroid 2024-09-22 01:49:12 +0800
  • aed1a0fcac LoRA: Extract the load_custom_hf_dataset function madroid 2024-09-22 01:46:44 +0800
  • fee252aa4f LoRA: support huggingface dataset via data argument madroid 2024-09-22 01:02:09 +0800
  • 59e29ba6bf LoRA: remove ToolsDataset madroid 2024-09-22 00:50:12 +0800
  • 1034aa4af1 Revert "LoRA: Split small function" madroid 2024-09-22 00:44:59 +0800
  • 03390a7e52 Revert "LoRA: pre-commit fix" madroid 2024-09-22 00:44:21 +0800
  • 796d5e40e4
    Fix export to gguf (#993) Angelos Katharopoulos 2024-09-20 13:33:45 -0700
  • b94b7e0fe7 LoRA: pre-commit fix madroid 2024-09-20 12:10:25 +0800
  • 30b39c7115 LoRA: add tools format to lora docs madroid 2024-09-20 11:34:27 +0800
  • 3f6a5f19fd LoRA: Split small function madroid 2024-09-20 11:06:06 +0800
  • bfd4ba2347 LoRA: support fine-tuning tools datasets madroid 2024-09-20 10:58:11 +0800
  • 9457329ed3 reformat and big clean up, final commit Goekdeniz-Guelmez 2024-09-19 18:13:40 +0200
  • f3c6ed99c8 fix example, support batched inference Awni Hannun 2024-09-18 06:01:04 -0700
  • 30e58ccf10 audio encodec Awni Hannun 2024-09-17 17:08:22 -0700
  • c0afcb7a5d use custom LSTM kernel Awni Hannun 2024-09-17 17:02:31 -0700
  • a6516b100f fix mlx version Awni Hannun 2024-09-16 16:21:04 -0700
  • 82e5c66385 fix for rnn layer Awni Hannun 2024-09-16 16:20:37 -0700
  • 0a73862430 use fast group norm Awni Hannun 2024-09-15 08:47:10 -0700
  • c3209fd29a nits Awni Hannun 2024-09-14 17:03:59 -0700
  • cead1541c5 works Awni Hannun 2024-09-14 16:35:51 -0700
  • 2d82579bc8 initial encodec Awni Hannun 2024-09-13 14:22:39 -0700
  • 13af75d88a clean up, reformating will come later Goekdeniz-Guelmez 2024-09-18 14:44:49 +0200
  • 399de78f51 removing extra mamba-infer.py file Goekdeniz-Guelmez 2024-09-18 13:37:43 +0200
  • 40f9e83306 Forther update to trainer/utils for correct layer selection. Successfull training Goekdeniz-Guelmez 2024-09-18 13:24:39 +0200
  • 602c9f18bd clean up Goekdeniz-Guelmez 2024-09-18 11:13:22 +0200
  • 511cdf89b1 update trainer/utils for right initialisation of the layers for LoRA, but not working. Goekdeniz-Guelmez 2024-09-18 11:04:01 +0200
  • 407bdefc76 Fix export to gguf Angelos Katharopoulos 2024-09-17 22:34:10 -0700
  • f530f56df2
    don't use internal exception (#990) Awni Hannun 2024-09-17 16:22:48 -0700
  • b0ba2342b1 don't use internal exception Awni Hannun 2024-09-17 16:19:38 -0700
  • 3ef1011aff modify generate_step in utils for cache history restoration with python api nath1295 2024-09-17 20:49:46 +0100
  • dee703eacf
    feat: simplify data handling Saurav Maheshkar 2024-09-14 23:31:17 +0100
  • 6c2369e4b9
    Fix bug in upload + docs nit (#981) Awni Hannun 2024-09-07 14:46:57 -0700
  • 29023ea61d nit Awni Hannun 2024-09-07 08:25:35 -0700
  • 11a2daebf3 fix bug in upload + docs nit Awni Hannun 2024-09-07 08:23:49 -0700
  • c3e3411756
    Update LLM generation docs to use chat template (#973) Awni Hannun 2024-09-07 06:06:15 -0700
  • 0a0ededeb0 version Awni Hannun 2024-09-06 20:21:14 -0700
  • c32f689f94 revert Awni Hannun 2024-09-06 20:20:30 -0700
  • 19144e892c add template to model cards as well Awni Hannun 2024-09-03 13:34:58 -0700
  • ae62c7f429 fix docs Awni Hannun 2024-09-03 13:25:37 -0700
  • 324184d670
    Fix the cache_prompt (#979) Angelos Katharopoulos 2024-09-06 20:19:27 -0700
  • e568161b75 Fix the cache_prompt Angelos Katharopoulos 2024-09-06 18:08:48 -0700
  • 9868743762 removed max file size Goekdeniz-Guelmez 2024-09-05 10:01:58 +0200