Commit Graph

  • 8071aacd98 fix-mistral-download-link (#418) Long Sha 2024-02-06 22:56:56 -05:00
  • 2303238e44 Update olmo.py (#419) Chris McMaster 2024-02-07 11:16:46 +11:00
  • 8b77677c05 chore(mlx-lm): add model weight index in save_weights (#413) Anchen 2024-02-07 00:32:15 +11:00
  • a7d139f484 fix(mlx-lm): olmo 1b model (#417) Anchen 2024-02-07 00:27:05 +11:00
  • aa7447efa2 Olmo in MLX LM (#415) Awni Hannun 2024-02-05 21:13:49 -08:00
  • 7fbca214b1 Add max sequence length argument in lora.py (#408) Ivan Fioravanti 2024-02-04 21:28:21 +01:00
  • 9d0dd34403 add qwen2 (#411) Junyang Lin 2024-02-05 00:31:38 +08:00
  • e9b32747b4 Add grad checkpointing and PE in the transformer example (#387) Angelos Katharopoulos 2024-02-01 13:04:03 -08:00
  • ec14583c2a work with tuple shape (#393) Awni Hannun 2024-02-01 13:03:47 -08:00
  • 0340113e02 BUG FIX: Decoding results in garbled text when multiple tokens represent a single character (e.g., Chinese). (#398) ZHAOKAI WANG 2024-02-01 11:27:29 +08:00
  • 94358219cf CLIP (ViT) (#315) Gabrijel Boduljak 2024-01-31 23:19:53 +01:00
  • ba3a9355d1 LoRA: Remove unnecessary model type judgments (#388) Madroid Ma 2024-02-01 03:55:27 +08:00
  • 0a49ba0697 fix(mlx-lm): apply lora layer doesn't update the lora weights (#396) Anchen 2024-02-01 06:51:26 +11:00
  • ab8bde1590 Add py.typed to support PEP-561 (type-hinting) (#389) Sugato Ray 2024-01-31 00:17:38 -05:00
  • f8fadf7a17 Fix token count computation to fix tps measurements (#392) David Koski 2024-01-30 11:24:16 -08:00
  • 614de6652f chore(mlx-lm): add reset lora layers helper (#377) Anchen 2024-01-30 15:54:49 +11:00
  • 20b969b412 Replace time.time() with time.perf_counter() as it is more suited for benchmarking (#380) Ashish 2024-01-26 15:11:38 -07:00
  • 5aa652d3c2 remove simplify (#379) Awni Hannun 2024-01-26 13:54:49 -08:00
  • 0b57f0eae6 Add StableLM-2 1.6B (#378) Ashish 2024-01-26 11:28:00 -07:00
  • 854ad8747a feat(mlx-lm): add de-quant for fuse.py (#365) Anchen 2024-01-26 13:59:32 +11:00
  • f51e98fcf1 chore(mlx-lm): truncate the input sentence to max seq len in lora iterate_batches (#373) Anchen 2024-01-26 07:38:04 +11:00
  • 0f19237fb8 docs: added missing imports (#375) Yiğit Ö. Ünver 2024-01-25 21:44:53 +03:00
  • b1dec281b3 feat(mlx-lm): add lora hypeparameters in lora layer (#366) Anchen 2024-01-25 03:11:25 +11:00
  • 5fc8668a53 fix(mlx-lm): handle legacy quant models (#369) Anchen 2024-01-25 02:44:05 +11:00
  • ab91ac1075 chore(mlx-lm): add load model with adapter and fix bug in sample (#360) Anchen 2024-01-23 19:47:39 -08:00
  • f5b80c95fb Example reading directly from gguf file (#222) Juarez Bochi 2024-01-23 18:41:54 -05:00
  • 40b61c1719 fix the chinese character generation as same as PR #321 (#342) iLoveBug 2024-01-24 04:44:23 +08:00
  • 21aa8038fb MLX LM version bump (#358) Awni Hannun 2024-01-23 09:05:57 -08:00
  • 362e88a744 feat: move lora into mlx-lm (#337) Anchen 2024-01-23 08:44:37 -08:00
  • 85c1ff8fd6 Add PLaMo-13B model as an LLM example (#303) Shunta Saito 2024-01-24 00:17:24 +09:00
  • c45c2311bd Add colorized output option to generate script (#347) Ivan Fioravanti 2024-01-23 14:25:44 +01:00
  • a445ac2895 Update docs with conda install option (#354) Sugato Ray 2024-01-23 00:14:48 -05:00
  • 42672f5446 add an option to apply the tokenizer chat template (#338) Baptiste Canton 2024-01-23 04:52:42 +01:00
  • 8022083979 feat(lora): add de-quantized support for fuse.py (#351) Anchen 2024-01-22 17:32:24 -08:00
  • 30be4c4734 refactor(qwen): moving qwen into mlx-lm (#312) Anchen 2024-01-22 15:00:07 -08:00
  • de15532da8 Feat: Bump isort version (#350) Nripesh Niketan 2024-01-21 18:35:15 +04:00
  • 1415595409 chore(lora): support mixtral in lora example (#343) Anchen 2024-01-20 06:07:45 -08:00
  • 527cea4027 chore: fix the convert.py script for weights are not sanitized and support quant for non-32 dimensions (#340) Anchen 2024-01-19 21:07:21 -08:00
  • 61297f547b Missing requirements needed for convert script (#320) bojanbabic 2024-01-18 19:04:24 -08:00
  • bcc9fc3581 two minor fixes (#335) Awni Hannun 2024-01-18 14:18:13 -08:00
  • d8680a89f9 Add argument --save-every N to lora.py for saving model regularly (#310) Zheng Qu 2024-01-17 05:03:33 +01:00
  • b4c20cc7f7 Stable Diffusion: Input image downsampling (#276) LeonEricsson 2024-01-16 22:45:00 +01:00
  • 2ba5d3db14 Refactor activation function and loss calculation (#325) AtomicVar 2024-01-17 05:42:56 +08:00
  • ce7b65e8c4 Fix import order of normalizing_flow (#326) AtomicVar 2024-01-17 00:45:55 +08:00
  • 2287294723 fix mlx_lm generator for chinese (#321) someone 2024-01-16 23:13:33 +08:00
  • b0870ed679 fix response + bump version (#319) Awni Hannun 2024-01-15 11:51:21 -08:00
  • 195bec2fa3 feat(mlx_lm): add mixtral support in mlx_lm (#318) Anchen 2024-01-15 07:18:14 -08:00
  • 19b6167d81 Normalizing flow example (#133) Siddharth Mishra-Sharma 2024-01-13 19:58:48 -05:00
  • cd3cff0858 Phixtral (#290) Marcel Bischoff 2024-01-13 11:35:03 -05:00
  • a39b735c3b chore(mlx-lm): update phi2 model args to sync with hf config format. (#311) Anchen 2024-01-13 07:51:45 -08:00
  • 7575125d5d Added lora support for Phi-2 (#302) Yousif 2024-01-12 13:45:30 -08:00
  • 3ac731dd4f Fix TypeError in whisper benchmark script (#306) Alexandre Boucaud 2024-01-12 22:07:15 +01:00
  • ef93979973 Update model card uploaded with converted models (#309) Pedro Cuenca 2024-01-12 22:03:52 +01:00
  • 1fa40067fe Change tuple type definitions to use Tuple (#308) Angelos Katharopoulos 2024-01-12 11:15:09 -08:00
  • c1342b8e89 Use pip for mlx data with speech commands (#307) Awni Hannun 2024-01-12 11:06:33 -08:00
  • c6440416a2 Mlx llm package (#301) Awni Hannun 2024-01-12 10:25:56 -08:00
  • 2b61d9deb6 Updated CIFAR-10 ResNet example to use BatchNorm instead of LayerNorm (#257) Markus Enzweiler 2024-01-12 14:43:11 +01:00
  • 6217d7acd0 Delete llms/hf_llm/models/.gitignore (#300) Anchen 2024-01-11 16:56:50 -08:00
  • a2402116ae refactor(hf_llm): moving phi2 example into hf_llm (#293) Anchen 2024-01-11 12:29:12 -08:00
  • e74889d0fa prompt parameter (#291) Anjor Kanekar 2024-01-11 14:04:57 +00:00
  • 7380ebfb0d fix: undefined hf_path (#292) Anchen 2024-01-11 05:53:52 -08:00
  • 047d4650c4 Add -local flag to llms/hf_llm/convert.py for reading source HF models from filesystem. (#260) Konstantin Kerekovski 2024-01-10 22:53:01 -05:00
  • 80d18671ad [Lora] Fix generate (#282) Awni Hannun 2024-01-10 16:13:06 -08:00
  • a2bc8426f2 Update txt2image.py (#285) Rishi Narang 2024-01-11 01:31:59 +08:00
  • 2bbe9d3bd8 fix use of args in generate function (#284) Alwin Arrasyid 2024-01-10 23:09:21 +07:00
  • 44f86092ea Fix Tokenizer save error. (#278) Vaibhav Srivastav 2024-01-10 19:19:32 +05:30
  • 841c8f7b30 fix max tokens (#275) Awni Hannun 2024-01-09 21:41:12 -08:00
  • 7cfda327fd fix(lora): tokenizer return incompatible mx array (#271) Anchen 2024-01-09 19:46:38 -08:00
  • 7b258f33ac Move lora example to use the same model format / conversion as hf_llm (#252) Awni Hannun 2024-01-09 11:14:52 -08:00
  • bbd7172eef Some fixes / cleanup for BERT example (#269) Awni Hannun 2024-01-09 08:44:51 -08:00
  • 6759dfddf1 Fix SD image conversion (#266) Awni Hannun 2024-01-09 08:41:31 -08:00
  • 6e6eff326e fix: use of undefined args in generate function in phi-2 example (#265) Alwin Arrasyid 2024-01-09 21:43:59 +07:00
  • bb35e878cb [Whisper] Add load from Hub. (#255) Vaibhav Srivastav 2024-01-08 19:50:00 +05:30
  • d4c3a9cb54 [Whisper] Add HF Hub upload option. (#254) Vaibhav Srivastav 2024-01-08 19:48:24 +05:30
  • 6e5b0de4d3 refactor: make the phi2 example can be directly load the model from hf without convert needed (#253) Anchen 2024-01-08 06:01:23 -08:00
  • 9742ad0f51 Update README.md (#248) Nino Risteski 2024-01-08 05:13:58 +01:00
  • 485fb9ac0f quantize linear (#250) Awni Hannun 2024-01-07 18:48:59 -08:00
  • 737b4c81a3 Update README.md (#251) Ikko Eltociear Ashimine 2024-01-08 04:35:39 +09:00
  • bf9926489e [Whisper] Add word timestamps and confidence scores (#201) bofeng huang 2024-01-07 19:01:29 +01:00
  • 25ebd36112 Fix typo in lora convert.py (#245) mc0ps 2024-01-07 06:30:30 -05:00
  • b152d12d7b Update README.md (#243) Nino Risteski 2024-01-06 20:44:49 +01:00
  • 758f05c09a refactor: merge deepseek coder example into hf_llm example (#234) Anchen 2024-01-06 07:53:46 -08:00
  • cf0ad26a89 force fp16 for quantized models (#240) Awni Hannun 2024-01-05 21:29:15 -08:00
  • 37856f70a8 add numpy as a requirement to run lora.py (#238) Lawrence Wu 2024-01-05 16:16:28 -08:00
  • 37b41cec60 Qlora (#219) Awni Hannun 2024-01-04 21:05:59 -08:00
  • 4fa659acbd Handle receiving 0 tokens gracefully (#231) Christian Bieniak 2024-01-05 14:14:13 +11:00
  • 12c9bafbf5 Update README.md to fix --hf-model param call. (#229) Andy Peatling 2024-01-04 11:53:51 -08:00
  • e14afb3e77 fix to use actual prompt (#227) Awni Hannun 2024-01-04 11:12:05 -08:00
  • f95cf30a31 Fix upload to hub for HF LLMs conversion script. (#221) Vaibhav Srivastav 2024-01-04 19:36:05 +05:30
  • a5d6d0436c Support Hugging Face models (#215) Awni Hannun 2024-01-03 15:13:26 -08:00
  • 1d09c4fecd keep dtype on model conversion (#186) Daniel Strobusch 2024-01-02 20:20:29 +01:00
  • 85258b2be7 make parameter naming consistent with other examples. (#214) Daniel Strobusch 2024-01-02 17:18:12 +01:00
  • e632d7aaaa fix: deepseek coder tokenizer error (#211) Anchen 2024-01-02 01:10:37 +11:00
  • ee3c44d231 chore: make the Deepseek example compatible with Yi models. (#205) Anchen 2023-12-31 01:11:33 +11:00
  • 581a5733a1 [Whisper] Load customized MLX model & Quantization (#191) bofeng huang 2023-12-29 19:22:15 +01:00
  • 1cdbf9e886 chore: fix the load quantization model for deepseek coder (#203) Anchen 2023-12-30 00:25:38 +11:00
  • 31ddbd7806 add deepseek coder example (#172) Anchen 2023-12-29 16:42:22 +11:00
  • 37fd2464dc Add an image2image example in the stable diffusion (#198) Angelos Katharopoulos 2023-12-28 18:31:45 -08:00
  • 09566c7257 add speculative decoding example for llama (#149) Benjamin Anderson 2023-12-28 17:20:43 -06:00
  • 07c163d9d9 [Whisper] Large-v3 requires 128 Mel frequency bins (#193) Dimo 2023-12-28 22:50:35 +01:00