mlx-examples/llms/tests
Prince Canuma 2fce02acd8
Add support for Gemma3 (#1336)
* add support for gemma3

* fix model loading

* revert rmsnorm

* revert is sliding pattern

* revert

* add tests

* formatting

* Update llms/mlx_lm/models/gemma3_text.py

Co-authored-by: Awni Hannun <awni.hannun@gmail.com>

* Update llms/mlx_lm/models/gemma3_text.py

Co-authored-by: Awni Hannun <awni.hannun@gmail.com>

* Update llms/mlx_lm/models/gemma3_text.py

Co-authored-by: Awni Hannun <awni.hannun@gmail.com>

* Update llms/mlx_lm/models/gemma3_text.py

Co-authored-by: Awni Hannun <awni.hannun@gmail.com>

* Update llms/mlx_lm/models/gemma3_text.py

Co-authored-by: Awni Hannun <awni.hannun@gmail.com>

* Update llms/mlx_lm/models/gemma3_text.py

Co-authored-by: Awni Hannun <awni.hannun@gmail.com>

* Update llms/mlx_lm/models/gemma3_text.py

Co-authored-by: Awni Hannun <awni.hannun@gmail.com>

* fix sliding window mask

---------

Co-authored-by: Awni Hannun <awni.hannun@gmail.com>
Co-authored-by: Awni Hannun <awni@apple.com>
2025-03-13 08:14:25 -07:00
..
test_datsets.py Completion only fine-tuning of instruction models with collections of HF datasets (#1103) 2025-02-09 20:12:34 -08:00
test_finetune.py reduction moved to CPU in case of distributed training (#1200) 2025-01-14 17:20:42 -08:00
test_generate.py Add "from_draft" to GenerationResponse (#1272) 2025-02-11 15:41:02 -08:00
test_gguf.py fix(mlx-lm): type hints in gguf.py (#621) 2024-03-26 07:56:01 -07:00
test_models.py Add support for Gemma3 (#1336) 2025-03-13 08:14:25 -07:00
test_prompt_cache.py fix flaky test (#1322) 2025-03-05 14:00:09 -08:00
test_sample_utils.py make_sampler creates sampler chain with all sampling parameters (#1330) 2025-03-11 13:37:35 -07:00
test_server.py chore(mlx-lm): support text type content in messages (#1225) 2025-01-27 17:13:50 -08:00
test_tokenizers.py Fix decoding manually added tokens (#1164) 2024-12-17 09:54:29 -08:00
test_tuner_utils.py LoRA: Extract small function (#614) 2024-06-02 06:38:42 -07:00
test_utils_load_model.py deepseek v3 model with pipeline parallelism (#1191) 2025-01-09 15:55:53 -08:00
test_utils.py Fix whipser conversion for safetensors models (#935) 2024-08-14 10:22:04 -07:00