Files
mlx-examples/llms/mlx_lm/models
L fc93c55723 feat(mlx_lm): Nemotron (#949)
* feat: Nemotron

https://huggingface.co/nvidia/Minitron-4B-Base

This is basically Llama with partial RoPE and LayerNorm instead of
BatchNorm. Also they add 1 to the LayerNorm weight for some reason.

* fixup! feat: Nemotron

* nits

---------

Co-authored-by: Awni Hannun <awni@apple.com>
2024-08-29 21:08:57 -07:00
..
2024-01-12 10:25:56 -08:00
2024-08-16 15:28:39 -07:00
2024-08-23 13:18:51 -07:00
2024-08-17 07:18:09 -07:00
2024-08-16 15:28:39 -07:00
2024-08-23 13:18:51 -07:00
2024-08-29 21:08:57 -07:00
2024-08-16 15:28:39 -07:00
2024-08-23 13:18:51 -07:00
2024-08-16 15:28:39 -07:00
2024-08-24 06:52:33 -07:00
2024-08-16 15:28:39 -07:00
2024-08-24 06:52:33 -07:00