mlx-examples/llms/mlx_lm/models
Prince Canuma b044ce2acf
Add support for ibm granite (#758)
* add support for granite 3-8B config

* add gpt_bigcode

* add positional embedding condition.

* add support for granite 3-8B config

* add gpt_bigcode

* add positional embedding condition.

* remove unused function

* rebase fix

* move position emebedding to mask creation

* add to tuner and format

* add support for granite 3-8B config

* add gpt_bigcode

* add positional embedding condition.

* add support for granite 3-8B config

* add gpt_bigcode

* add positional embedding condition.

* rebase fix

* move position emebedding to mask creation

* add to tuner and format

* refactor mask

* remove dropout layers
2024-05-21 20:16:31 -07:00
..
__init__.py Mlx llm package (#301) 2024-01-12 10:25:56 -08:00
base.py Add support for ibm granite (#758) 2024-05-21 20:16:31 -07:00
cohere.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
dbrx.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
gemma.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
gpt_bigcode.py Add support for ibm granite (#758) 2024-05-21 20:16:31 -07:00
llama.py Add support for ibm granite (#758) 2024-05-21 20:16:31 -07:00
minicpm.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
mixtral.py Block sparse MM MoEs (#782) 2024-05-21 15:58:08 -07:00
olmo.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
openelm.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
phi3.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
phi.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
phixtral.py Block sparse MM MoEs (#782) 2024-05-21 15:58:08 -07:00
plamo.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
qwen2_moe.py Block sparse MM MoEs (#782) 2024-05-21 15:58:08 -07:00
qwen2.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
qwen.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
stablelm.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
starcoder2.py Kv cache (#643) 2024-05-08 08:18:13 -07:00
switch_layers.py Block sparse MM MoEs (#782) 2024-05-21 15:58:08 -07:00