mirror of
https://github.com/ml-explore/mlx-examples.git
synced 2025-09-27 09:48:07 +08:00
Phixtral (#290)
* initial * file * remove debug * Adding README * typo * simplify readme * nits in readmes --------- Co-authored-by: Marcel Bischoff <marcel.bischoff@awarehq.com> Co-authored-by: Awni Hannun <awni@apple.com>
This commit is contained in:
28
llms/phixtral/README.md
Normal file
28
llms/phixtral/README.md
Normal file
@@ -0,0 +1,28 @@
|
||||
# Phixtral
|
||||
|
||||
Phixtral is a Mixture of Experts (MoE) architecture inspired by
|
||||
[Mixtral](../mixtral/README.md) but made by combinding fine-tuned versions of
|
||||
Phi-2.[^1][^2]
|
||||
|
||||
### Setup
|
||||
|
||||
Install the dependencies:
|
||||
|
||||
```
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
### Run
|
||||
|
||||
```
|
||||
python generate.py \
|
||||
--model mlabonne/phixtral-4x2_8 \
|
||||
--prompt "write a quick sort in Python"
|
||||
```
|
||||
|
||||
Run `python generate.py --help` to see all the options.
|
||||
|
||||
[^1]: For more details on Phixtral, see the [Hugging Face repo](https://huggingface.co/mlabonne/phixtral-4x2_8).
|
||||
[^2]: For more details on Phi-2 see Microsoft's [blog post](
|
||||
https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/)
|
||||
and the [Hugging Face repo](https://huggingface.co/microsoft/phi-2).
|
Reference in New Issue
Block a user