Vaibhav Srivastav
|
aed14618ca
|
Add config.json to Mixtral. (#158)
* Add config.json to Mixtral.
* Update mixtral/mixtral.py
Co-authored-by: Pedro Cuenca <pedro@huggingface.co>
---------
Co-authored-by: Pedro Cuenca <pedro@huggingface.co>
|
2023-12-20 09:47:23 -08:00 |
|
Awni Hannun
|
ec11763527
|
fix RoPE bug + minor updates
|
2023-12-14 21:45:25 -08:00 |
|
Awni Hannun
|
b863e7cca0
|
format
|
2023-12-14 16:56:50 -08:00 |
|
Awni Hannun
|
e434e7e5c2
|
incude instruct option
|
2023-12-14 15:40:38 -08:00 |
|
Awni Hannun
|
078fed3d8d
|
use official HF for mixtral
|
2023-12-14 15:30:32 -08:00 |
|
jbax3
|
1505e49a62
|
Update README.md to fix git-lfs command
|
2023-12-13 15:51:27 -06:00 |
|
Awni Hannun
|
a614e951c4
|
Merge pull request #82 from ml-explore/llamav2
llama v2 with sharded weights
|
2023-12-12 17:08:24 -08:00 |
|
Awni Hannun
|
a99e9d551e
|
hf correction
|
2023-12-12 17:08:04 -08:00 |
|
Merrick Christensen
|
2206e8f7d9
|
Update convert.py
Docs are right, however, the code has a typo.
|
2023-12-12 14:33:33 -07:00 |
|
Awni Hannun
|
f0c57c1361
|
llama v2 with sharded weights
|
2023-12-12 12:48:15 -08:00 |
|
805karansaini
|
eae9431143
|
Typo Fix
|
2023-12-13 01:45:50 +05:30 |
|
Awni Hannun
|
034d0cfc2e
|
nit
|
2023-12-12 08:42:32 -08:00 |
|
Awni Hannun
|
0f66a12721
|
typos in readme
|
2023-12-12 08:41:28 -08:00 |
|
Awni Hannun
|
2ffd0da009
|
mixtral runs a bit faster
|
2023-12-12 08:36:40 -08:00 |
|
Awni Hannun
|
e42682dced
|
initial mixtral
|
2023-12-12 07:44:23 -08:00 |
|