Quantize example (#162)

* testing quantization

* conversion + quantization working

* one config processor

* quantization in mistral / nits in llama

* args for quantization

* llama / mistral conversion in good shape

* phi2 quantized

* mixtral

* qwen conversion
This commit is contained in:
Awni Hannun
2023-12-21 12:59:37 -08:00
committed by GitHub
parent 4c9db80ed2
commit 3cf436b529
17 changed files with 553 additions and 126 deletions

View File

@@ -23,10 +23,17 @@ tar -xf mistral-7B-v0.1.tar
Then, convert the weights with:
```
python convert.py
python convert.py --torch-path <path_to_torch>
```
The conversion script will save the converted weights in the same location.
To generate a 4-bit quantized model, use ``-q``. For a full list of options:
```
python convert.py --help
```
By default, the conversion script will make the directory `mlx_model` and save
the converted `weights.npz`, `tokenizer.model`, and `config.json` there.
> [!TIP]
> Alternatively, you can also download a few converted checkpoints from the
@@ -40,7 +47,7 @@ Once you've converted the weights to MLX format, you can generate text with
the Mistral model:
```
python mistral.py --prompt "It is a truth universally acknowledged," --temp 0
python mistral.py --prompt "It is a truth universally acknowledged,"
```
Run `python mistral.py --help` for more details.

View File

@@ -1,32 +1,98 @@
# Copyright © 2023 Apple Inc.
import argparse
import copy
import json
import shutil
from pathlib import Path
import mlx.core as mx
import mlx.nn as nn
import numpy as np
import torch
from mistral import Mistral, ModelArgs
from mlx.utils import tree_flatten, tree_map, tree_unflatten
def quantize(weights, config, args):
quantized_config = copy.deepcopy(config)
# Load the model:
config.pop("sliding_window", None)
model = Mistral(ModelArgs(**config))
weights = tree_map(mx.array, weights)
model.update(tree_unflatten(list(weights.items())))
# Quantize the model:
nn.QuantizedLinear.quantize_module(model, args.q_group_size, args.q_bits)
# Update the config:
quantized_config["quantization"] = {
"group_size": args.q_group_size,
"bits": args.q_bits,
}
quantized_weights = dict(tree_flatten(model.parameters()))
return quantized_weights, quantized_config
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert Mistral weights to MLX.")
parser.add_argument(
"--model-path",
"--torch-path",
type=str,
default="mistral-7B-v0.1/",
help="The path to the Mistral model. The MLX weights will also be saved there.",
default="mistral-7B-v0.1",
help="The path to the PyTorch model.",
)
parser.add_argument(
"--mlx-path",
type=str,
default="mlx_model",
help="The path to save the MLX model.",
)
parser.add_argument(
"-q",
"--quantize",
help="Generate a quantized model.",
action="store_true",
)
parser.add_argument(
"--q_group_size",
help="Group size for quantization.",
type=int,
default=64,
)
parser.add_argument(
"--q_bits",
help="Bits per weight for quantization.",
type=int,
default=4,
)
args = parser.parse_args()
model_path = Path(args.model_path)
state = torch.load(str(model_path / "consolidated.00.pth"))
np.savez(
str(model_path / "weights.npz"),
**{k: v.to(torch.float16).numpy() for k, v in state.items()}
torch_path = Path(args.torch_path)
state = torch.load(str(torch_path / "consolidated.00.pth"))
mlx_path = Path(args.mlx_path)
mlx_path.mkdir(parents=True, exist_ok=True)
weights = {k: v.to(torch.float16).numpy() for k, v in state.items()}
with open(torch_path / "params.json", "r") as f:
config = json.loads(f.read())
if args.quantize:
print("[INFO] Quantizing")
weights, config = quantize(weights, config, args)
# Save weights
np.savez(str(mlx_path / "weights.npz"), **weights)
# Copy tokenizer
shutil.copyfile(
str(torch_path / "tokenizer.model"),
str(mlx_path / "tokenizer.model"),
)
# Save config.json with model_type
with open(model_path / "params.json", "r") as f:
config = json.loads(f.read())
with open(mlx_path / "config.json", "w") as f:
config["model_type"] = "mistral"
with open(model_path / "config.json", "w") as f:
json.dump(config, f, indent=4)

View File

@@ -8,7 +8,7 @@ from typing import List, Optional, Tuple
import mlx.core as mx
import mlx.nn as nn
from mlx.utils import tree_map, tree_unflatten
from mlx.utils import tree_unflatten
from sentencepiece import SentencePieceProcessor
@@ -189,18 +189,20 @@ class Tokenizer:
return out
def load_model(folder: str, dtype=mx.float16):
def load_model(folder: str):
model_path = Path(folder)
tokenizer = Tokenizer(str(model_path / "tokenizer.model"))
with open(model_path / "config.json", "r") as f:
config = json.loads(f.read())
config.pop("sliding_window", None)
config.pop("model_type", None)
quantization = config.pop("quantization", None)
model_args = ModelArgs(**config)
weights = mx.load(str(model_path / "weights.npz"))
weights = tree_unflatten(list(weights.items()))
weights = tree_map(lambda p: p.astype(dtype), weights)
model = Mistral(model_args)
if quantization is not None:
nn.QuantizedLinear.quantize_module(model, **quantization)
model.update(weights)
return model, tokenizer
@@ -227,7 +229,7 @@ if __name__ == "__main__":
parser.add_argument(
"--model-path",
type=str,
default="mistral-7B-v0.1",
default="mlx_model",
help="The path to the model weights and tokenizer",
)
parser.add_argument(
@@ -236,7 +238,7 @@ if __name__ == "__main__":
default="In the beginning the Universe was created.",
)
parser.add_argument(
"--max_tokens",
"--max-tokens",
"-m",
type=int,
default=100,
@@ -246,7 +248,7 @@ if __name__ == "__main__":
"--temp",
help="The sampling temperature.",
type=float,
default=1.0,
default=0.0,
)
parser.add_argument(
"--tokens_per_eval",