chore: fix the load quantization model

This commit is contained in:
Anchen
2023-12-29 23:01:20 +11:00
committed by GitHub
parent 31ddbd7806
commit a7da4ad7a3

View File

@@ -244,7 +244,7 @@ def load_model(model_path: str):
model = DeepseekCoder(model_args)
weights = mx.load(str(model_path / "weights.npz"))
if quantization := config.get("quantization", False):
if quantization:
nn.QuantizedLinear.quantize_module(model, **quantization)
model.update(tree_unflatten(list(weights.items())))