chore: fix the load quantization model for deepseek coder (#203)

* chore: fix the load quantization model

* change to explicitly check for quantization config
This commit is contained in:
Anchen 2023-12-30 00:25:38 +11:00 committed by GitHub
parent 31ddbd7806
commit 1cdbf9e886
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -244,7 +244,7 @@ def load_model(model_path: str):
model = DeepseekCoder(model_args)
weights = mx.load(str(model_path / "weights.npz"))
if quantization := config.get("quantization", False):
if quantization is not None:
nn.QuantizedLinear.quantize_module(model, **quantization)
model.update(tree_unflatten(list(weights.items())))