From 726b1ddec0d503a42267dd1ebf61a1b39a76d16f Mon Sep 17 00:00:00 2001 From: Madroid Ma Date: Fri, 16 Feb 2024 22:03:33 +0800 Subject: [PATCH] fix: check LoRA layers number error (#446) --- llms/mlx_lm/tuner/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llms/mlx_lm/tuner/utils.py b/llms/mlx_lm/tuner/utils.py index 6756bfca..2059ebc4 100644 --- a/llms/mlx_lm/tuner/utils.py +++ b/llms/mlx_lm/tuner/utils.py @@ -21,7 +21,7 @@ def linear_to_lora_layers(model: nn.Module, num_lora_layers: int): if num_lora_layers > num_model: raise ValueError( f"Requested {num_lora_layers} LoRA layers " - f"but the model only has {num_model_layers} layers." + f"but the model only has {num_model} layers." ) if model.model_type in [