diff --git a/llms/mlx_lm/tuner/lora.py b/llms/mlx_lm/tuner/lora.py index adc1f8ca..d83b9025 100644 --- a/llms/mlx_lm/tuner/lora.py +++ b/llms/mlx_lm/tuner/lora.py @@ -10,7 +10,7 @@ class LoRALinear(nn.Module): linear: nn.Linear, r: int = 8, lora_alpha: float = 16, - lora_dropout: float = 0.05, + lora_dropout: float = 0.0, scale: float = 10.0, ): # TODO remove when input_dims and output_dims are attributes