From 2d30f6787aa15b4da62acf2795dba8e734ed54e9 Mon Sep 17 00:00:00 2001 From: Prince Canuma Date: Wed, 12 Mar 2025 09:48:14 +0100 Subject: [PATCH] revert is sliding pattern --- llms/mlx_lm/models/gemma3_text.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/llms/mlx_lm/models/gemma3_text.py b/llms/mlx_lm/models/gemma3_text.py index 8843dc47..b0f76e16 100644 --- a/llms/mlx_lm/models/gemma3_text.py +++ b/llms/mlx_lm/models/gemma3_text.py @@ -60,7 +60,7 @@ class Attention(nn.Module): self.q_norm = RMSNorm(dims=head_dim, eps=args.rms_norm_eps) self.k_norm = RMSNorm(dims=head_dim, eps=args.rms_norm_eps) - self.is_sliding = (layer_idx + 1) % args.sliding_window_pattern == 0 + self.is_sliding = (layer_idx + 1) % args.sliding_window_pattern != 0 self.rope = nn.RoPE( head_dim, @@ -102,6 +102,7 @@ class Attention(nn.Module): if mask.shape[-1] != key_len: mask = mask[..., :key_len] + output = mx.fast.scaled_dot_product_attention( queries, keys, values, scale=self.scale, mask=mask )