From df1f8aa3be981cce7f6a873152720ccd439f7ef6 Mon Sep 17 00:00:00 2001 From: Jyun1998 Date: Fri, 15 Dec 2023 03:27:33 +0800 Subject: [PATCH] precommit changes --- python/mlx/nn/layers/transformer.py | 1 + 1 file changed, 1 insertion(+) diff --git a/python/mlx/nn/layers/transformer.py b/python/mlx/nn/layers/transformer.py index 971a2ad56..9b70221ff 100644 --- a/python/mlx/nn/layers/transformer.py +++ b/python/mlx/nn/layers/transformer.py @@ -9,6 +9,7 @@ from mlx.nn.layers.dropout import Dropout from mlx.nn.layers.linear import Linear from mlx.nn.layers.normalization import LayerNorm + class MultiHeadAttention(Module): """Implements the scaled dot product attention with multiple heads.