From 9284e50c67b37662640e6cbebd16640354f84a31 Mon Sep 17 00:00:00 2001 From: hschaeufler <9865991+hschaeufler@users.noreply.github.com> Date: Wed, 23 Oct 2024 22:20:35 +0300 Subject: [PATCH] Update lora_config.yaml with new param: num_layers --- llms/mlx_lm/examples/lora_config.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llms/mlx_lm/examples/lora_config.yaml b/llms/mlx_lm/examples/lora_config.yaml index 4ec9a23c..530272c7 100644 --- a/llms/mlx_lm/examples/lora_config.yaml +++ b/llms/mlx_lm/examples/lora_config.yaml @@ -14,7 +14,7 @@ data: "/path/to/training/data" seed: 0 # Number of layers to fine-tune -lora_layers: 16 +num_layers: 16 # Minibatch size. batch_size: 4