From 06a9f5d106bfb5e2ac0f0f05732414d347b56938 Mon Sep 17 00:00:00 2001 From: Goekdeniz-Guelmez Date: Sun, 19 Jan 2025 00:53:41 +0100 Subject: [PATCH] update lora_config.yaml --- llms/mlx_lm/examples/lora_config.yaml | 13 +++++++++++++ 1 file changed, 13 insertions(+) diff --git a/llms/mlx_lm/examples/lora_config.yaml b/llms/mlx_lm/examples/lora_config.yaml index 530272c7..88caedd6 100644 --- a/llms/mlx_lm/examples/lora_config.yaml +++ b/llms/mlx_lm/examples/lora_config.yaml @@ -7,6 +7,19 @@ train: true # The fine-tuning method: "lora", "dora", or "full". fine_tune_type: lora +# The training-mode: "normal", or "dpo" +training_mode: normal + +# If you set training_mode to "dpo" +# beta: 0.1 +# The dpo-lodd-type: "sigmoid", "hinge", "ipo", or "dpop" +# dpo_loss_type: "sigmoid" +# is_reference_free: False +# delta: 50.0 +# If reference_model_path is not given it will just use the same model +# reference_model_path: "mlx_model" +# train_bias_only: False + # Directory with {train, valid, test}.jsonl files data: "/path/to/training/data"