From 7fbca214b142dbedb73d696329f154d6683e23fb Mon Sep 17 00:00:00 2001 From: Ivan Fioravanti Date: Sun, 4 Feb 2024 21:28:21 +0100 Subject: [PATCH] Add max sequence length argument in lora.py (#408) A new argument "--max_seq_length" has been added to the command-line parser and passed as a parameter to the main function of the lora.py script. This allows users to specify and control the maximum sequence length during training. --- llms/mlx_lm/lora.py | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/llms/mlx_lm/lora.py b/llms/mlx_lm/lora.py index f5418bbc..29fdcdcb 100644 --- a/llms/mlx_lm/lora.py +++ b/llms/mlx_lm/lora.py @@ -109,6 +109,12 @@ def build_parser(): default=500, help="Number of test set batches, -1 uses the entire test set.", ) + parser.add_argument( + "--max_seq_length", + type=int, + default=2048, + help="Maximum sequence length.", + ) parser.add_argument("--seed", type=int, default=0, help="The PRNG seed") return parser @@ -197,6 +203,7 @@ if __name__ == "__main__": steps_per_eval=args.steps_per_eval, steps_per_save=args.save_every, adapter_file=args.adapter_file, + max_seq_length=args.max_seq_length ) if args.train: print("Training")