From 135c5818c1b2fbea5970b41a10172518dfa8a73a Mon Sep 17 00:00:00 2001 From: Alex Barron Date: Tue, 10 Dec 2024 11:26:04 -0800 Subject: [PATCH] Fix max_tokens (#1148) --- llms/mlx_lm/chat.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llms/mlx_lm/chat.py b/llms/mlx_lm/chat.py index 7795d8d7..5a8245ef 100644 --- a/llms/mlx_lm/chat.py +++ b/llms/mlx_lm/chat.py @@ -79,7 +79,7 @@ def main(): model, tokenizer, prompt, - args.max_tokens, + max_tokens=args.max_tokens, sampler=make_sampler(args.temp, args.top_p), prompt_cache=prompt_cache, ):