From 9e28bedc8176be25a44e9295b8312d1a082a340b Mon Sep 17 00:00:00 2001 From: Matt Clayton Date: Mon, 10 Feb 2025 12:32:00 -0500 Subject: [PATCH] Fix comment --- llms/mlx_lm/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llms/mlx_lm/utils.py b/llms/mlx_lm/utils.py index 2cf87194..64813123 100644 --- a/llms/mlx_lm/utils.py +++ b/llms/mlx_lm/utils.py @@ -76,7 +76,7 @@ class GenerationResponse: Args: text (str): The next segment of decoded text. This can be an empty string. token (int): The next token. - from_draft (bool): Whether the token was generated by a draft model. + from_draft (bool): Whether the token was generated by the draft model. logprobs (mx.array): A vector of log probabilities. prompt_tokens (int): The number of tokens in the prompt. prompt_tps (float): The prompt processing tokens-per-second.