From e14afb3e77d210388679c95c606fc40cee3982ff Mon Sep 17 00:00:00 2001 From: Awni Hannun Date: Thu, 4 Jan 2024 11:12:05 -0800 Subject: [PATCH] fix to use actual prompt (#227) --- llms/hf_llm/generate.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llms/hf_llm/generate.py b/llms/hf_llm/generate.py index e3b1136d..d0b41fe0 100644 --- a/llms/hf_llm/generate.py +++ b/llms/hf_llm/generate.py @@ -16,7 +16,7 @@ def generate( temp: float = 0.0, ): prompt = tokenizer( - args.prompt, + prompt, return_tensors="np", return_attention_mask=False, )[