From f691e00e5a5d84c196abd093cb6f8db88ccff6df Mon Sep 17 00:00:00 2001 From: Burak Budanur Date: Thu, 14 Dec 2023 14:02:11 +0100 Subject: [PATCH] Corrected the typo in 'ffn_dim_multiplier' in and added 'rope_theta' to the list unused. Without these, llama examples did not run. --- llama/llama.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama/llama.py b/llama/llama.py index db9c8db3..9b8157b7 100644 --- a/llama/llama.py +++ b/llama/llama.py @@ -315,7 +315,7 @@ def load_model(model_path): config["hidden_dim"] = weights["layers.0.feed_forward.w1.weight"].shape[0] if config.get("vocab_size", -1) < 0: config["vocab_size"] = weights["output.weight"].shape[-1] - unused = ["multiple_of", "ffn_dim_multiplie"] + unused = ["multiple_of", "ffn_dim_multiplier", 'rope_theta'] for k in unused: if k in config: config.pop(k)