diff --git a/llama/llama.py b/llama/llama.py index db9c8db3..9b8157b7 100644 --- a/llama/llama.py +++ b/llama/llama.py @@ -315,7 +315,7 @@ def load_model(model_path): config["hidden_dim"] = weights["layers.0.feed_forward.w1.weight"].shape[0] if config.get("vocab_size", -1) < 0: config["vocab_size"] = weights["output.weight"].shape[-1] - unused = ["multiple_of", "ffn_dim_multiplie"] + unused = ["multiple_of", "ffn_dim_multiplier", 'rope_theta'] for k in unused: if k in config: config.pop(k)