Fix cache.py to support non-top level layers

This commit is contained in:
Shunta Saito 2025-02-13 13:44:31 +09:00
parent 58686bbcac
commit 9a6e6541de

View File

@ -26,7 +26,10 @@ def make_prompt_cache(
if hasattr(model, "make_cache"):
return model.make_cache()
num_layers = len(model.layers)
if hasattr(model, "layers"):
num_layers = len(model.layers)
else:
num_layers = len(model.model.layers)
if max_kv_size is not None:
return [
RotatingKVCache(max_size=max_kv_size, keep=4) for _ in range(num_layers)