docs update

This commit is contained in:
Awni Hannun
2024-08-10 09:24:35 -07:00
committed by CircleCI Docs
parent 8f68182d95
commit d8f7b8cda6
724 changed files with 14529 additions and 11046 deletions

View File

@@ -15,7 +15,7 @@ module to concisely define the model architecture.
Attention layer
^^^^^^^^^^^^^^^^
We will start with the llama attention layer which notably uses the RoPE
We will start with the Llama attention layer which notably uses the RoPE
positional encoding. [1]_ In addition, our attention layer will optionally use a
key/value cache that will be concatenated with the provided keys and values to
support efficient inference.