diff --git a/docs/src/python/nn/functions.rst b/docs/src/python/nn/functions.rst index f13cbe7b4..fd4302ef1 100644 --- a/docs/src/python/nn/functions.rst +++ b/docs/src/python/nn/functions.rst @@ -15,9 +15,9 @@ simple functions. gelu gelu_approx gelu_fast_approx - relu + mish prelu + relu + selu silu step - selu - mish diff --git a/docs/src/python/nn/layers.rst b/docs/src/python/nn/layers.rst index 4b2107446..fc8848c54 100644 --- a/docs/src/python/nn/layers.rst +++ b/docs/src/python/nn/layers.rst @@ -9,29 +9,29 @@ Layers :toctree: _autosummary :template: nn-module-template.rst - Sequential - ReLU - PReLU - GELU - SiLU - Step - SELU - Mish - Embedding - Linear - QuantizedLinear + ALiBi + BatchNorm Conv1d Conv2d - BatchNorm - LayerNorm - RMSNorm - GroupNorm - InstanceNorm Dropout Dropout2d Dropout3d - Transformer + Embedding + GELU + GroupNorm + InstanceNorm + LayerNorm + Linear + Mish MultiHeadAttention - ALiBi + PReLU + QuantizedLinear + RMSNorm + ReLU RoPE + SELU + Sequential + SiLU SinusoidalPositionalEncoding + Step + Transformer diff --git a/docs/src/python/nn/losses.rst b/docs/src/python/nn/losses.rst index 5a80ba947..31f40fb1f 100644 --- a/docs/src/python/nn/losses.rst +++ b/docs/src/python/nn/losses.rst @@ -10,14 +10,14 @@ Loss Functions :template: nn-module-template.rst binary_cross_entropy + cosine_similarity_loss cross_entropy + hinge_loss + huber_loss kl_div_loss l1_loss + log_cosh_loss mse_loss nll_loss smooth_l1_loss - triplet_loss - hinge_loss - huber_loss - log_cosh_loss - cosine_similarity_loss \ No newline at end of file + triplet_loss \ No newline at end of file diff --git a/docs/src/python/random.rst b/docs/src/python/random.rst index 8ac3eaa11..706378f9d 100644 --- a/docs/src/python/random.rst +++ b/docs/src/python/random.rst @@ -33,13 +33,13 @@ we use a splittable version of Threefry, which is a counter-based PRNG. .. autosummary:: :toctree: _autosummary - seed - key - split bernoulli categorical gumbel + key normal randint - uniform + seed + split truncated_normal + uniform