mirror of
				https://github.com/ml-explore/mlx.git
				synced 2025-10-29 06:53:18 +08:00 
			
		
		
		
	fix links (#32)
This commit is contained in:
		| @@ -321,7 +321,7 @@ which can then be used to update the model. Note that the method above incurs | ||||
| several unnecessary copies from disk to numpy and then from numpy to MLX. It | ||||
| will be replaced in the future with direct loading to MLX. | ||||
|  | ||||
| You can download the full example code in `mlx-examples <code>`_. Assuming, the | ||||
| You can download the full example code in `mlx-examples`_. Assuming, the | ||||
| existence of ``weights.pth`` and ``tokenizer.model`` in the current working | ||||
| directory we can play around with our inference script as follows (the timings | ||||
| are representative of an M1 Ultra and the 7B parameter Llama model): | ||||
| @@ -369,9 +369,9 @@ Scripts | ||||
|  | ||||
| .. admonition:: Download the code | ||||
|  | ||||
|    The full example code is available in `mlx-examples <code>`_. | ||||
|    The full example code is available in `mlx-examples`_. | ||||
|  | ||||
| .. code: `https://github.com/ml-explore/mlx-examples/tree/main/llama`_ | ||||
| .. _mlx-examples: https://github.com/ml-explore/mlx-examples/tree/main/llama | ||||
|  | ||||
| .. [1] Su, J., Lu, Y., Pan, S., Murtadha, A., Wen, B. and Liu, Y., 2021. | ||||
|    Roformer: Enhanced transformer with rotary position embedding. arXiv | ||||
|   | ||||
| @@ -127,5 +127,5 @@ Finally, we put it all together by instantiating the model, the | ||||
|   This should not be confused with :func:`mlx.core.value_and_grad`. | ||||
|  | ||||
| The model should train to a decent accuracy (about 95%) after just a few passes | ||||
| over the training set. The `full example <https://github.com/ml-explore/mlx-examples/tree/main/mlp>`_ | ||||
| over the training set. The `full example <https://github.com/ml-explore/mlx-examples/tree/main/mnist>`_ | ||||
| is available in the MLX GitHub repo. | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Awni Hannun
					Awni Hannun