mlx/benchmarks/python
2025-07-10 06:20:43 -07:00
..
blas
comparative [CUDA] Fix reductions (#2314) 2025-06-27 12:59:20 -07:00
batch_matmul_bench.py
compile_bench.py Add softmin, hardshrink, hardtanh (#1180) 2024-06-04 15:48:18 -07:00
conv1d_bench.py
conv2d_bench_cpu.py Conv cpu improvements (#1410) 2024-09-15 18:45:10 -07:00
conv2d_train_bench_cpu.py
conv2d_transpose_bench_cpu.py Conv cpu improvements (#1410) 2024-09-15 18:45:10 -07:00
conv3d_bench_cpu.py Conv cpu improvements (#1410) 2024-09-15 18:45:10 -07:00
conv3d_train_bench_cpu.py
conv3d_transpose_bench_cpu.py Conv cpu improvements (#1410) 2024-09-15 18:45:10 -07:00
conv_bench.py Add softmin, hardshrink, hardtanh (#1180) 2024-06-04 15:48:18 -07:00
conv_transpose_bench.py Transposed Convolution (#1245) 2024-09-06 19:52:38 -07:00
conv_unaligned_bench.py Add load_safe to the general conv loaders (#2258) 2025-06-10 20:58:16 -07:00
distributed_bench.py MPI ops in GPU stream for faster comms (#1356) 2024-08-26 15:12:50 -07:00
einsum_bench.py Einsum (#1269) 2024-07-25 09:36:44 -07:00
fft_bench.py Feature complete Metal FFT (#1102) 2024-06-06 12:57:25 -07:00
gather_bench.py Remove unused modules (#1949) 2025-03-10 06:05:26 -07:00
gather_mm_bench.py Gather qmm batched kernel and refactoring of quantized (#2078) 2025-04-17 13:53:11 -07:00
gather_qmm_bench.py Gather qmm batched kernel and refactoring of quantized (#2078) 2025-04-17 13:53:11 -07:00
hadamard_bench.py
layer_norm_bench.py
rms_norm_bench.py
rope_bench.py
scatter_bench.py improvements to scatter / gather (#1541) 2024-10-30 19:30:54 -07:00
sdpa_bench.py Support fused masking in Attention (#1924) 2025-03-20 11:01:32 -07:00
sdpa_vector_bench.py Allow different value dimensions in sdpa_vector (#1811) 2025-01-31 20:58:59 -08:00
single_ops.py Align mlx::core::min op nan propagation with NumPy (#2346) 2025-07-10 06:20:43 -07:00
synchronize_bench.py Faster synchronization Fence primitive (#1773) 2025-01-17 18:42:19 -08:00
time_utils.py