mirror of
https://github.com/ml-explore/mlx.git
synced 2025-07-15 21:21:16 +08:00
![]() * Make max op NaN propagation rules align with numpy * Adding benchmarks and testing for max op nanpropagation * Pre-commit formatting * Fix max complex64 nan propagation and add test * Improve the cpp unittest * Only check nans on non-integral types in simd_reduce_impl. * Cleanup using namespace alias * Add cpu Max nanpropagation. Fix a small fib in cpu max dispatch data types for int8/int16. * Make the max nanpropagation test more meaningful for integer types * Remove tuple unpacking syntax to comply with earlier python versions. Add cuda skip to nanpropagation tests, fix cuda implementation in a separate PR. |
||
---|---|---|
.. | ||
blas | ||
comparative | ||
batch_matmul_bench.py | ||
compile_bench.py | ||
conv1d_bench.py | ||
conv2d_bench_cpu.py | ||
conv2d_train_bench_cpu.py | ||
conv2d_transpose_bench_cpu.py | ||
conv3d_bench_cpu.py | ||
conv3d_train_bench_cpu.py | ||
conv3d_transpose_bench_cpu.py | ||
conv_bench.py | ||
conv_transpose_bench.py | ||
conv_unaligned_bench.py | ||
distributed_bench.py | ||
einsum_bench.py | ||
fft_bench.py | ||
gather_bench.py | ||
gather_mm_bench.py | ||
gather_qmm_bench.py | ||
hadamard_bench.py | ||
layer_norm_bench.py | ||
rms_norm_bench.py | ||
rope_bench.py | ||
scatter_bench.py | ||
sdpa_bench.py | ||
sdpa_vector_bench.py | ||
single_ops.py | ||
synchronize_bench.py | ||
time_utils.py |