mlx/cmake
Jagrit Digani 89d327075f
Enabling fused attention for head dim 128 (#1899)
* Share KV smem

* Fix bfloat error

* Unroll O = S @ V loop

* Perf upgrade

* Remove commented out function

* Add -Wno-c++17-extensions flag to metal flags

* Add -Wno-c++17-extensions flag to metal extension flags
2025-02-26 10:02:06 -08:00
..
extension.cmake Enabling fused attention for head dim 128 (#1899) 2025-02-26 10:02:06 -08:00