diff --git a/mlx/backend/cuda/matmul.cpp b/mlx/backend/cuda/matmul.cpp index 5b86961da..c32cecc03 100644 --- a/mlx/backend/cuda/matmul.cpp +++ b/mlx/backend/cuda/matmul.cpp @@ -162,14 +162,14 @@ class MatMul { } } - void *workspace_ptr = nullptr; + void* workspace_ptr = nullptr; if (heuristic_.workspaceSize > 0) { - array workspace( - allocator::malloc(heuristic_.workspaceSize), - {static_cast(heuristic_.workspaceSize)}, - int8); - encoder.add_temporary(workspace); - workspace_ptr = workspace.data(); + array workspace( + allocator::malloc(heuristic_.workspaceSize), + {static_cast(heuristic_.workspaceSize)}, + int8); + encoder.add_temporary(workspace); + workspace_ptr = workspace.data(); } encoder.launch_kernel([&](cudaStream_t stream) { @@ -464,7 +464,14 @@ void AddMM::eval_gpu(const std::vector& inputs, array& out) { auto nbatch = batch_count / batch_shape.back(); if (nbatch == 1) { - matmul.run(encoder, out.data(), a.data(), b.data(), c.data(), alpha_, beta_); + matmul.run( + encoder, + out.data(), + a.data(), + b.data(), + c.data(), + alpha_, + beta_); return; }