Skip to content

Commit e562eec

Browse files
CUDA: fix typo in FlashAttention code (#13926)
1 parent b47ab7b commit e562eec

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

ggml/src/ggml-cuda/fattn-mma-f16.cuh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1246,7 +1246,7 @@ static __global__ void flash_attn_ext_f16(
12461246
NO_DEVICE_CODE;
12471247
return;
12481248
}
1249-
#endif __CUDA_ARCH__ == GGML_CUDA_CC_TURING
1249+
#endif // __CUDA_ARCH__ == GGML_CUDA_CC_TURING
12501250

12511251
static_assert(!mla || DKQ >= DV, "MLA needs DKQ >= DV");
12521252

0 commit comments

Comments
 (0)