Skip to content

Commit 3132a93

Browse files
[Bugfix][Kernel] FA3 Fix - RuntimeError: This flash attention build only supports pack_gqa (for build size reasons). (#12405)
Signed-off-by: Lucas Wilkinson <[email protected]>
1 parent df5dafa commit 3132a93

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

CMakeLists.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -576,7 +576,7 @@ else()
576576
FetchContent_Declare(
577577
vllm-flash-attn
578578
GIT_REPOSITORY https://github.com/vllm-project/flash-attention.git
579-
GIT_TAG 0aff05f577e8a10086066a00618609199b25231d
579+
GIT_TAG 9732b0ce005d1e6216864788502d5570004678f5
580580
GIT_PROGRESS TRUE
581581
# Don't share the vllm-flash-attn build between build types
582582
BINARY_DIR ${CMAKE_BINARY_DIR}/vllm-flash-attn

0 commit comments

Comments
 (0)