Skip to content

Commit 67df989

Browse files
heheda12345mzusman
authored andcommitted
[Bugfix][Kernel] Give unique name to BlockSparseFlashAttention (vllm-project#12040)
Signed-off-by: Chen Zhang <[email protected]>
1 parent 0b7ba44 commit 67df989

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

vllm/attention/backends/blocksparse_attn.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -89,8 +89,7 @@ class BlocksparseFlashAttentionBackend(AttentionBackend):
8989

9090
@staticmethod
9191
def get_name() -> str:
92-
# For attention layer compatibility
93-
return "FLASH_ATTN"
92+
return "BLOCK_SPARSE_FLASH_ATTN"
9493

9594
@staticmethod
9695
def get_impl_cls() -> Type["BlocksparseFlashAttentionImpl"]:

vllm/platforms/interface.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,7 @@ class _Backend(enum.Enum):
3333
HPU_ATTN = enum.auto()
3434
PALLAS = enum.auto()
3535
IPEX = enum.auto()
36+
BLOCK_SPARSE_FLASH_ATTN = enum.auto()
3637
NO_ATTENTION = enum.auto()
3738

3839

0 commit comments

Comments
 (0)