Skip to content

Commit a164275

Browse files
committed
Update test_flash_attn.py
fix flash_attn_fp8 test Signed-off-by: ShuaibinLi <[email protected]>
1 parent 14288d1 commit a164275

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

tests/kernels/attention/test_flash_attn.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -145,7 +145,7 @@ def test_flash_attn_with_paged_kv(
145145
v_descale = None
146146
if q_dtype is not None:
147147
# QKV are drawn from N(0, 1): no need for a fp8 scaling factor
148-
maybe_quantized_query = query.to(q_dtype)
148+
maybe_quantized_query = q.to(q_dtype)
149149
maybe_quantized_key_cache = key_cache.to(q_dtype)
150150
maybe_quantized_value_cache = value_cache.to(q_dtype)
151151

0 commit comments

Comments
 (0)