Skip to content

Commit 1f768b8

Browse files
committed
fix formatting
1 parent 0eb0a48 commit 1f768b8

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

vllm/model_executor/layers/attention.py

+2-1
Original file line numberDiff line numberDiff line change
@@ -264,7 +264,8 @@ def __init__(
264264
self.is_neox_style = is_neox_style
265265

266266
# Create the cos and sin cache.
267-
inv_freq = 1.0 / (base**(torch.arange(0, rotary_dim, 2, dtype=torch.float32) / rotary_dim))
267+
inv_freq = 1.0 / (base**(
268+
torch.arange(0, rotary_dim, 2, dtype=torch.float32) / rotary_dim))
268269
t = torch.arange(max_position, dtype=torch.float32)
269270
freqs = torch.einsum("i,j -> ij", t, inv_freq)
270271
cos = freqs.cos()

0 commit comments

Comments
 (0)