Skip to content

Commit c02167c

Browse files
heheda12345mzusman
authored andcommitted
[Kernel] Revert the API change of Attention.forward (vllm-project#12038)
Signed-off-by: Chen Zhang <zhangch99@outlook.com>
1 parent 16cf314 commit c02167c

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

vllm/attention/layer.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -134,8 +134,8 @@ def forward(
134134
query: torch.Tensor,
135135
key: torch.Tensor,
136136
value: torch.Tensor,
137-
_kv_cache: torch.Tensor,
138-
_attn_metadata: AttentionMetadata,
137+
kv_cache: torch.Tensor,
138+
attn_metadata: AttentionMetadata,
139139
) -> torch.Tensor:
140140
if self.use_output:
141141
output = torch.empty_like(query)

0 commit comments

Comments
 (0)