Skip to content

Commit

Permalink
prefill only attention
Browse files Browse the repository at this point in the history
  • Loading branch information
noooop committed Oct 7, 2024
1 parent 748fdee commit 0333601
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion vllm/attention/layer.py
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,7 @@ def __init__(
dtype = torch.get_default_dtype()
self.attn_backend = get_attn_backend(
num_heads, head_size, num_kv_heads, sliding_window, dtype,
kv_cache_dtype, block_size, blocksparse_params is not None)
kv_cache_dtype, block_size, blocksparse_params is not None)()
else:
self.attn_backend = attn_backend

Expand Down

0 comments on commit 0333601

Please sign in to comment.