From ff36712effb97e272f47800c4bfaa56a955c5c41 Mon Sep 17 00:00:00 2001 From: Aleksandr Malyshev Date: Tue, 22 Oct 2024 18:12:35 -0500 Subject: [PATCH] after merge correction --- vllm/attention/backends/rocm_flash_attn.py | 5 ----- 1 file changed, 5 deletions(-) diff --git a/vllm/attention/backends/rocm_flash_attn.py b/vllm/attention/backends/rocm_flash_attn.py index 756202a45e3ef..677e3f6bb029c 100644 --- a/vllm/attention/backends/rocm_flash_attn.py +++ b/vllm/attention/backends/rocm_flash_attn.py @@ -621,11 +621,6 @@ def forward( value = value[:num_prefill_tokens] if prefill_meta := attn_metadata.prefill_metadata: - (query_seq_start_loc, query_max_seq_len, key_seq_start_loc, - key_max_seq_len, seq_lens, - causal_mask) = _get_seq_len_block_table_args( - prefill_meta, attn_type) - # Prompt run. # normal attention and DECODER if attn_type == AttentionType.DECODER and (