Skip to content

Commit f4b4a35

Browse files
qixiang-99symphonylyh
authored andcommitted
fix: fix rebase format issue
Signed-off-by: Qixiang Lin <[email protected]>
1 parent 829db20 commit f4b4a35

File tree

2 files changed

+4
-2
lines changed

2 files changed

+4
-2
lines changed

tensorrt_llm/_torch/attention_backend/trtllm.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -706,7 +706,8 @@ def forward(
706706
use_paged_context_fmha = (
707707
metadata.runtime_features.chunked_prefill
708708
or metadata.runtime_features.cache_reuse
709-
or metadata.runtime_features.has_speculative_draft_tokens) if metadata.runtime_features else False
709+
or metadata.runtime_features.has_speculative_draft_tokens
710+
) if metadata.runtime_features else False
710711

711712
if use_paged_context_fmha and self.has_fp8_kv_cache:
712713
# NOTE: W4A8_AWQ can be included too, exclude for now since

tensorrt_llm/_torch/pyexecutor/model_engine.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -556,7 +556,8 @@ def no_cuda_graph():
556556
resource_manager=resource_manager)
557557
torch.cuda.synchronize()
558558

559-
def _set_up_attn_metadata(self, kv_cache_manager: KVCacheManager,
559+
def _set_up_attn_metadata(self,
560+
kv_cache_manager: KVCacheManager,
560561
is_dummy_forward: bool = False):
561562
# is_dummy_forward is used to indicate whether the forward is
562563
# a dummy forward for memory estimation OR

0 commit comments

Comments
 (0)