From f906cefc43102fa05a92bf54a24da03aa7d78052 Mon Sep 17 00:00:00 2001 From: Abhinav Goyal Date: Mon, 19 Aug 2024 10:51:38 +0530 Subject: [PATCH] formatting --- vllm/sequence.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/vllm/sequence.py b/vllm/sequence.py index 56a2b4141843f..f6c4a5a50ffc0 100644 --- a/vllm/sequence.py +++ b/vllm/sequence.py @@ -1192,7 +1192,7 @@ class HiddenStates(msgspec.Struct, array_like=True, # last proposed token is accepted (i.e., in case of bonus tokens). For the # case of no bonus tokens, these are ignored. second_last_token_hidden_states: Optional[torch.Tensor] = None - + _seq_ids: List[int] = msgspec.field(default_factory=list) def __post_init__(self): @@ -1227,8 +1227,8 @@ def prune(self, seq_group_metadata_list: List[SequenceGroupMetadata]) -> None: """Prune to provided list of sequence ids. Only used for decode steps. """ - # Currently this prunes all seq_ids not present in - # seq_group_metadata_list which might cause problems where a sequence + # Currently this prunes all seq_ids not present in + # seq_group_metadata_list which might cause problems where a sequence # may be "paused" then "resumed" later. This should only prune sequences # which are confirmed to be aborted. seq_ids = get_all_seq_ids(seq_group_metadata_list)