From 1e9b2b9a01638569a5e5b1c54a3c613f9829824a Mon Sep 17 00:00:00 2001 From: kevin Date: Tue, 21 Jan 2025 20:24:52 +0000 Subject: [PATCH] mypy fix Signed-off-by: kevin --- vllm/v1/stats/common.py | 24 +++++++++++++----------- 1 file changed, 13 insertions(+), 11 deletions(-) diff --git a/vllm/v1/stats/common.py b/vllm/v1/stats/common.py index 099d82c5904cf..500bc356fc179 100644 --- a/vllm/v1/stats/common.py +++ b/vllm/v1/stats/common.py @@ -10,10 +10,11 @@ from vllm.sampling_params import SamplingParams -class RequestStatsUpdate(msgspec.Struct, - array_like=True, - omit_defaults=True, - gc=False): +class RequestStatsUpdate( + msgspec.Struct, # type: ignore + array_like=True, + omit_defaults=True, + gc=False): """ An update to the request stats. @@ -341,8 +342,8 @@ def update_from(self, update: "RequestStatsUpdate"): self.queued_ts_s = ts elif update.type == RequestStatsUpdate.Type.PREFILLING: self.prefill_start_ts_s_lst.append(ts) - self.num_cached_tokens = update.num_cached_tokens - self.num_computed_tokens = update.num_computed_tokens + self.num_cached_tokens = update.num_cached_tokens or 0 + self.num_computed_tokens = update.num_computed_tokens or 0 elif update.type == RequestStatsUpdate.Type.PREEMPTED: self._reset_for_preemption(ts) elif update.type == RequestStatsUpdate.Type.DECODING: @@ -350,7 +351,7 @@ def update_from(self, update: "RequestStatsUpdate"): elif update.type == RequestStatsUpdate.Type.DETOKENIZED: self._record_detokenized_output( ts, - update.num_new_tokens, + update.num_new_tokens or 0, ) elif update.type == RequestStatsUpdate.Type.FINISHED: self.finished_ts_s = ts @@ -425,10 +426,11 @@ class EngineCoreProcessStats: output_queue_size: Optional[int] = None -class EngineCoreStatsSnapshot(msgspec.Struct, - array_like=True, - omit_defaults=True, - gc=False): +class EngineCoreStatsSnapshot( + msgspec.Struct, # type: ignore + array_like=True, + omit_defaults=True, + gc=False): """ A snapshot of the EngineCore's current stats over a period of time. """