From d1dd5f53e4bdb78d5e3b7b4dc1c03f6e2a991e1e Mon Sep 17 00:00:00 2001 From: Roy Wang Date: Fri, 7 Nov 2025 00:25:39 +0800 Subject: [PATCH] [Frontend] Fix logging format when enable response logging (#28049) Signed-off-by: esmeetu --- vllm/entrypoints/openai/api_server.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/vllm/entrypoints/openai/api_server.py b/vllm/entrypoints/openai/api_server.py index 11a0c8a07..aba961ac6 100644 --- a/vllm/entrypoints/openai/api_server.py +++ b/vllm/entrypoints/openai/api_server.py @@ -1572,8 +1572,7 @@ def _log_streaming_response(response, response_body: list) -> None: full_content = full_content[:2048] + "" "...[truncated]" logger.info( - "response_body={streaming_complete: " - "content='%s', chunks=%d}", + "response_body={streaming_complete: content=%r, chunks=%d}", full_content, chunk_count, )