[Frontend] Improve StreamingResponse Exception Handling (#11752)

This commit is contained in:
Robert Shaw
2025-01-05 16:35:01 -05:00
committed by GitHub
parent eba17173d3
commit 33fc1e2e86
2 changed files with 3 additions and 3 deletions

View File

@@ -301,7 +301,7 @@ class OpenAIServingChat(OpenAIServing):
] * num_choices
else:
tool_parsers = [None] * num_choices
except RuntimeError as e:
except Exception as e:
logger.exception("Error in tool parser creation.")
data = self.create_streaming_error_response(str(e))
yield f"data: {data}\n\n"
@@ -591,7 +591,7 @@ class OpenAIServingChat(OpenAIServing):
completion_tokens=num_completion_tokens,
total_tokens=num_prompt_tokens + num_completion_tokens)
except ValueError as e:
except Exception as e:
# TODO: Use a vllm-specific Validation Error
logger.exception("Error in chat completion stream generator.")
data = self.create_streaming_error_response(str(e))