fix(reasoning): Qwen3ReasoningParser returns truncated output as reasoning (#35230)

Signed-off-by: stakeswky <stakeswky@users.noreply.github.com>
Co-authored-by: stakeswky <stakeswky@users.noreply.github.com>
This commit is contained in:
不做了睡大觉
2026-02-27 04:30:45 +08:00
committed by GitHub
parent 3d66502e1b
commit 967572dd5f
2 changed files with 97 additions and 10 deletions

View File

@@ -9,6 +9,7 @@ from tests.reasoning.utils import (
run_reasoning_extraction,
run_reasoning_extraction_streaming,
)
from vllm.entrypoints.openai.chat_completion.protocol import ChatCompletionRequest
from vllm.reasoning import ReasoningParser, ReasoningParserManager
parser_name = "qwen3"
@@ -58,12 +59,14 @@ WITH_THINK_STREAM = {
"content": "This is the rest",
}
# --- No think tokens at all (thinking disabled) ---
# --- No think tokens at all (thinking enabled, truncated) ---
# With thinking enabled (default), no think tokens means the output was
# truncated before </think> could be generated. All output is reasoning.
WITHOUT_THINK = {
"output": "This is the rest",
"reasoning": None,
"content": "This is the rest",
"reasoning": "This is the rest",
"content": None,
}
# In streaming, the parser cannot distinguish "thinking disabled" from
# "reasoning in progress" when no think tokens have appeared yet.
@@ -87,10 +90,12 @@ MULTILINE_REASONING = {
"reasoning": "This is a reasoning\nsection",
"content": "This is the rest\nThat",
}
# Truncated output: <think> present but no </think> (thinking enabled).
# Everything is reasoning because the output was cut off mid-thought.
ONLY_OPEN_TAG = {
"output": "<think>This is a reasoning section",
"reasoning": None,
"content": "This is a reasoning section",
"reasoning": "This is a reasoning section",
"content": None,
}
ONLY_OPEN_TAG_STREAM = {
@@ -99,6 +104,20 @@ ONLY_OPEN_TAG_STREAM = {
"content": None,
}
# Truncated output without <think> prefix (Qwen3.5 style where <think>
# is in the prompt). No </think> means truncation — all is reasoning.
TRUNCATED_NO_START_TOKEN = {
"output": "This is a reasoning section",
"reasoning": "This is a reasoning section",
"content": None,
}
TRUNCATED_NO_START_TOKEN_STREAM = {
"output": "This is a reasoning section",
"reasoning": "This is a reasoning section",
"content": None,
}
TEST_CASES = [
pytest.param(
False,
@@ -170,6 +189,16 @@ TEST_CASES = [
ONLY_OPEN_TAG_STREAM,
id="only_open_tag_stream",
),
pytest.param(
False,
TRUNCATED_NO_START_TOKEN,
id="truncated_no_start_token",
),
pytest.param(
True,
TRUNCATED_NO_START_TOKEN_STREAM,
id="truncated_no_start_token_stream",
),
]
@@ -249,3 +278,46 @@ def test_reasoning_streaming_multi_token_deltas(
assert reconstructor.reasoning == expected_reasoning
assert (reconstructor.other_content or None) == expected_content
# --- Tests for enable_thinking=False (thinking explicitly disabled) ---
THINKING_DISABLED_CASES = [
pytest.param(
"This is plain content",
None,
"This is plain content",
id="thinking_disabled_plain_content",
),
pytest.param(
"Some output without think tokens",
None,
"Some output without think tokens",
id="thinking_disabled_no_think_tokens",
),
]
@pytest.mark.parametrize(
"output, expected_reasoning, expected_content", THINKING_DISABLED_CASES
)
def test_reasoning_thinking_disabled(
output: str,
expected_reasoning: str | None,
expected_content: str | None,
qwen3_tokenizer,
):
"""When enable_thinking=False, output without </think> is all content."""
parser: ReasoningParser = ReasoningParserManager.get_reasoning_parser(parser_name)(
qwen3_tokenizer,
chat_template_kwargs={"enable_thinking": False},
)
reasoning, content = parser.extract_reasoning(
model_output=output,
request=ChatCompletionRequest(messages=[], model="test-model"),
)
assert reasoning == expected_reasoning
assert content == expected_content