Signed-off-by: JartX <sagformas@epdcenter.es> Signed-off-by: Ifta Khairul Alam Adil <ikaadil007@gmail.com> Signed-off-by: Netanel Haber <58652339+netanel-haber@users.noreply.github.com> Signed-off-by: yewentao256 <zhyanwentao@126.com> Signed-off-by: Philip Ottesen <phiott256@gmail.com> Signed-off-by: Woosuk Kwon <woosuk@inferact.ai> Signed-off-by: Michael Goin <mgoin64@gmail.com> Signed-off-by: Giancarlo Delfin <gdelfin@inferact.ai> Signed-off-by: Andy Lo <andy@mistral.ai> Signed-off-by: Thillai Chithambaram <thillaichithambaram.a@gmail.com> Signed-off-by: sihao.li <sihao.li@intel.com> Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com> Co-authored-by: JartX <sagformas@epdcenter.es> Co-authored-by: Netanel Haber <58652339+netanel-haber@users.noreply.github.com> Co-authored-by: Wentao Ye <44945378+yewentao256@users.noreply.github.com> Co-authored-by: Philip Ottesen <phiott256@gmail.com> Co-authored-by: Woosuk Kwon <woosuk.kwon@berkeley.edu> Co-authored-by: Michael Goin <mgoin64@gmail.com> Co-authored-by: Giancarlo Delfin <32987265+TheEpicDolphin@users.noreply.github.com> Co-authored-by: Andy Lo <andy@mistral.ai> Co-authored-by: Thillai Chithambaram <79466435+thillai-c@users.noreply.github.com> Co-authored-by: sihao_li <165983188+1643661061leo@users.noreply.github.com> Co-authored-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
342 lines
9.1 KiB
Python
342 lines
9.1 KiB
Python
# SPDX-License-Identifier: Apache-2.0
|
|
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
|
|
|
|
import pytest
|
|
from transformers import AutoTokenizer
|
|
|
|
from tests.reasoning.utils import run_reasoning_extraction
|
|
from vllm.reasoning import ReasoningParser, ReasoningParserManager
|
|
|
|
parser_name = "step3p5"
|
|
start_token = "<think>"
|
|
end_token = "</think>"
|
|
|
|
REASONING_MODEL_NAME = "stepfun-ai/Step-3.5-Flash"
|
|
|
|
|
|
@pytest.fixture(scope="module")
|
|
def step3p5_tokenizer():
|
|
return AutoTokenizer.from_pretrained(REASONING_MODEL_NAME)
|
|
|
|
|
|
SIMPLE_REASONING = {
|
|
"output": "This is a reasoning section</think>This is the rest",
|
|
"reasoning": "This is a reasoning section",
|
|
"content": "This is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
# need to get into parser again to remove newline after </think>
|
|
COMPLETE_REASONING = {
|
|
"output": "This is a reasoning section</think>",
|
|
"reasoning": "This is a reasoning section",
|
|
"content": None,
|
|
"is_reasoning_end": False,
|
|
}
|
|
NO_CONTENT = {
|
|
"output": "This is content",
|
|
"reasoning": "This is content",
|
|
"content": None,
|
|
"is_reasoning_end": False,
|
|
}
|
|
NO_REASONING_STREAMING = {
|
|
"output": "This is a reasoning section",
|
|
"reasoning": "This is a reasoning section",
|
|
"content": None,
|
|
"is_reasoning_end": False,
|
|
}
|
|
MULTIPLE_LINES = {
|
|
"output": "This\nThat</think>This is the rest\nThat",
|
|
"reasoning": "This\nThat",
|
|
"content": "This is the rest\nThat",
|
|
"is_reasoning_end": True,
|
|
}
|
|
SHORTEST_REASONING_NO_STREAMING = {
|
|
"output": "</think>This is the rest",
|
|
"reasoning": None,
|
|
"content": "This is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
SHORTEST_REASONING = {
|
|
"output": "</think>This is the rest",
|
|
"reasoning": None,
|
|
"content": "This is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
REASONING_WITH_THINK = {
|
|
"output": "<think>This is a reasoning section</think>This is the rest",
|
|
"reasoning": "This is a reasoning section",
|
|
"content": "This is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
COMPLETE_REASONING_WITH_THINK = {
|
|
"output": "<think>This is a reasoning section</think>",
|
|
"reasoning": "This is a reasoning section",
|
|
"content": None,
|
|
"is_reasoning_end": False,
|
|
}
|
|
MULTIPLE_LINES_WITH_THINK = {
|
|
"output": "<think>This\nThat</think>This is the rest\nThat",
|
|
"reasoning": "This\nThat",
|
|
"content": "This is the rest\nThat",
|
|
"is_reasoning_end": True,
|
|
}
|
|
SHORTEST_REASONING_NO_STREAMING_WITH_THINK = {
|
|
"output": "</think>This is the rest",
|
|
"reasoning": None,
|
|
"content": "This is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
SHORTEST_REASONING_WITH_THINK = {
|
|
"output": "</think>This is the rest",
|
|
"reasoning": None,
|
|
"content": "This is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
THINK_NO_END = {
|
|
"output": "<think>This is a reasoning section",
|
|
"reasoning": "This is a reasoning section",
|
|
"content": None,
|
|
"is_reasoning_end": False,
|
|
}
|
|
EMPTY = {
|
|
"output": "",
|
|
"reasoning": None,
|
|
"content": None,
|
|
"is_reasoning_end": False,
|
|
}
|
|
EMPTY_STREAMING = {
|
|
"output": "",
|
|
"reasoning": None,
|
|
"content": None,
|
|
"is_reasoning_end": False,
|
|
}
|
|
NEW_LINE = {
|
|
"output": "\n<think>This is a reasoning section</think>\nThis is the rest",
|
|
"reasoning": "This is a reasoning section",
|
|
"content": "This is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
|
|
NEW_LINE_STREAMING = {
|
|
"output": "\n<think>This is a reasoning section\n</think>\nThis is the rest",
|
|
"reasoning": "\nThis is a reasoning section",
|
|
"content": "This is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
|
|
NEW_LINE_STREAMING_COMPLEX_CONTENT = {
|
|
"output": "\n This is a \n reasoning section\n\n\n</think>\n\nThis is the rest",
|
|
"reasoning": "\n This is a \n reasoning section\n\n",
|
|
"content": "\nThis is the rest",
|
|
"is_reasoning_end": True,
|
|
}
|
|
|
|
MULTI_TURN_PROMPT_CONTENT = {
|
|
"output": "<think> This is last turn's reasoning section </think> hello <think>",
|
|
"reasoning": "",
|
|
"content": "",
|
|
"is_reasoning_end": False,
|
|
}
|
|
|
|
TEST_CASES = [
|
|
pytest.param(
|
|
False,
|
|
SIMPLE_REASONING,
|
|
id="simple_reasoning",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
SIMPLE_REASONING,
|
|
id="simple_reasoning_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
COMPLETE_REASONING,
|
|
id="complete_reasoning",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
COMPLETE_REASONING,
|
|
id="complete_reasoning_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
NO_CONTENT,
|
|
id="no_content_token",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
NO_REASONING_STREAMING,
|
|
id="no_reasoning_token_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
MULTIPLE_LINES,
|
|
id="multiple_lines",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
MULTIPLE_LINES,
|
|
id="multiple_lines_streaming",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
SHORTEST_REASONING,
|
|
id="shortest",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
SHORTEST_REASONING_NO_STREAMING,
|
|
id="shortest_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
REASONING_WITH_THINK,
|
|
id="reasoning_with_think",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
REASONING_WITH_THINK,
|
|
id="reasoning_with_think_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
COMPLETE_REASONING_WITH_THINK,
|
|
id="complete_reasoning_with_think",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
COMPLETE_REASONING_WITH_THINK,
|
|
id="complete_reasoning_with_think_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
MULTIPLE_LINES_WITH_THINK,
|
|
id="multiple_lines_with_think",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
MULTIPLE_LINES_WITH_THINK,
|
|
id="multiple_lines_with_think_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
SHORTEST_REASONING_NO_STREAMING_WITH_THINK,
|
|
id="shortest_with_think",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
SHORTEST_REASONING_WITH_THINK,
|
|
id="shortest_with_think_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
THINK_NO_END,
|
|
id="think_no_end",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
THINK_NO_END,
|
|
id="think_no_end_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
EMPTY,
|
|
id="empty",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
EMPTY_STREAMING,
|
|
id="empty_streaming",
|
|
),
|
|
pytest.param(
|
|
False,
|
|
NEW_LINE,
|
|
id="new_line",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
NEW_LINE_STREAMING,
|
|
id="new_line_streaming",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
NEW_LINE_STREAMING_COMPLEX_CONTENT,
|
|
id="new_line_streaming_complex_content",
|
|
),
|
|
pytest.param(
|
|
True,
|
|
MULTI_TURN_PROMPT_CONTENT,
|
|
id="multi_turn_prompt_content",
|
|
),
|
|
]
|
|
|
|
|
|
@pytest.mark.parametrize("streaming, param_dict", TEST_CASES)
|
|
def test_reasoning(
|
|
streaming: bool,
|
|
param_dict: dict,
|
|
step3p5_tokenizer,
|
|
request,
|
|
):
|
|
output = step3p5_tokenizer.tokenize(param_dict["output"])
|
|
# decode everything to tokens
|
|
output_tokens: list[str] = [
|
|
step3p5_tokenizer.convert_tokens_to_string([token]) for token in output
|
|
]
|
|
parser: ReasoningParser = ReasoningParserManager.get_reasoning_parser(parser_name)(
|
|
step3p5_tokenizer
|
|
)
|
|
|
|
reasoning, content = run_reasoning_extraction(
|
|
parser, output_tokens, streaming=streaming
|
|
)
|
|
|
|
print(f"reasoning: {reasoning}")
|
|
print(f"content: {content}")
|
|
test_id = request.node.callspec.id if hasattr(request.node, "callspec") else None
|
|
if request.node.callspec.id != "multi_turn_prompt_content":
|
|
assert reasoning == param_dict["reasoning"]
|
|
assert content == param_dict["content"]
|
|
|
|
# Test is_reasoning_end
|
|
output_ids = step3p5_tokenizer.convert_tokens_to_ids(output)
|
|
if streaming:
|
|
is_reasoning_end = parser.is_reasoning_end(output_ids)
|
|
assert is_reasoning_end == param_dict["is_reasoning_end"]
|
|
|
|
# Test extract_content
|
|
if param_dict["content"] is not None:
|
|
content = parser.extract_content_ids(output_ids)
|
|
# Fixed expected token ids for specific test cases
|
|
test_id = (
|
|
request.node.callspec.id if hasattr(request.node, "callspec") else None
|
|
)
|
|
# Match most specific first
|
|
if test_id not in [
|
|
"new_line_streaming_complex_content",
|
|
"new_line_streaming",
|
|
"new_line",
|
|
"multi_turn_prompt_content",
|
|
]:
|
|
expected_content_ids = step3p5_tokenizer.convert_tokens_to_ids(
|
|
step3p5_tokenizer.tokenize(param_dict["content"])
|
|
)
|
|
assert content == expected_content_ids
|
|
else:
|
|
content = parser.extract_content_ids(output)
|
|
assert content == []
|
|
|
|
|
|
def test_step3p5_streaming_drops_leading_newline(step3p5_tokenizer):
|
|
parser_cls = ReasoningParserManager.get_reasoning_parser("step3p5")
|
|
parser = parser_cls(step3p5_tokenizer)
|
|
output = "<think>calc</think>\nAnswer"
|
|
tokens = step3p5_tokenizer.tokenize(output)
|
|
output_tokens = [
|
|
step3p5_tokenizer.convert_tokens_to_string([token]) for token in tokens
|
|
]
|
|
|
|
_, content = run_reasoning_extraction(parser, output_tokens, streaming=True)
|
|
assert content == "Answer"
|