reasoning_content -> reasoning (#27752)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
Harry Mellor
2025-11-08 04:15:08 -08:00
committed by GitHub
parent 608bb14462
commit d9ab1ad9d1
46 changed files with 428 additions and 438 deletions

View File

@@ -9,25 +9,28 @@ from vllm.transformers_utils.tokenizers.mistral import MistralTokenizer
class StreamingReasoningReconstructor:
def __init__(self):
self.reasoning_content = None
self.reasoning = None
self.other_content = None
def append_delta(self, delta: DeltaMessage):
# content and the reasoning content should not be present
# at the same time
assert delta.content is None or delta.reasoning_content is None, (
assert delta.content is None or delta.reasoning is None, (
"Both content and reasoning content are present in the delta message"
)
assert delta.reasoning == delta.reasoning_content, (
"reasoning_content should be present for backwards compatibility"
)
if delta.content is not None:
if self.other_content is None:
self.other_content = delta.content
else:
self.other_content += delta.content
else:
if self.reasoning_content is None:
self.reasoning_content = delta.reasoning_content
if self.reasoning is None:
self.reasoning = delta.reasoning
else:
self.reasoning_content += delta.reasoning_content
self.reasoning += delta.reasoning
def run_reasoning_extraction(
@@ -43,7 +46,7 @@ def run_reasoning_extraction(
request,
)
return (
reconstructor.reasoning_content,
reconstructor.reasoning,
reconstructor.other_content or None,
)
else:
@@ -69,7 +72,7 @@ def run_reasoning_extraction_mistral(
request,
)
return (
reconstructor.reasoning_content,
reconstructor.reasoning,
reconstructor.other_content or None,
)
else:
@@ -88,7 +91,7 @@ def run_reasoning_extraction_nonstreaming(
request: ChatCompletionRequest | None = None,
) -> tuple[str | None, str | None]:
request = request or ChatCompletionRequest(messages=[], model="test-model")
return reasoning_parser.extract_reasoning_content(
return reasoning_parser.extract_reasoning(
model_output="".join(model_output), request=request
)
@@ -110,7 +113,7 @@ def run_reasoning_extraction_streaming(
]
current_text = previous_text + delta
current_tokens = previous_tokens + token_delta
delta_message = reasoning_parser.extract_reasoning_content_streaming(
delta_message = reasoning_parser.extract_reasoning_streaming(
previous_text,
current_text,
delta,
@@ -142,7 +145,7 @@ def run_reasoning_extraction_streaming_mistral(
delta = reasoning_parser.model_tokenizer.convert_ids_to_tokens([model_delta])[0]
current_text = previous_text + delta
current_tokens = previous_tokens + token_delta
delta_message = reasoning_parser.extract_reasoning_content_streaming(
delta_message = reasoning_parser.extract_reasoning_streaming(
previous_text,
current_text,
delta,