add test_devstral.py, restore chat_template_kwargs+logprobs to all tests (vLLM-compat spec)

This commit is contained in:
Jinx
2026-04-12 20:27:44 +00:00
parent 1186c9d816
commit b285ffc2d5
4 changed files with 527 additions and 12 deletions

View File

@@ -94,7 +94,10 @@ def test_tool_call_response_flow(streaming: bool = True):
"tools": tools,
"tool_choice": "auto",
"stream": True,
"max_tokens": 512
"max_tokens": 512,
"chat_template_kwargs": {"enable_thinking": False},
"logprobs": True,
"top_logprobs": 5
}
) as response:
print(f"[{timestamp()}] Response status: {response.status_code}")
@@ -152,7 +155,10 @@ def test_tool_call_response_flow(streaming: bool = True):
"tools": tools,
"tool_choice": "auto",
"stream": False,
"max_tokens": 512
"max_tokens": 512,
"chat_template_kwargs": {"enable_thinking": False},
"logprobs": True,
"top_logprobs": 5
}
)
@@ -224,7 +230,10 @@ def test_tool_call_response_flow(streaming: bool = True):
"messages": messages,
"tools": tools,
"stream": True,
"max_tokens": 512
"max_tokens": 512,
"chat_template_kwargs": {"enable_thinking": False},
"logprobs": True,
"top_logprobs": 5
}
) as response:
for line in response.iter_lines():
@@ -258,7 +267,10 @@ def test_tool_call_response_flow(streaming: bool = True):
"messages": messages,
"tools": tools,
"stream": False,
"max_tokens": 512
"max_tokens": 512,
"chat_template_kwargs": {"enable_thinking": False},
"logprobs": True,
"top_logprobs": 5
}
)
@@ -358,7 +370,10 @@ def test_tool_response_with_debug_info():
"tools": tools,
"tool_choice": "auto",
"stream": False,
"max_tokens": 256
"max_tokens": 256,
"chat_template_kwargs": {"enable_thinking": False},
"logprobs": True,
"top_logprobs": 5
}
)
@@ -401,7 +416,10 @@ def test_tool_response_with_debug_info():
"messages": messages,
"tools": tools,
"stream": False,
"max_tokens": 256
"max_tokens": 256,
"chat_template_kwargs": {"enable_thinking": False},
"logprobs": True,
"top_logprobs": 5
}
)