add test_devstral.py, restore chat_template_kwargs+logprobs to all tests (vLLM-compat spec)
This commit is contained in:
@@ -91,7 +91,10 @@ def test_streaming_tool_call_with_code():
|
||||
"tools": tools,
|
||||
"tool_choice": "auto",
|
||||
"stream": True,
|
||||
"max_tokens": 4096
|
||||
"max_tokens": 4096,
|
||||
"chat_template_kwargs": {"enable_thinking": False},
|
||||
"logprobs": True,
|
||||
"top_logprobs": 5
|
||||
}
|
||||
) as response:
|
||||
print(f"[{timestamp()}] Response status: {response.status_code}")
|
||||
@@ -242,7 +245,10 @@ def test_streaming_tool_call_with_json():
|
||||
"tools": tools,
|
||||
"tool_choice": "auto",
|
||||
"stream": True,
|
||||
"max_tokens": 2048
|
||||
"max_tokens": 2048,
|
||||
"chat_template_kwargs": {"enable_thinking": False},
|
||||
"logprobs": True,
|
||||
"top_logprobs": 5
|
||||
}
|
||||
) as response:
|
||||
for line in response.iter_lines():
|
||||
@@ -328,7 +334,10 @@ def test_non_streaming_tool_call():
|
||||
"tools": tools,
|
||||
"tool_choice": "auto",
|
||||
"stream": False,
|
||||
"max_tokens": 1024
|
||||
"max_tokens": 1024,
|
||||
"chat_template_kwargs": {"enable_thinking": False},
|
||||
"logprobs": True,
|
||||
"top_logprobs": 5
|
||||
}
|
||||
)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user