Fix Responses JSON schema alias serialization (#38519)

Signed-off-by: noobhappylife <aratar1991@hotmail.com>
Co-authored-by: OpenAI Codex <codex@openai.com>
This commit is contained in:
noobHappylife
2026-04-09 10:50:16 +08:00
committed by GitHub
parent d37b378762
commit 2a49284c8a
6 changed files with 104 additions and 32 deletions

View File

@@ -999,17 +999,21 @@ async def test_mcp_tool_multi_turn(client: OpenAI, model_name: str, server):
(msg.get("recipient") or "").startswith("python")
for msg in response1.output_messages
)
parsed_output_messages = [
Message.from_dict(msg) for msg in response1.output_messages
]
tool_response_found = any(
msg.get("author", {}).get("role") == "tool"
and (msg.get("author", {}).get("name") or "").startswith("python")
for msg in response1.output_messages
(msg.author.role == "tool" and (msg.author.name or "").startswith("python"))
for msg in parsed_output_messages
)
assert tool_call_found, "MCP tool call not found in output_messages"
assert tool_response_found, "MCP tool response not found in output_messages"
# No developer messages expected for elevated tools
developer_msgs = [
msg for msg in response1.input_messages if msg["author"]["role"] == "developer"
msg
for msg in (Message.from_dict(raw) for raw in response1.input_messages)
if msg.author.role == "developer"
]
assert len(developer_msgs) == 0, "No developer message expected for elevated tools"
@@ -1119,12 +1123,10 @@ async def test_function_call_with_previous_input_messages(
num_system = 0
num_developer = 0
num_tool = 0
for msg_dict in response_2.input_messages:
# input_messages use {"author": {"role": "..."}} format,
# not the top-level {"role": "..."} that Message.from_dict
# expects.
author = msg_dict.get("author", {})
role = author.get("role") if isinstance(author, dict) else None
for message in (
Message.from_dict(msg_dict) for msg_dict in response_2.input_messages
):
role = message.author.role
if role == "system":
num_system += 1
elif role == "developer":
@@ -1183,12 +1185,8 @@ async def test_system_prompt_override_no_duplication(client: OpenAI, model_name:
assert response.output_text is not None
num_system = 0
for msg in response.input_messages:
# input_messages use {"author": {"role": "system"}} format,
# not the top-level {"role": "system"} that Message.from_dict expects.
author = msg.get("author", {})
role = author.get("role") if isinstance(author, dict) else None
if role == "system":
for message in (Message.from_dict(msg) for msg in response.input_messages):
if message.author.role == "system":
num_system += 1
assert num_system == 1, f"Expected 1 system message, got {num_system}"