From f5432e35a3a4f0bd6e7d49c51a35a0a01bc32452 Mon Sep 17 00:00:00 2001 From: Micah Williamson Date: Thu, 19 Feb 2026 23:37:49 -0600 Subject: [PATCH] [ROCm][CI] Loosen RemoteOpenAIServer Startup Timeout (#34922) Signed-off-by: Micah Williamson --- tests/entrypoints/openai/test_serving_chat.py | 2 +- tests/utils.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/tests/entrypoints/openai/test_serving_chat.py b/tests/entrypoints/openai/test_serving_chat.py index 1d96b05ac..33c69578c 100644 --- a/tests/entrypoints/openai/test_serving_chat.py +++ b/tests/entrypoints/openai/test_serving_chat.py @@ -126,7 +126,7 @@ def gptoss_speculative_server(default_server_args: list[str]): if is_aiter_found_and_supported(): env_dict = {"VLLM_ROCM_USE_AITER": "1"} with RemoteOpenAIServer( - GPT_OSS_MODEL_NAME, server_args, env_dict=env_dict + GPT_OSS_MODEL_NAME, server_args, env_dict=env_dict, max_wait_seconds=480 ) as remote_server: yield remote_server diff --git a/tests/utils.py b/tests/utils.py index 5252115f2..9ab6df9e2 100644 --- a/tests/utils.py +++ b/tests/utils.py @@ -190,7 +190,7 @@ class RemoteOpenAIServer: model_loader.download_model(model_config) self._start_server(model, vllm_serve_args, env_dict) - max_wait_seconds = max_wait_seconds or 240 + max_wait_seconds = max_wait_seconds or 360 self._wait_for_server(url=self.url_for("health"), timeout=max_wait_seconds) def __enter__(self):