diff --git a/tests/entrypoints/openai/test_serving_chat.py b/tests/entrypoints/openai/test_serving_chat.py index 1d96b05ac..33c69578c 100644 --- a/tests/entrypoints/openai/test_serving_chat.py +++ b/tests/entrypoints/openai/test_serving_chat.py @@ -126,7 +126,7 @@ def gptoss_speculative_server(default_server_args: list[str]): if is_aiter_found_and_supported(): env_dict = {"VLLM_ROCM_USE_AITER": "1"} with RemoteOpenAIServer( - GPT_OSS_MODEL_NAME, server_args, env_dict=env_dict + GPT_OSS_MODEL_NAME, server_args, env_dict=env_dict, max_wait_seconds=480 ) as remote_server: yield remote_server diff --git a/tests/utils.py b/tests/utils.py index 5252115f2..9ab6df9e2 100644 --- a/tests/utils.py +++ b/tests/utils.py @@ -190,7 +190,7 @@ class RemoteOpenAIServer: model_loader.download_model(model_config) self._start_server(model, vllm_serve_args, env_dict) - max_wait_seconds = max_wait_seconds or 240 + max_wait_seconds = max_wait_seconds or 360 self._wait_for_server(url=self.url_for("health"), timeout=max_wait_seconds) def __enter__(self):