[CI] Skip the pooling models that do not support transformers v4.55 (#22411)

Signed-off-by: wang.yuqi <noooop@126.com>
This commit is contained in:
wang.yuqi
2025-08-07 14:05:03 +08:00
committed by GitHub
parent 4be02a3776
commit 2a4c825523
4 changed files with 28 additions and 1 deletions

View File

@@ -7,7 +7,7 @@ import pytest
from vllm.config import PoolerConfig
from vllm.platforms import current_platform
from ...utils import check_embeddings_close
from ...utils import check_embeddings_close, check_transformers_version
@pytest.fixture(autouse=True)
@@ -56,6 +56,9 @@ def test_models(
model,
monkeypatch,
) -> None:
if model == "Alibaba-NLP/gte-Qwen2-1.5B-instruct":
check_transformers_version(model, max_transformers_version="4.53.2")
if model == "BAAI/bge-multilingual-gemma2" and current_platform.is_rocm():
# ROCm Triton FA does not currently support sliding window attention
# switch to use ROCm CK FA backend