[CI/Build] Fix VLM test failures when using transformers v4.46 (#9666)

This commit is contained in:
Cyrus Leung
2024-10-25 01:40:40 +08:00
committed by GitHub
parent d27cfbf791
commit c866e0079d
4 changed files with 28 additions and 12 deletions

View File

@@ -1,6 +1,7 @@
from typing import List, Optional, Type
import pytest
import transformers
from transformers import AutoModelForVision2Seq, BatchEncoding
from vllm.multimodal.utils import rescale_image_size
@@ -93,6 +94,10 @@ def run_test(
)
@pytest.mark.skipif(
transformers.__version__.startswith("4.46.0"),
reason="Model broken in HF, see huggingface/transformers#34379",
)
@pytest.mark.parametrize("model", models)
@pytest.mark.parametrize(
"size_factors",