[CI/Build] Bump test transformers version (#10106)

Signed-off-by: Isotr0py <2037008807@qq.com>
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Co-authored-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
Isotr0py
2024-12-06 00:05:52 +08:00
committed by GitHub
parent 571da8fc43
commit 998eeafe58
5 changed files with 3 additions and 35 deletions

View File

@@ -6,7 +6,6 @@ from pathlib import PosixPath
from typing import Type
import pytest
import transformers
from transformers import AutoModelForVision2Seq
from transformers.utils import is_flash_attn_2_available
@@ -187,12 +186,6 @@ VLM_TEST_SETTINGS = {
comparator=check_outputs_equal,
max_tokens=8,
dtype="bfloat16",
marks=[
pytest.mark.skipif(
transformers.__version__ < "4.46.2",
reason="Model broken in HF, see huggingface/transformers#34379"
),
]
),
"fuyu": VLMTestInfo(
models=["adept/fuyu-8b"],
@@ -243,13 +236,7 @@ VLM_TEST_SETTINGS = {
max_model_len=8192,
max_num_seqs=2,
auto_cls=AutoModelForVision2Seq,
marks=[
pytest.mark.skipif(
transformers.__version__ < "4.46.0",
reason="Model introduced in HF >= 4.46.0"
),
large_gpu_mark(min_gb=48),
],
marks=[large_gpu_mark(min_gb=48)],
),
"intern_vl": VLMTestInfo(
models=[
@@ -318,12 +305,6 @@ VLM_TEST_SETTINGS = {
auto_cls=AutoModelForVision2Seq,
vllm_output_post_proc=model_utils.llava_video_vllm_to_hf_output,
image_sizes=[((1669, 2560), (2560, 1669), (183, 488), (488, 183))],
marks=[
pytest.mark.skipif(
transformers.__version__ < "4.46.2",
reason="Model broken with changes in transformers 4.46"
)
],
),
"minicpmv_25": VLMTestInfo(
models=["openbmb/MiniCPM-Llama3-V-2_5"],
@@ -404,10 +385,6 @@ VLM_TEST_SETTINGS = {
cuda_device_count_stateless() < 2,
reason="Need at least 2 GPUs to run the test.",
),
pytest.mark.skipif(
transformers.__version__ < "4.46.2",
reason="Model broken in HF, see huggingface/transformers#34379"
)
],
**COMMON_BROADCAST_SETTINGS # type: ignore
),