[bnb] Skip moe + bnb test (#36896)

Signed-off-by: Marc Sun <marc@huggingface.co>
This commit is contained in:
Marc Sun
2026-03-12 19:03:25 +01:00
committed by GitHub
parent e39257a552
commit c973ecdead

View File

@@ -6,7 +6,9 @@ Run `pytest tests/quantization/test_bitsandbytes.py`.
"""
import pytest
from packaging.version import Version
from transformers import BitsAndBytesConfig
from transformers import __version__ as TRANSFORMERS_VERSION
from tests.quantization.utils import is_quant_method_supported
from vllm.platforms import current_platform
@@ -138,6 +140,12 @@ def test_load_pp_4bit_bnb_model(model_name, description) -> None:
compare_two_settings(model_name, common_args, pp_args)
@pytest.mark.skipif(
Version(TRANSFORMERS_VERSION) >= Version("5.0.0"),
reason="Need to add support for quantizing MoE experts with bnb"
" in transformers v5. See"
" https://github.com/bitsandbytes-foundation/bitsandbytes/issues/1849",
)
@pytest.mark.skipif(
not is_quant_method_supported("bitsandbytes"),
reason="bitsandbytes is not supported on this GPU type.",