[CI/Build][Bugfix]Fix Quantized Models Test on AMD (#27712)

Signed-off-by: zhewenli <zhewenli@meta.com>
This commit is contained in:
Zhewen Li
2025-10-28 23:27:30 -07:00
committed by GitHub
parent a4a4f0f617
commit 83fd49b1fc
3 changed files with 8 additions and 2 deletions

View File

@@ -9,10 +9,16 @@ import pytest
from transformers import BitsAndBytesConfig
from tests.quantization.utils import is_quant_method_supported
from vllm.platforms import current_platform
from ...utils import compare_two_settings, multi_gpu_test
from ..utils import check_embeddings_close, check_logprobs_close
pytestmark = pytest.mark.skipif(
current_platform.is_rocm(),
reason="bitsandbytes quantization not supported on ROCm (CUDA-only kernels)",
)
models_4bit_to_test = [
("facebook/opt-125m", "quantize opt model inflight"),
(