Adding "Basic Models Test" and "Multi-Modal Models Test (Extended) 3" in AMD Pipeline (#18106)

Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com>
Co-authored-by: Cyrus Leung <cyrus.tl.leung@gmail.com>
This commit is contained in:
Alexei-V-Ivanov-AMD
2025-05-15 10:49:23 -05:00
committed by GitHub
parent 01c22335ba
commit 566ec04c3d
4 changed files with 25 additions and 3 deletions

View File

@@ -2,6 +2,8 @@
"""Test the functionality of the Transformers backend."""
import pytest
from vllm.platforms import current_platform
from ..conftest import HfRunner, VllmRunner
from ..utils import multi_gpu_test
from .utils import check_logprobs_close
@@ -33,6 +35,9 @@ def check_implementation(
)
@pytest.mark.skipif(
current_platform.is_rocm(),
reason="Llama-3.2-1B-Instruct, Ilama-3.2-1B produce memory access fault.")
@pytest.mark.parametrize(
"model,model_impl",
[
@@ -64,6 +69,9 @@ def test_distributed(
"meta-llama/Llama-3.2-1B-Instruct", **kwargs)
@pytest.mark.skipif(
current_platform.is_rocm(),
reason="bitsandbytes quantization is currently not supported in rocm.")
@pytest.mark.parametrize("model, quantization_kwargs", [
(
"meta-llama/Llama-3.2-1B-Instruct",