Adding "Basic Models Test" and "Multi-Modal Models Test (Extended) 3" in AMD Pipeline (#18106)
Signed-off-by: Alexei V. Ivanov <alexei.ivanov@amd.com> Co-authored-by: Cyrus Leung <cyrus.tl.leung@gmail.com>
This commit is contained in:
committed by
GitHub
parent
01c22335ba
commit
566ec04c3d
@@ -2,6 +2,8 @@
|
||||
"""Test the functionality of the Transformers backend."""
|
||||
import pytest
|
||||
|
||||
from vllm.platforms import current_platform
|
||||
|
||||
from ..conftest import HfRunner, VllmRunner
|
||||
from ..utils import multi_gpu_test
|
||||
from .utils import check_logprobs_close
|
||||
@@ -33,6 +35,9 @@ def check_implementation(
|
||||
)
|
||||
|
||||
|
||||
@pytest.mark.skipif(
|
||||
current_platform.is_rocm(),
|
||||
reason="Llama-3.2-1B-Instruct, Ilama-3.2-1B produce memory access fault.")
|
||||
@pytest.mark.parametrize(
|
||||
"model,model_impl",
|
||||
[
|
||||
@@ -64,6 +69,9 @@ def test_distributed(
|
||||
"meta-llama/Llama-3.2-1B-Instruct", **kwargs)
|
||||
|
||||
|
||||
@pytest.mark.skipif(
|
||||
current_platform.is_rocm(),
|
||||
reason="bitsandbytes quantization is currently not supported in rocm.")
|
||||
@pytest.mark.parametrize("model, quantization_kwargs", [
|
||||
(
|
||||
"meta-llama/Llama-3.2-1B-Instruct",
|
||||
|
||||
Reference in New Issue
Block a user