[Model] Support Llama4 in vLLM (#16104)
This commit is contained in:
@@ -280,6 +280,7 @@ def _test_processing_correctness_mistral(
|
||||
"Skywork/Skywork-R1V-38B",
|
||||
"fixie-ai/ultravox-v0_5-llama-3_2-1b",
|
||||
"openai/whisper-large-v3",
|
||||
"meta-llama/Llama-4-Scout-17B-16E-Instruct",
|
||||
])
|
||||
@pytest.mark.parametrize("hit_rate", [0.3, 0.5, 1.0])
|
||||
@pytest.mark.parametrize("num_batches", [32])
|
||||
|
||||
Reference in New Issue
Block a user