Use Transformers v5 WeightRenaming for Transformers modeling backend (#31545)
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
@@ -206,9 +206,7 @@ VLM_TEST_SETTINGS = {
|
||||
"model_impl": "transformers",
|
||||
"default_torch_num_threads": 1,
|
||||
},
|
||||
# FIXME: Investigate why the test hangs
|
||||
# when processing the 3rd prompt in vLLM
|
||||
marks=[pytest.mark.core_model, pytest.mark.skip(reason="Test hangs")],
|
||||
marks=[pytest.mark.core_model],
|
||||
),
|
||||
# Gemma3 has bidirectional mask on images
|
||||
"gemma3-transformers": VLMTestInfo(
|
||||
|
||||
Reference in New Issue
Block a user