[1/n][CI] Load models in CI from S3 instead of HF (#13205)
Signed-off-by: <> Co-authored-by: EC2 Default User <ec2-user@ip-172-31-20-117.us-west-2.compute.internal>
This commit is contained in:
@@ -7,6 +7,7 @@ from transformers import PretrainedConfig
|
||||
|
||||
from vllm import LLM
|
||||
|
||||
from ..conftest import MODELS_ON_S3
|
||||
from .registry import HF_EXAMPLE_MODELS
|
||||
|
||||
|
||||
@@ -42,8 +43,11 @@ def test_can_initialize(model_arch):
|
||||
|
||||
with patch.object(LLM.get_engine_class(), "_initialize_kv_caches",
|
||||
_initialize_kv_caches):
|
||||
model_name = model_info.default
|
||||
if model_name in MODELS_ON_S3:
|
||||
model_name = f"s3://vllm-ci-model-weights/{model_name.split('/')[-1]}"
|
||||
LLM(
|
||||
model_info.default,
|
||||
model_name,
|
||||
tokenizer=model_info.tokenizer,
|
||||
tokenizer_mode=model_info.tokenizer_mode,
|
||||
speculative_model=model_info.speculative_model,
|
||||
|
||||
Reference in New Issue
Block a user