[CI/Build] Fix model nightly tests (#26466)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
@@ -749,6 +749,7 @@ VLM_TEST_SETTINGS = {
|
||||
max_num_seqs=2,
|
||||
auto_cls=AutoModelForImageTextToText,
|
||||
hf_output_post_proc=model_utils.smolvlm_trunc_hf_output,
|
||||
num_logprobs=10,
|
||||
),
|
||||
"tarsier": VLMTestInfo(
|
||||
models=["omni-research/Tarsier-7b"],
|
||||
|
||||
@@ -45,14 +45,16 @@ def _run_test(
|
||||
|
||||
all_outputs = []
|
||||
for inputs in all_inputs:
|
||||
inputs = hf_model.wrap_device(inputs)
|
||||
|
||||
if "pixel_values" in inputs:
|
||||
inputs.pop("input_ids")
|
||||
pooled_output = hf_model.model.get_image_features(
|
||||
**hf_model.wrap_device(inputs)
|
||||
pixel_values=inputs.pixel_values,
|
||||
).squeeze(0)
|
||||
else:
|
||||
pooled_output = hf_model.model.get_text_features(
|
||||
**hf_model.wrap_device(inputs)
|
||||
input_ids=inputs.input_ids,
|
||||
attention_mask=inputs.attention_mask,
|
||||
).squeeze(0)
|
||||
|
||||
all_outputs.append(pooled_output.tolist())
|
||||
|
||||
Reference in New Issue
Block a user