[CI/Build] Fix VLM test failures when using transformers v4.46 (#9666)

This commit is contained in:
Cyrus Leung
2024-10-25 01:40:40 +08:00
committed by GitHub
parent d27cfbf791
commit c866e0079d
4 changed files with 28 additions and 12 deletions

View File

@@ -32,8 +32,8 @@ HF_MULTIIMAGE_IMAGE_PROMPT = \
models = ["openbmb/MiniCPM-Llama3-V-2_5"]
def _wrap_inputs(hf_inputs: BatchEncoding) -> BatchEncoding:
return BatchEncoding({"model_inputs": hf_inputs})
def _wrap_inputs(hf_inputs: BatchEncoding):
return {"model_inputs": hf_inputs}
def trunc_hf_output(hf_output: Tuple[List[int], str,