From 7693c8eabf88f72d5db14d376539782b34b09127 Mon Sep 17 00:00:00 2001 From: Harry Mellor <19981378+hmellor@users.noreply.github.com> Date: Fri, 20 Mar 2026 14:49:40 +0000 Subject: [PATCH] Fix attribute error in `isaac_patch_hf_runner` (#37685) Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com> (cherry picked from commit 9f6d9dd371c63154dddd2a8b85d7f337f3e10911) --- .../multimodal/generation/vlm_utils/model_utils.py | 13 +++++++++---- 1 file changed, 9 insertions(+), 4 deletions(-) diff --git a/tests/models/multimodal/generation/vlm_utils/model_utils.py b/tests/models/multimodal/generation/vlm_utils/model_utils.py index c4465657e..414feab26 100644 --- a/tests/models/multimodal/generation/vlm_utils/model_utils.py +++ b/tests/models/multimodal/generation/vlm_utils/model_utils.py @@ -24,6 +24,7 @@ from transformers import ( GenerationConfig, GenerationMixin, ) +from transformers.masking_utils import create_causal_mask from transformers.video_utils import VideoMetadata from vllm.logprobs import SampleLogprobs @@ -679,10 +680,14 @@ def isaac_patch_hf_runner(hf_model: HfRunner) -> HfRunner: sin = sin.to(inputs_embeds.dtype) # Prepare attention mask - if attention_mask is not None: - attention_mask = self._update_causal_mask( - attention_mask, inputs_embeds, cache_position, past_key_values, False - ) + attention_mask = create_causal_mask( + config=self.config, + input_embeds=inputs_embeds, + attention_mask=attention_mask, + past_key_values=past_key_values, + position_ids=position_ids, + cache_position=cache_position, + ) # Initialize and collect hidden states hidden_states = inputs_embeds