[Fix] Enable mm_processor_cache with vision LoRA (#31927)

Signed-off-by: prashanth058 <prashanth.dannamaneni@uipath.com>
This commit is contained in:
prashanth058
2026-01-08 01:31:51 -06:00
committed by GitHub
parent 287b37cda4
commit d3235cb503
5 changed files with 52 additions and 16 deletions

View File

@@ -562,15 +562,17 @@ class InputProcessor:
mm_features = []
for modality, idx in sorted_mm_idxs:
base_mm_hash = decoder_mm_hashes[modality][idx]
mm_features.append(
MultiModalFeatureSpec(
data=decoder_mm_inputs[modality][idx],
modality=modality,
identifier=self._get_mm_identifier(
decoder_mm_hashes[modality][idx],
base_mm_hash,
lora_request,
),
mm_position=decoder_mm_positions[modality][idx],
mm_hash=base_mm_hash,
)
)