[Bugfix] Disable multi-modal preprocessor cache for DP (#21896)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
@@ -1197,6 +1197,18 @@ class EngineArgs:
|
||||
enable_multimodal_encoder_data_parallel,
|
||||
)
|
||||
|
||||
supports_mm_preprocessor_cache = (self.data_parallel_size == 1
|
||||
or data_parallel_external_lb)
|
||||
if (not supports_mm_preprocessor_cache
|
||||
and model_config.is_multimodal_model
|
||||
and not model_config.disable_mm_preprocessor_cache):
|
||||
logger.warning(
|
||||
"Multi-modal preprocessor cache is not compatible "
|
||||
"with data parallelism when there does not exist a "
|
||||
"one-to-one correspondance between API process and "
|
||||
"EngineCore process, so the cache will be disabled.")
|
||||
model_config.set_disable_mm_preprocessor_cache(True)
|
||||
|
||||
speculative_config = self.create_speculative_config(
|
||||
target_model_config=model_config,
|
||||
target_parallel_config=parallel_config,
|
||||
|
||||
Reference in New Issue
Block a user