diff --git a/vllm/model_executor/models/kimi_k25.py b/vllm/model_executor/models/kimi_k25.py index dccf05c14..24c89d222 100644 --- a/vllm/model_executor/models/kimi_k25.py +++ b/vllm/model_executor/models/kimi_k25.py @@ -58,6 +58,7 @@ from vllm.multimodal.processing import ( PromptReplacement, PromptUpdate, ) +from vllm.platforms import current_platform from vllm.sequence import IntermediateTensors from vllm.transformers_utils.configs import KimiK25Config from vllm.transformers_utils.processor import cached_get_image_processor @@ -320,7 +321,7 @@ class KimiK25ForConditionalGeneration(nn.Module, SupportsMultiModal, SupportsPP) model_config.multimodal_config.mm_encoder_tp_mode == "data" ) self.hidden_size = config.text_config.hidden_size - self.device = torch.cuda.current_device() + self.device = current_platform.current_device() # Build vision tower directly with KimiK25VisionConfig self.vision_tower = MoonViT3dPretrainedModel( config.vision_config,