[Feature] Add SM103 (Blackwell Ultra) Support to vLLM (#30484)
Signed-off-by: LopezCastroRoberto <robertol.c510@gmail.com> Signed-off-by: Roberto L. Castro <38211239+LopezCastroRoberto@users.noreply.github.com> Co-authored-by: youkaichao <youkaichao@gmail.com>
This commit is contained in:
committed by
GitHub
parent
57e9bf1864
commit
4fa7ce46f3
@@ -363,7 +363,7 @@ class HybridAttentionMambaModelConfig(VerifyAndUpdateConfig):
|
||||
else:
|
||||
kernel_block_alignment_size = 16
|
||||
if (
|
||||
current_platform.is_device_capability(100)
|
||||
current_platform.is_device_capability_family(100)
|
||||
and model_config.get_head_size() == 256
|
||||
and (
|
||||
attention_config.backend is None
|
||||
|
||||
Reference in New Issue
Block a user