[Bugfix] Remove erroneous lower bound on LoRA vocab size constraint (#35354)

Signed-off-by: Lucas Wilkinson <lwilkins@redhat.com>
This commit is contained in:
Lucas Wilkinson
2026-02-26 13:44:50 -05:00
committed by GitHub
parent a1f53addb1
commit 5e58bdc711
2 changed files with 4 additions and 6 deletions

View File

@@ -469,7 +469,7 @@ def test_lm_head_logits_processor(
@torch.inference_mode()
@pytest.mark.parametrize("vocab_size", [512, 32000, 258049, 300000])
@pytest.mark.parametrize("vocab_size", [258049, 300000])
@pytest.mark.parametrize("device", DEVICES)
def test_lm_head_logits_processor_invalid_vocab_size(
default_vllm_config, dist_init, vocab_size, device
@@ -489,7 +489,7 @@ def test_lm_head_logits_processor_invalid_vocab_size(
logits_processor, 1024, torch.float16, device, None
)
with pytest.raises(ValueError, match="vocab size must be > 32000 and <= 258048"):
with pytest.raises(ValueError, match="vocab size must be <= 258048"):
lora_logits_processor.create_lora_weights(max_loras, lora_config)

View File

@@ -88,10 +88,8 @@ class LogitsProcessorWithLoRA(BaseLayerWithLoRA):
model_config: PretrainedConfig | None = None,
) -> None:
# TODO: Verify if this condition can be further relaxed
if self.base_layer.vocab_size <= 32000 or self.base_layer.vocab_size > 258048:
raise ValueError(
"When using LoRA, vocab size must be > 32000 and <= 258048"
)
if self.base_layer.vocab_size > 258048:
raise ValueError("When using LoRA, vocab size must be <= 258048")
self.lora_a_stacked = torch.zeros(
(
max_loras,