[EASY] Drop duplicate KV-cache initialization (#38799)

Signed-off-by: namgyu-youn <namgyu.dev@gmail.com>
This commit is contained in:
namgyu-youn
2026-04-07 03:05:39 +09:00
committed by GitHub
parent 419e73cdfa
commit 94fbb09894

View File

@@ -131,9 +131,6 @@ def _init_kv_cache_quant(
quant_config: Optional quantization configuration.
prefix: Layer name prefix for quantization method lookup.
"""
quant_method = (
quant_config.get_quant_method(layer, prefix=prefix) if quant_config else None
)
# Note [Register q/k/v/prob scales in state dict]
# When calling model.to(device), only parameters/buffers in state dict are