Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
Signed-off-by: wang.yuqi <noooop@126.com>
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
(cherry picked from commit abb34ac43a)
620 lines
22 KiB
Python
620 lines
22 KiB
Python
# SPDX-License-Identifier: Apache-2.0
|
|
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
|
|
|
|
from collections.abc import Iterable
|
|
from contextlib import contextmanager
|
|
from typing import TYPE_CHECKING, Any, TypeVar, cast
|
|
|
|
import torch
|
|
import torch.nn as nn
|
|
|
|
from vllm.config import VllmConfig
|
|
from vllm.logger import init_logger
|
|
from vllm.model_executor.layers.activation import get_act_fn
|
|
from vllm.model_executor.models.config import VerifyAndUpdateConfig
|
|
from vllm.transformers_utils.config import (
|
|
try_get_dense_modules,
|
|
)
|
|
from vllm.transformers_utils.repo_utils import get_hf_file_bytes
|
|
|
|
from .interfaces import supports_multimodal
|
|
from .interfaces_base import VllmModelForPooling, is_pooling_model
|
|
|
|
if TYPE_CHECKING:
|
|
from vllm.config import ModelConfig, VllmConfig
|
|
from vllm.model_executor.layers.pooler import Pooler
|
|
|
|
_T = TypeVar("_T", bound=type[nn.Module])
|
|
|
|
logger = init_logger(__name__)
|
|
|
|
_GENERATE_SUFFIXES = [
|
|
"ForCausalLM",
|
|
"ForConditionalGeneration",
|
|
"ChatModel",
|
|
"LMHeadModel",
|
|
]
|
|
|
|
|
|
def _load_st_projector(model_config: "ModelConfig") -> nn.Module | None:
|
|
"""Load Sentence-Transformers Dense projection layers."""
|
|
|
|
dense_modules = try_get_dense_modules(
|
|
model_config.model, revision=model_config.revision
|
|
)
|
|
|
|
if dense_modules is None:
|
|
return
|
|
|
|
try:
|
|
layers = []
|
|
for layer_config in dense_modules:
|
|
folder = layer_config["folder"]
|
|
linear = nn.Linear(
|
|
layer_config["in_features"],
|
|
layer_config["out_features"],
|
|
bias=layer_config.get("bias", True),
|
|
dtype=model_config.head_dtype,
|
|
)
|
|
if not _load_dense_weights(linear, folder, model_config):
|
|
continue
|
|
layers.append(linear)
|
|
if act_name := layer_config.get("activation_function"):
|
|
layers.append(get_act_fn(act_name))
|
|
return nn.Sequential(*layers).to(dtype=model_config.head_dtype)
|
|
except Exception:
|
|
logger.exception("ST projector loading failed")
|
|
|
|
return None
|
|
|
|
|
|
def _load_dense_weights(
|
|
linear: nn.Linear, folder: str, model_config: "ModelConfig"
|
|
) -> bool:
|
|
"""Load weights using vLLM's weight_loader pattern."""
|
|
from vllm.model_executor.model_loader.weight_utils import default_weight_loader
|
|
|
|
for filename in ["model.safetensors", "pytorch_model.bin"]:
|
|
file_path = f"{folder}/{filename}" if folder else filename
|
|
|
|
try:
|
|
file_bytes = get_hf_file_bytes(
|
|
file_path, model_config.model, model_config.revision
|
|
)
|
|
if not file_bytes:
|
|
continue
|
|
|
|
if filename.endswith(".safetensors"):
|
|
from safetensors.torch import load as load_safetensors
|
|
|
|
state_dict = load_safetensors(file_bytes)
|
|
else:
|
|
import io
|
|
|
|
state_dict = torch.load(
|
|
io.BytesIO(file_bytes), map_location="cpu", weights_only=True
|
|
)
|
|
|
|
for weight_key in ["weight", "linear.weight", "dense.weight"]:
|
|
if weight_key in state_dict:
|
|
weight_loader = getattr(
|
|
linear.weight, "weight_loader", default_weight_loader
|
|
)
|
|
weight_loader(linear.weight, state_dict[weight_key])
|
|
|
|
bias_key = weight_key.replace("weight", "bias")
|
|
if linear.bias is not None and bias_key in state_dict:
|
|
bias_loader = getattr(
|
|
linear.bias, "weight_loader", default_weight_loader
|
|
)
|
|
bias_loader(linear.bias, state_dict[bias_key])
|
|
return True
|
|
except Exception:
|
|
logger.exception("Failed to load %s", filename)
|
|
continue
|
|
|
|
return False
|
|
|
|
|
|
def _get_pooling_model_name(orig_model_name: str, pooling_suffix: str) -> str:
|
|
model_name = orig_model_name
|
|
|
|
for generate_suffix in _GENERATE_SUFFIXES:
|
|
model_name = model_name.removesuffix(generate_suffix)
|
|
|
|
return model_name + pooling_suffix
|
|
|
|
|
|
def _create_pooling_model_cls(orig_cls: _T) -> _T:
|
|
# Lazy import
|
|
from vllm.model_executor.layers.logits_processor import LogitsProcessor
|
|
from vllm.model_executor.layers.vocab_parallel_embedding import ParallelLMHead
|
|
|
|
from .utils import AutoWeightsLoader, StageMissingLayer, no_init_weights
|
|
|
|
class ModelForPooling(orig_cls, VllmModelForPooling):
|
|
is_pooling_model = True
|
|
|
|
def __init__(
|
|
self,
|
|
*,
|
|
vllm_config: "VllmConfig",
|
|
prefix: str = "",
|
|
**kwargs: Any,
|
|
) -> None:
|
|
with no_init_weights(
|
|
self,
|
|
lambda mod: StageMissingLayer("output", mod),
|
|
targets=(LogitsProcessor, ParallelLMHead),
|
|
):
|
|
super().__init__(vllm_config=vllm_config, prefix=prefix, **kwargs)
|
|
|
|
# Used by SEQ_CLS_LOAD_METHODS
|
|
self.vllm_config = vllm_config
|
|
|
|
# If the model already defines a pooler instance, don't overwrite it
|
|
pooler = getattr(self, "pooler", None)
|
|
if not pooler and supports_multimodal(self):
|
|
# Try to get the pooler from the LM backbone
|
|
language_model = self.get_language_model()
|
|
if hasattr(language_model, "pooler"):
|
|
pooler = language_model.pooler
|
|
|
|
if not pooler:
|
|
pooler = self._init_pooler(vllm_config, prefix=prefix)
|
|
|
|
self.pooler = pooler
|
|
|
|
def _init_pooler(
|
|
self,
|
|
vllm_config: "VllmConfig",
|
|
prefix: str = "",
|
|
) -> "Pooler":
|
|
raise NotImplementedError
|
|
|
|
def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]):
|
|
params_dict = dict(self.named_parameters())
|
|
|
|
# We support loading from both `*ForCausalLM` and `*Model`
|
|
candidate_prefixes = ["", "model."]
|
|
target_prefix = ""
|
|
|
|
seen_weights = list[tuple[str, torch.Tensor]]()
|
|
for name, loaded_weight in weights:
|
|
seen_weights.append((name, loaded_weight))
|
|
|
|
try:
|
|
target_prefix = next(
|
|
prefix
|
|
for prefix in candidate_prefixes
|
|
if prefix + name in params_dict
|
|
)
|
|
break
|
|
except StopIteration:
|
|
# The weight might not exist on the model
|
|
# (to be handled by AutoWeightsLoader)
|
|
pass
|
|
|
|
if target_prefix:
|
|
target_model = self
|
|
for attr in target_prefix.split("."):
|
|
if attr:
|
|
target_model = getattr(self, attr)
|
|
|
|
logger.info(
|
|
"Mapping weights to %s as they are "
|
|
"relative to this model instead of %s.",
|
|
target_model._get_name(),
|
|
self._get_name(),
|
|
)
|
|
|
|
mapped_weights = (
|
|
(target_prefix + name, weight)
|
|
for name, weight in (*seen_weights, *weights)
|
|
)
|
|
|
|
def default_load_weights(weights):
|
|
loader = AutoWeightsLoader(self)
|
|
return loader.load_weights(weights)
|
|
|
|
load_weights = getattr(super(), "load_weights", default_load_weights)
|
|
return load_weights(mapped_weights)
|
|
|
|
return ModelForPooling # type: ignore
|
|
|
|
|
|
def as_embedding_model(cls: _T) -> _T:
|
|
"""
|
|
Subclass an existing vLLM model to support embeddings.
|
|
|
|
By default, the embeddings of the whole prompt are extracted from the
|
|
normalized hidden state corresponding to the last token.
|
|
|
|
Note:
|
|
We assume that no extra layers are added to the original model;
|
|
please implement your own model if this is not the case.
|
|
"""
|
|
# Avoid modifying existing embedding models
|
|
if is_pooling_model(cls):
|
|
return cls
|
|
|
|
# Lazy import
|
|
from vllm.model_executor.layers.pooler import DispatchPooler
|
|
|
|
class ModelForEmbedding(_create_pooling_model_cls(cls)):
|
|
def _init_pooler(
|
|
self,
|
|
vllm_config: "VllmConfig",
|
|
prefix: str = "",
|
|
) -> "Pooler":
|
|
pooler_config = vllm_config.model_config.pooler_config
|
|
assert pooler_config is not None
|
|
|
|
return DispatchPooler.for_embedding(pooler_config)
|
|
|
|
ModelForEmbedding.__name__ = _get_pooling_model_name(cls.__name__, "ForEmbedding")
|
|
|
|
return ModelForEmbedding # type: ignore
|
|
|
|
|
|
def as_seq_cls_model(cls: _T) -> _T:
|
|
"""
|
|
Subclass an existing vLLM model to support classify and score tasks.
|
|
|
|
By default, the class probabilities are extracted from the softmaxed
|
|
hidden state corresponding to the last token.
|
|
|
|
Note:
|
|
We assume that the classification head is a single linear layer
|
|
stored as the attribute `score` of the top-level model;
|
|
please implement your own model if this is not the case.
|
|
"""
|
|
# Avoid modifying existing classification models
|
|
if is_pooling_model(cls):
|
|
return cls
|
|
|
|
# Lazy import
|
|
from vllm.model_executor.layers.linear import ReplicatedLinear
|
|
from vllm.model_executor.layers.pooler import DispatchPooler
|
|
from vllm.model_executor.models.interfaces import SupportsCrossEncoding
|
|
|
|
from .utils import maybe_prefix
|
|
|
|
class ModelForSequenceClassification(
|
|
_create_pooling_model_cls(cls), SupportsCrossEncoding
|
|
):
|
|
def _init_pooler(
|
|
self,
|
|
vllm_config: "VllmConfig",
|
|
prefix: str = "",
|
|
) -> "Pooler":
|
|
text_config = vllm_config.model_config.hf_config.get_text_config()
|
|
model_config = vllm_config.model_config
|
|
quant_config = vllm_config.quant_config
|
|
|
|
self.score = ReplicatedLinear(
|
|
model_config.get_hidden_size(),
|
|
text_config.num_labels,
|
|
bias=False,
|
|
params_dtype=vllm_config.model_config.head_dtype,
|
|
quant_config=quant_config,
|
|
return_bias=False,
|
|
prefix=maybe_prefix(prefix, "score"),
|
|
)
|
|
|
|
pooler_config = vllm_config.model_config.pooler_config
|
|
assert pooler_config is not None
|
|
|
|
return DispatchPooler.for_seq_cls(pooler_config, classifier=self.score)
|
|
|
|
def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]):
|
|
hf_config = self.config
|
|
text_config = hf_config.get_text_config()
|
|
tokens = getattr(
|
|
hf_config,
|
|
"classifier_from_token",
|
|
getattr(text_config, "classifier_from_token", None),
|
|
)
|
|
method = getattr(hf_config, "method", getattr(text_config, "method", None))
|
|
|
|
def auto_set_score_bias(weights):
|
|
for name, weight in weights:
|
|
if name == "score.bias":
|
|
device = self.score.weight.device
|
|
dtype = self.score.weight.dtype
|
|
bias = weight.to(device).to(dtype)
|
|
self.score.bias = torch.nn.Parameter(bias)
|
|
self.score.skip_bias_add = False
|
|
else:
|
|
yield name, weight
|
|
|
|
weights = auto_set_score_bias(weights)
|
|
if tokens is None and method is None:
|
|
return super().load_weights(weights)
|
|
else:
|
|
# Online convert ForCausalLM into
|
|
# ForSequenceClassification model.
|
|
return seq_cls_model_loader(self, weights)
|
|
|
|
ModelForSequenceClassification.__name__ = _get_pooling_model_name(
|
|
cls.__name__, "ForSequenceClassification"
|
|
)
|
|
|
|
return ModelForSequenceClassification # type: ignore
|
|
|
|
|
|
class SequenceClassificationConfig(VerifyAndUpdateConfig):
|
|
@staticmethod
|
|
def verify_and_update_config(vllm_config: "VllmConfig") -> None:
|
|
hf_config = vllm_config.model_config.hf_config
|
|
text_config = hf_config.get_text_config()
|
|
method = getattr(hf_config, "method", getattr(text_config, "method", None))
|
|
tokens = getattr(
|
|
hf_config,
|
|
"classifier_from_token",
|
|
getattr(text_config, "classifier_from_token", None),
|
|
)
|
|
|
|
if method is None:
|
|
return
|
|
|
|
assert tokens is not None
|
|
assert method in SEQ_CLS_LOAD_METHODS, f"method {method} not supported"
|
|
|
|
if method == "from_2_way_softmax":
|
|
assert len(tokens) == 2
|
|
hf_config.num_labels = 1
|
|
text_config.num_labels = 1
|
|
else:
|
|
hf_config.num_labels = len(tokens)
|
|
text_config.num_labels = len(tokens)
|
|
|
|
# `llm as reranker` defaults to not using separating token.
|
|
use_sep_token = getattr(text_config, "use_sep_token", False)
|
|
text_config.use_sep_token = use_sep_token
|
|
|
|
|
|
def _get_language_model_for_seq_cls(model) -> nn.Module:
|
|
"""
|
|
Get the language model component for sequence classification conversion.
|
|
For VLMs, returns the inner language model. For standard LLMs, returns model itself.
|
|
"""
|
|
if supports_multimodal(model):
|
|
try:
|
|
lm = model.get_language_model()
|
|
if lm is not model:
|
|
return lm
|
|
except Exception:
|
|
pass
|
|
|
|
for attr_name in ("language_model", "lm", "text_model"):
|
|
if hasattr(model, attr_name):
|
|
candidate = getattr(model, attr_name)
|
|
if (
|
|
isinstance(candidate, nn.Module)
|
|
and candidate is not model
|
|
and hasattr(candidate, "model")
|
|
):
|
|
return candidate
|
|
|
|
for name, child in model.named_children():
|
|
child_name = type(child).__name__
|
|
if ("ForCausalLM" in child_name or "LMHead" in child_name) and hasattr(
|
|
child, "model"
|
|
):
|
|
return child
|
|
|
|
return model
|
|
|
|
|
|
@contextmanager
|
|
def _disable_seq_cls_loading_on_inner_model(language_model, is_vlm: bool):
|
|
"""
|
|
Context manager to temporarily disable sequence classification loading
|
|
on inner VLM models to prevent recursive seq_cls_model_loader calls.
|
|
"""
|
|
if not is_vlm:
|
|
yield
|
|
return
|
|
|
|
inner_hf_config = getattr(language_model, "config", None)
|
|
if inner_hf_config is None:
|
|
yield
|
|
return
|
|
|
|
inner_text_config = inner_hf_config.get_text_config()
|
|
original_method = getattr(inner_text_config, "method", None)
|
|
original_tokens = getattr(inner_text_config, "classifier_from_token", None)
|
|
original_hf_tokens = getattr(inner_hf_config, "classifier_from_token", None)
|
|
|
|
try:
|
|
if original_method is not None:
|
|
inner_text_config.method = None
|
|
if original_tokens is not None:
|
|
inner_text_config.classifier_from_token = None
|
|
if original_hf_tokens is not None:
|
|
inner_hf_config.classifier_from_token = None
|
|
yield
|
|
finally:
|
|
if original_method is not None:
|
|
inner_text_config.method = original_method
|
|
if original_tokens is not None:
|
|
inner_text_config.classifier_from_token = original_tokens
|
|
if original_hf_tokens is not None:
|
|
inner_hf_config.classifier_from_token = original_hf_tokens
|
|
|
|
|
|
def load_weights_using_from_2_way_softmax(
|
|
model, weights: Iterable[tuple[str, torch.Tensor]]
|
|
):
|
|
# refer to https://huggingface.co/Qwen/Qwen3-Reranker-0.6B/discussions/3
|
|
from vllm.model_executor.layers.vocab_parallel_embedding import ParallelLMHead
|
|
from vllm.model_executor.model_loader.weight_utils import default_weight_loader
|
|
|
|
model_config = model.vllm_config.model_config
|
|
quant_config = model.vllm_config.quant_config
|
|
hf_config = model.config
|
|
text_config = hf_config.get_text_config()
|
|
|
|
tokens = getattr(
|
|
hf_config,
|
|
"classifier_from_token",
|
|
getattr(text_config, "classifier_from_token", []),
|
|
)
|
|
tokens = cast(list[int], tokens)
|
|
assert len(tokens) == 2
|
|
|
|
language_model = _get_language_model_for_seq_cls(model)
|
|
is_vlm = language_model is not model
|
|
using_vlm_head = is_vlm and hasattr(language_model, "score")
|
|
|
|
language_model.lm_head = ParallelLMHead(
|
|
text_config.vocab_size, text_config.hidden_size, quant_config=quant_config
|
|
)
|
|
if text_config.tie_word_embeddings:
|
|
# embed_tokens is the assumed name for input embeddings. If the model does not
|
|
# have this attribute, we fall back to get_input_embeddings(), which is used by
|
|
# the Transformers modeling backend.
|
|
text_backbone = language_model.model
|
|
embed_tokens = (
|
|
text_backbone.embed_tokens
|
|
if hasattr(text_backbone, "embed_tokens")
|
|
else text_backbone.get_input_embeddings()
|
|
)
|
|
language_model.lm_head = language_model.lm_head.tie_weights(embed_tokens)
|
|
|
|
with _disable_seq_cls_loading_on_inner_model(language_model, is_vlm):
|
|
# ModelForPooling is dynamically defined inside the _create_pooling_model_cls
|
|
# function, so we need use this hacky method to obtain it.
|
|
pooling_model_cls = next(
|
|
x for x in type(model).__mro__ if x.__name__ == "ModelForPooling"
|
|
)
|
|
loaded_weights = pooling_model_cls.load_weights(model, weights)
|
|
|
|
from vllm.tokenizers import get_tokenizer
|
|
|
|
tokenizer = get_tokenizer(
|
|
model_config.tokenizer,
|
|
revision=model_config.tokenizer_revision,
|
|
tokenizer_mode=model_config.tokenizer_mode,
|
|
trust_remote_code=model_config.trust_remote_code,
|
|
)
|
|
|
|
false_id = tokenizer.convert_tokens_to_ids(tokens[0])
|
|
true_id = tokenizer.convert_tokens_to_ids(tokens[1])
|
|
lm_head_weight = language_model.lm_head.weight
|
|
score_weight = lm_head_weight.data[[true_id]].to(
|
|
torch.float32
|
|
) - lm_head_weight.data[[false_id]].to(torch.float32)
|
|
|
|
score_layer = language_model.score if using_vlm_head else model.score
|
|
param = score_layer.weight
|
|
weight_loader = getattr(param, "weight_loader", default_weight_loader)
|
|
weight_loader(param, score_weight)
|
|
|
|
del language_model.lm_head
|
|
|
|
score_weight_name = (
|
|
"language_model.score.weight" if using_vlm_head else "score.weight"
|
|
)
|
|
loaded_weights.add(score_weight_name)
|
|
|
|
lm_head_name = "lm_head.weight"
|
|
if hf_to_vllm_mapper := getattr(model, "hf_to_vllm_mapper", None):
|
|
lm_head_name = hf_to_vllm_mapper._map_name(lm_head_name)
|
|
loaded_weights.discard(lm_head_name)
|
|
return loaded_weights
|
|
|
|
|
|
def load_weights_no_post_processing(model, weights: Iterable[tuple[str, torch.Tensor]]):
|
|
from vllm.model_executor.layers.vocab_parallel_embedding import ParallelLMHead
|
|
from vllm.model_executor.model_loader.weight_utils import default_weight_loader
|
|
|
|
model_config = model.vllm_config.model_config
|
|
quant_config = model.vllm_config.quant_config
|
|
text_config = model.config.get_text_config()
|
|
|
|
tokens = getattr(text_config, "classifier_from_token", [])
|
|
tokens = cast(list[int], tokens)
|
|
assert len(tokens) > 0
|
|
|
|
language_model = _get_language_model_for_seq_cls(model)
|
|
is_vlm = language_model is not model
|
|
using_vlm_head = is_vlm and hasattr(language_model, "score")
|
|
|
|
language_model.lm_head = ParallelLMHead(
|
|
text_config.vocab_size, text_config.hidden_size, quant_config=quant_config
|
|
)
|
|
if text_config.tie_word_embeddings:
|
|
# embed_tokens is the assumed name for input embeddings. If the model does not
|
|
# have this attribute, we fall back to get_input_embeddings(), which is used by
|
|
# the Transformers modeling backend.
|
|
text_backbone = language_model.model
|
|
embed_tokens = (
|
|
text_backbone.embed_tokens
|
|
if hasattr(text_backbone, "embed_tokens")
|
|
else text_backbone.get_input_embeddings()
|
|
)
|
|
language_model.lm_head = language_model.lm_head.tie_weights(embed_tokens)
|
|
|
|
with _disable_seq_cls_loading_on_inner_model(language_model, is_vlm):
|
|
pooling_model_cls = next(
|
|
x for x in type(model).__mro__ if x.__name__ == "ModelForPooling"
|
|
)
|
|
# Skip ModelForSequenceClassification in MRO to avoid infinite recursion
|
|
loaded_weights = pooling_model_cls.load_weights(model, weights)
|
|
|
|
from vllm.tokenizers import get_tokenizer
|
|
|
|
tokenizer = get_tokenizer(
|
|
model_config.tokenizer,
|
|
revision=model_config.tokenizer_revision,
|
|
tokenizer_mode=model_config.tokenizer_mode,
|
|
trust_remote_code=model_config.trust_remote_code,
|
|
)
|
|
|
|
token_ids = [tokenizer.convert_tokens_to_ids(t) for t in tokens]
|
|
score_weight = language_model.lm_head.weight.data[token_ids]
|
|
|
|
score_layer = language_model.score if using_vlm_head else model.score
|
|
param = score_layer.weight
|
|
weight_loader = getattr(param, "weight_loader", default_weight_loader)
|
|
weight_loader(param, score_weight)
|
|
|
|
del language_model.lm_head
|
|
|
|
score_weight_name = (
|
|
"language_model.score.weight" if using_vlm_head else "score.weight"
|
|
)
|
|
loaded_weights.add(score_weight_name)
|
|
|
|
lm_head_name = "lm_head.weight"
|
|
if hf_to_vllm_mapper := getattr(model, "hf_to_vllm_mapper", None):
|
|
lm_head_name = hf_to_vllm_mapper._map_name(lm_head_name)
|
|
loaded_weights.discard(lm_head_name)
|
|
return loaded_weights
|
|
|
|
|
|
SEQ_CLS_LOAD_METHODS = {
|
|
"from_2_way_softmax": load_weights_using_from_2_way_softmax,
|
|
"no_post_processing": load_weights_no_post_processing,
|
|
}
|
|
|
|
|
|
def seq_cls_model_loader(model, weights: Iterable[tuple[str, torch.Tensor]]):
|
|
# Online convert ForCausalLM into ForSequenceClassification model.
|
|
# - from_2_way_softmax:
|
|
# - Qwen3ForCausalLM
|
|
# - Qwen3-Reranker
|
|
# - Qwen2ForCausalLM
|
|
# - mxbai-rerank-v2
|
|
# - no_post_processing:
|
|
# - GemmaForCausalLM
|
|
# - bge-reranker-v2-gemma
|
|
|
|
hf_config = model.vllm_config.model_config.hf_config
|
|
text_config = hf_config.get_text_config()
|
|
method = getattr(hf_config, "method", getattr(text_config, "method", None))
|
|
assert method in SEQ_CLS_LOAD_METHODS, f"method {method} not supported"
|
|
return SEQ_CLS_LOAD_METHODS[method](model, weights)
|