[Misc] Remove redundant config definitions (#21891)

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
Cyrus Leung
2025-07-30 14:54:18 +08:00
committed by GitHub
parent 6f8d261882
commit 2ca5f82c2a
23 changed files with 54 additions and 1910 deletions

View File

@@ -8,6 +8,7 @@ from typing import Optional
import torch
import torch.nn as nn
from transformers import PretrainedConfig
from vllm.attention.layer import MultiHeadAttention
from vllm.distributed import get_tensor_model_parallel_world_size
@@ -20,13 +21,12 @@ from vllm.model_executor.layers.linear import (MergedColumnParallelLinear,
from vllm.model_executor.layers.quantization.base_config import (
QuantizationConfig)
from vllm.model_executor.model_loader.weight_utils import default_weight_loader
from vllm.transformers_utils.configs.ovis import AIMv2Config
class AIMv2SwiGLUFFN(nn.Module):
def __init__(self, config: AIMv2Config, quant_config: QuantizationConfig,
prefix: str):
def __init__(self, config: PretrainedConfig,
quant_config: QuantizationConfig, prefix: str):
super().__init__()
hidden_features = config.intermediate_size
in_features = config.hidden_size
@@ -57,7 +57,7 @@ class AIMv2SwiGLUFFN(nn.Module):
class AIMv2PatchEmbed(nn.Module):
def __init__(self, config: AIMv2Config):
def __init__(self, config: PretrainedConfig):
super().__init__()
self.proj = nn.Conv2d(
config.num_channels,
@@ -75,7 +75,7 @@ class AIMv2PatchEmbed(nn.Module):
class AIMv2ViTPreprocessor(nn.Module):
def __init__(self, config: AIMv2Config):
def __init__(self, config: PretrainedConfig):
super().__init__()
num_patches = (config.image_size // config.patch_size)**2
@@ -93,8 +93,8 @@ class AIMv2ViTPreprocessor(nn.Module):
class AIMv2Attention(nn.Module):
def __init__(self, config: AIMv2Config, quant_config: QuantizationConfig,
prefix: str):
def __init__(self, config: PretrainedConfig,
quant_config: QuantizationConfig, prefix: str):
super().__init__()
self.config = config
self.embed_dim = config.hidden_size
@@ -141,8 +141,8 @@ class AIMv2Attention(nn.Module):
class AIMv2Block(nn.Module):
def __init__(self, config: AIMv2Config, quant_config: QuantizationConfig,
prefix: str):
def __init__(self, config: PretrainedConfig,
quant_config: QuantizationConfig, prefix: str):
super().__init__()
self.attn = AIMv2Attention(config,
quant_config=quant_config,
@@ -163,7 +163,7 @@ class AIMv2Transformer(nn.Module):
def __init__(
self,
config: AIMv2Config,
config: PretrainedConfig,
quant_config: QuantizationConfig,
*,
require_post_norm: Optional[bool] = None,
@@ -193,7 +193,7 @@ class AIMv2Transformer(nn.Module):
class AIMv2Model(torch.nn.Module):
def __init__(self,
config: AIMv2Config,
config: PretrainedConfig,
quant_config: QuantizationConfig,
*,
require_post_norm: Optional[bool] = None,