Bump transformers version for Llama 3.1 hotfix and patch Chameleon (#6690)

This commit is contained in:
Roger Wang
2024-07-23 13:47:48 -07:00
committed by GitHub
parent 507ef787d8
commit 1bedf210e3
7 changed files with 32 additions and 177 deletions

View File

@@ -1,5 +1,3 @@
from vllm.transformers_utils.configs.chameleon import (ChameleonConfig,
ChameleonVQVAEConfig)
from vllm.transformers_utils.configs.chatglm import ChatGLMConfig
from vllm.transformers_utils.configs.dbrx import DbrxConfig
# RWConfig is for the original tiiuae/falcon-40b(-instruct) and
@@ -12,8 +10,6 @@ from vllm.transformers_utils.configs.mlp_speculator import MLPSpeculatorConfig
from vllm.transformers_utils.configs.mpt import MPTConfig
__all__ = [
"ChameleonConfig",
"ChameleonVQVAEConfig",
"ChatGLMConfig",
"DbrxConfig",
"MPTConfig",