This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
4e824d1c835d9b57db621297e8d9119bfc32fb2e
vllm
/
vllm
/
model_executor
/
models
/
transformers
History
Yufeng He
ec2280611a
[Bugfix] Fix RoBERTa position_ids accumulation on CUDA graph padding (
#37884
)
2026-03-23 15:15:12 +00:00
..
__init__.py
[Docs] Update the name of
Transformers backend
->
Transformers modeling backend
(
#28725
)
2025-11-14 16:34:14 +00:00
base.py
Fix pipeline parallel with multimodal models with the Transformers modelling backend (
#37057
)
2026-03-16 10:20:37 +00:00
causal.py
Fix pipeline parallel with multimodal models with the Transformers modelling backend (
#37057
)
2026-03-16 10:20:37 +00:00
legacy.py
[Bugfix] Fix RoBERTa position_ids accumulation on CUDA graph padding (
#37884
)
2026-03-23 15:15:12 +00:00
moe.py
Enable loading of fused expert weights in the Transformers modelling backend (
#36997
)
2026-03-14 07:01:06 +00:00
multimodal.py
Fix text only inputs for MRoPE models with the Transformers modelling backend (
#37055
)
2026-03-16 10:31:16 +00:00
pooling.py
[Doc] Fix duplicate words in comments (
#36713
)
2026-03-10 21:28:31 -07:00
utils.py
Replace
nn.ConvNd
with vLLM's
ConvNdLayer
for Transformers modeling backend (
#31498
)
2025-12-29 16:20:01 +00:00