This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
9103ed16967390f3bbd6df104dcd162db43d3148
vllm
/
vllm
/
model_executor
History
Andy Liu
0dd63639be
[MTP][GLM][Bugfix] Fixed .weight_scale loading logic that dropped MTP prediction accuracy with fp8+mtp (
#32101
)
...
Signed-off-by: Andy Liu <
andyliu@roblox.com
>
2026-01-10 23:14:54 -08:00
..
layers
[Bugfix][Quantization] Ensure input contiguity in per_token_quant_int8 (
#31637
)
2026-01-10 12:40:02 -08:00
model_loader
[Core] Use weights_only=True with torch.load (
#32045
)
2026-01-10 00:28:57 +00:00
models
[MTP][GLM][Bugfix] Fixed .weight_scale loading logic that dropped MTP prediction accuracy with fp8+mtp (
#32101
)
2026-01-10 23:14:54 -08:00
warmup
[UX] Reduce DeepGEMM warmup log output to single progress bar (
#30903
)
2025-12-17 20:21:51 -08:00
__init__.py
[Platform] Deprecate seed_everything (
#31659
)
2026-01-04 18:34:04 -08:00
custom_op.py
[Doc] Add developer guide for CustomOp (
#30886
)
2026-01-09 16:21:11 +00:00
parameter.py
[Docs] Replace
rst
style double-backtick with
md
single-backtick (
#27091
)
2025-10-17 02:47:34 -07:00
utils.py
[Platform] Deprecate seed_everything (
#31659
)
2026-01-04 18:34:04 -08:00