This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
1e8438a89a6453a6b1ba28798bb9c51d6364ed96
vllm
/
vllm
/
model_executor
History
Eldar Kurtić
1e8438a89a
[Llama4,CI] Bring back Llama-4 bug fixes, and also fix Maverick tests (
#35033
)
...
Signed-off-by: Eldar Kurtic <
you@example.com
> Co-authored-by: Eldar Kurtic <
you@example.com
>
2026-02-23 09:04:34 -05:00
..
layers
[XPU] allow TORCH_SDPA/TRITON_ATTN as XPU vit Backend (
#35010
)
2026-02-23 05:06:44 -08:00
model_loader
[Quantization] - Added uses_meta_device_weights to quant config (
#34645
)
2026-02-17 23:43:44 -08:00
models
[Llama4,CI] Bring back Llama-4 bug fixes, and also fix Maverick tests (
#35033
)
2026-02-23 09:04:34 -05:00
warmup
[Kernel] Add KernelConfig flag to enable/disable FlashInfer autotune (
#34006
)
2026-02-07 05:24:44 -08:00
__init__.py
[Platform] Deprecate seed_everything (
#31659
)
2026-01-04 18:34:04 -08:00
custom_op.py
[Model Bash][DSR1] Add selective dynamic shape marking for CustomOp (
#34900
)
2026-02-21 19:28:01 -05:00
parameter.py
[QeRL] Layerwise Reloading (
#32133
)
2026-01-30 08:50:05 -07:00
utils.py
[BugFix] Fix EPLB fail for MoeFP4 model with Marlin backend (
#33262
)
2026-01-29 16:52:11 +08:00