This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
05a015d6a52e6093f1ac0b76ada5b7da4d6a5671
vllm
/
vllm
/
v1
/
attention
/
backends
History
Yong Hoon Shin
05a015d6a5
Add warning for Attention backends that do not support irope yet (
#16212
)
2025-04-08 03:59:26 +00:00
..
mla
[V1] Enable V1 Fp8 cache for FA3 in the oracle (
#15191
)
2025-03-23 15:07:04 -07:00
__init__.py
[V1] Implement vLLM V1 [1/N] (
#9289
)
2024-10-22 01:24:07 -07:00
flash_attn.py
Upstream Llama4 Support to Main (
#16113
)
2025-04-07 08:06:27 -07:00
pallas.py
Add warning for Attention backends that do not support irope yet (
#16212
)
2025-04-08 03:59:26 +00:00
triton_attn.py
Upstream Llama4 Support to Main (
#16113
)
2025-04-07 08:06:27 -07:00