This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
699bca76c00b81ba6c7ead38fed01712f5f56aa1
vllm
/
vllm
/
attention
History
Roger Wang
0ff70821c9
[Core] Deprecate
xformers
(
#29262
)
...
Signed-off-by: Roger Wang <
hey@rogerw.io
>
2025-11-24 04:18:55 +00:00
..
backends
[Core] Deprecate
xformers
(
#29262
)
2025-11-24 04:18:55 +00:00
layers
[Bugfix] Fix ChunkedLocalAttention CUDA Graph setting (
#28739
)
2025-11-14 14:14:46 -08:00
ops
[Core] Deprecate
xformers
(
#29262
)
2025-11-24 04:18:55 +00:00
utils
[Misc] Refactor Attention kv transfer methods into decorator (
#27816
)
2025-11-12 16:05:44 +00:00
__init__.py
[Model][Mamba] Add selector for mamba attention backend and make it pluggable for other device (
#26487
)
2025-11-19 16:24:55 +00:00
layer.py
[Core] Deprecate
xformers
(
#29262
)
2025-11-24 04:18:55 +00:00
selector.py
[Core] Deprecate
xformers
(
#29262
)
2025-11-24 04:18:55 +00:00