This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
551ef1631a98d60fe9e82f0282e49c4a59a7887b
vllm
/
vllm
/
attention
History
Chendi.Xue
dec197e3e5
Quick Fix by adding conditional import for flash_attn_varlen_func in flash_attn (
#20143
)
...
Signed-off-by: Chendi.Xue <
chendi.xue@intel.com
>
2025-06-27 05:48:13 +00:00
..
backends
[Kernel] mark TorchSDPABackend swap_blocks NotImplementedError (
#19749
)
2025-06-20 18:18:11 +00:00
ops
[TPU] add kv cache update kernel (
#19928
)
2025-06-26 10:01:37 -07:00
utils
Quick Fix by adding conditional import for flash_attn_varlen_func in flash_attn (
#20143
)
2025-06-27 05:48:13 +00:00
__init__.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00
layer.py
[Bugfix][V1][ROCm] Fix AITER Flash Attention Backend (Fix API Break and Local Attention Logic: affecting Llama4) (
#19904
)
2025-06-26 12:42:31 +00:00
selector.py
[Misc] Add SPDX-FileCopyrightText (
#19100
)
2025-06-03 11:20:17 -07:00