This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
4ef0bb1fcfc6e948b8c9ac196ca5497bd2c9e157
vllm
/
vllm
/
attention
History
Liangfu Chen
d2b58ca203
[Neuron][kernel] Fuse kv cache into a single tensor (
#15911
)
...
Signed-off-by: Liangfu Chen <
liangfc@amazon.com
>
2025-04-03 09:51:32 -07:00
..
backends
[ROCM][KERNEL] Paged attention for V1 (
#15720
)
2025-04-02 19:48:00 -07:00
ops
[Neuron][kernel] Fuse kv cache into a single tensor (
#15911
)
2025-04-03 09:51:32 -07:00
__init__.py
[Attention] Flash Attention 3 - fp8 (
#14570
)
2025-03-20 01:14:20 -04:00
layer.py
[TPU][V1] MHA Pallas backend (
#15288
)
2025-03-21 08:50:39 -07:00
selector.py
Correct capitalisation:
VLLM
->
vLLM
(
#14562
)
2025-03-10 16:36:21 +00:00