This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
b9a1c4c8a2430555f1959938b87b10640c751e57
vllm
/
benchmarks
/
kernels
/
benchmark_trtllm_decode_attention.py
elvischenv
bba1042c6f
[Flashinfer] Support Flashinfer TRTLLM FP8-qkv BF16/FP16-out Attention Kernel (
#23647
)
...
Signed-off-by: elvischenv <
219235043+elvischenv@users.noreply.github.com
>
2025-09-08 20:53:07 -07:00
9.0 KiB
Raw
Blame
History
View Raw
Reference in New Issue
View Git Blame
Copy Permalink