This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
e919d6f549f4da22fa60ea394f00aaf93ef23aa0
vllm
/
vllm
/
attention
History
Burkhard Ringlein
6d80ae83e1
[Bugfix] Fixing division by zero in triton_attn if query_heads/kv_heads > 16 (
#23424
)
...
Signed-off-by: Burkhard Ringlein <
ngl@zurich.ibm.com
>
2025-09-03 15:01:09 +00:00
..
backends
[Docs] Fix warnings in
mkdocs build
(
#23649
)
2025-08-26 18:19:23 +00:00
layers
[Misc] Modify CacheConfig import (
#23459
)
2025-08-23 06:05:27 +00:00
ops
[Bugfix] Fixing division by zero in triton_attn if query_heads/kv_heads > 16 (
#23424
)
2025-09-03 15:01:09 +00:00
utils
[MISC] Add init files for python package (
#20908
)
2025-07-15 12:16:33 +00:00
__init__.py
Remove duplicate entry in vllm.attention.__all__ (
#23296
)
2025-08-20 17:14:59 -07:00
layer.py
[XPU] Add xpu torch.compile support (
#22609
)
2025-08-27 05:33:27 +00:00
selector.py
[gpt-oss] Enable gpt-oss on ampere (
#22714
)
2025-08-12 03:21:44 -07:00