This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
f6227c22ab8976a24913122874c24624102da1b4
vllm
/
vllm
/
model_executor
/
layers
/
quantization
/
kernels
History
czhu-cohere
f6227c22ab
[Kernel]Support W4A8 Grouped GEMM on Hopper (
#29691
)
...
Signed-off-by: czhu-cohere <
conway.zhu@cohere.com
>
2025-12-08 19:29:06 -08:00
..
mixed_precision
[Kernel]Support W4A8 Grouped GEMM on Hopper (
#29691
)
2025-12-08 19:29:06 -08:00
scaled_mm
[ROCM] Fix ROCm warnings, environment flag access, and GEMM kernel naming for consistency in
_aiter_ops.py
(
#28464
)
2025-11-12 21:46:57 +00:00
__init__.py
[TPU][Quantization] TPU
W8A8
(
#11785
)
2025-01-08 19:33:29 +00:00