This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
5963b98b465007e3cfb0d39447e4459a8afa96dc
vllm
/
vllm
/
model_executor
/
layers
/
quantization
/
compressed_tensors
History
bnellnm
5963b98b46
[Kernel] Delegate construction of FusedMoEQuantConfig to FusedMoEMethodBase subclasses (
#22537
)
...
Signed-off-by: Bill Nell <
bnell@redhat.com
>
2025-09-17 17:43:31 -06:00
..
schemes
[quantization] use channel scales for w4a8 + misc fixes (
#23570
)
2025-08-26 18:23:23 -07:00
transform
[Transform] Deterministic Hadacore Transforms (
#24106
)
2025-09-15 12:59:31 -06:00
__init__.py
[Kernel] Initial Activation Quantization Support (
#4525
)
2024-05-23 21:29:18 +00:00
compressed_tensors_moe.py
[Kernel] Delegate construction of FusedMoEQuantConfig to FusedMoEMethodBase subclasses (
#22537
)
2025-09-17 17:43:31 -06:00
compressed_tensors.py
[Transform] Deterministic Hadacore Transforms (
#24106
)
2025-09-15 12:59:31 -06:00
triton_scaled_mm.py
[AMD][Kernel][BugFix] fix test_rocm_compressed_tensors_w8a8 for rocm (
#19509
)
2025-06-12 07:14:24 +00:00
utils.py
[Doc]: fix typos in Python comments (
#24093
)
2025-09-02 21:05:45 -07:00