Logo
Explore Help
Register Sign In
biondizzle/vllm
1
0
Fork 0
You've already forked vllm
Code Issues Pull Requests Actions 2 Packages Projects Releases Wiki Activity
Files
5963b98b465007e3cfb0d39447e4459a8afa96dc
vllm/vllm/model_executor/layers/quantization/compressed_tensors
History
bnellnm 5963b98b46 [Kernel] Delegate construction of FusedMoEQuantConfig to FusedMoEMethodBase subclasses (#22537)
Signed-off-by: Bill Nell <bnell@redhat.com>
2025-09-17 17:43:31 -06:00
..
schemes
[quantization] use channel scales for w4a8 + misc fixes (#23570)
2025-08-26 18:23:23 -07:00
transform
[Transform] Deterministic Hadacore Transforms (#24106)
2025-09-15 12:59:31 -06:00
__init__.py
[Kernel] Initial Activation Quantization Support (#4525)
2024-05-23 21:29:18 +00:00
compressed_tensors_moe.py
[Kernel] Delegate construction of FusedMoEQuantConfig to FusedMoEMethodBase subclasses (#22537)
2025-09-17 17:43:31 -06:00
compressed_tensors.py
[Transform] Deterministic Hadacore Transforms (#24106)
2025-09-15 12:59:31 -06:00
triton_scaled_mm.py
[AMD][Kernel][BugFix] fix test_rocm_compressed_tensors_w8a8 for rocm (#19509)
2025-06-12 07:14:24 +00:00
utils.py
[Doc]: fix typos in Python comments (#24093)
2025-09-02 21:05:45 -07:00
Powered by Gitea Version: 1.25.2 Page: 407ms Template: 2ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API