This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
0ed27ef66ca7d7fec3c9bb8e33709ade1c73b4ad
vllm
/
vllm
/
model_executor
/
layers
/
quantization
/
compressed_tensors
History
a2q1p
0ed27ef66c
Fix: Spelling of inference (
#17387
)
2025-04-29 09:23:39 -07:00
..
schemes
[Misc] Update
compressed-tensors
WNA16 to support zero-points (
#14211
)
2025-04-15 07:33:51 -06:00
__init__.py
[Kernel] Initial Activation Quantization Support (
#4525
)
2024-05-23 21:29:18 +00:00
compressed_tensors_moe.py
[FEAT] [ROCm]: AITER Fused MOE V1 Support (
#16752
)
2025-04-25 11:06:50 +08:00
compressed_tensors.py
Fix: Spelling of inference (
#17387
)
2025-04-29 09:23:39 -07:00
triton_scaled_mm.py
[Misc] Add SPDX-License-Identifier headers to python source files (
#12628
)
2025-02-02 11:58:18 -08:00
utils.py
[Model][Quant] Fix GLM, Fix fused module mappings for quantization (
#12634
)
2025-02-05 05:32:06 +00:00