[Misc] update fp8 to use vLLMParameter (#7437)

This commit is contained in:
Dipika Sikka
2024-08-22 08:36:18 -04:00
committed by GitHub
parent 55d63b1211
commit 955b5191c9
4 changed files with 51 additions and 17 deletions

View File

@@ -15,3 +15,4 @@ compressed-tensors, nm-testing/Phi-3-mini-128k-instruct-FP8, main
compressed-tensors, neuralmagic/Phi-3-medium-128k-instruct-quantized.w4a16, main
awq, casperhansen/mixtral-instruct-awq, main
awq_marlin, casperhansen/mixtral-instruct-awq, main
fp8, neuralmagic/Meta-Llama-3-8B-Instruct-FP8-KV, main