This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
ac2f3f7fee93cf9cd97c0078e362feab7b6c8299
vllm
/
vllm
/
engine
History
Joe Runde
ac2f3f7fee
[Bugfix] Validate lora adapters to avoid crashing server (
#11727
)
...
Signed-off-by: Joe Runde <
Joseph.Runde@ibm.com
> Co-authored-by: Jee Jee Li <
pandaleefree@gmail.com
>
2025-01-10 15:56:36 +08:00
..
multiprocessing
[Bugfix] Validate lora adapters to avoid crashing server (
#11727
)
2025-01-10 15:56:36 +08:00
output_processor
[Doc][2/N] Reorganize Models and Usage sections (
#11755
)
2025-01-06 21:40:31 +08:00
__init__.py
Change the name to vLLM (
#150
)
2023-06-17 03:07:40 -07:00
arg_utils.py
[Bugfix] Significant performance drop on CPUs with --num-scheduler-steps > 1 (
#11794
)
2025-01-07 16:15:50 -08:00
async_llm_engine.py
[Bugfix] Validate lora adapters to avoid crashing server (
#11727
)
2025-01-10 15:56:36 +08:00
async_timeout.py
[Bugfix] AsyncLLMEngine hangs with asyncio.run (
#5654
)
2024-06-19 13:57:12 -07:00
llm_engine.py
[Bugfix] Last token measurement fix (
#11376
)
2024-12-28 11:34:46 +08:00
metrics_types.py
monitor metrics of tokens per step using cudagraph batchsizes (
#11031
)
2024-12-09 22:35:36 -08:00
metrics.py
monitor metrics of tokens per step using cudagraph batchsizes (
#11031
)
2024-12-09 22:35:36 -08:00
protocol.py
[Bugfix] Validate lora adapters to avoid crashing server (
#11727
)
2025-01-10 15:56:36 +08:00