This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
554af9228df620a63d4736240a8f76a64a675f4d
vllm
/
vllm
/
entrypoints
History
Zijin Xiao
554af9228d
[Bugfix] use AF_INET6 for OpenAI Compatible Server with ipv6 (
#9583
)
...
Signed-off-by: xiaozijin <
xiaozijin@bytedance.com
>
2024-11-14 16:38:53 -08:00
..
openai
[Bugfix] use AF_INET6 for OpenAI Compatible Server with ipv6 (
#9583
)
2024-11-14 16:38:53 -08:00
__init__.py
Change the name to vLLM (
#150
)
2023-06-17 03:07:40 -07:00
api_server.py
bugfix: fix the bug that stream generate not work (
#2756
)
2024-11-09 10:09:48 +00:00
chat_utils.py
Online video support for VLMs (
#10020
)
2024-11-07 20:25:59 +00:00
launcher.py
[Core][Bugfix][Perf] Introduce
MQLLMEngine
to avoid
asyncio
OH (
#8157
)
2024-09-18 13:56:58 +00:00
llm.py
[V1]
AsyncLLM
Implementation (
#9826
)
2024-11-11 23:05:38 +00:00
logger.py
[Frontend] API support for beam search (
#9087
)
2024-10-05 23:39:03 -07:00