[cli] complete vllm cli help message (#31226)
Signed-off-by: Andy Xie <andy.xning@gmail.com>
This commit is contained in:
@@ -32,6 +32,7 @@ class BenchmarkSubcommand(CLISubcommand):
|
|||||||
) -> FlexibleArgumentParser:
|
) -> FlexibleArgumentParser:
|
||||||
bench_parser = subparsers.add_parser(
|
bench_parser = subparsers.add_parser(
|
||||||
self.name,
|
self.name,
|
||||||
|
help=self.help,
|
||||||
description=self.help,
|
description=self.help,
|
||||||
usage=f"vllm {self.name} <bench_type> [options]",
|
usage=f"vllm {self.name} <bench_type> [options]",
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -66,7 +66,11 @@ class ServeSubcommand(CLISubcommand):
|
|||||||
self, subparsers: argparse._SubParsersAction
|
self, subparsers: argparse._SubParsersAction
|
||||||
) -> FlexibleArgumentParser:
|
) -> FlexibleArgumentParser:
|
||||||
serve_parser = subparsers.add_parser(
|
serve_parser = subparsers.add_parser(
|
||||||
self.name, description=DESCRIPTION, usage="vllm serve [model_tag] [options]"
|
self.name,
|
||||||
|
help="Launch a local OpenAI-compatible API server to serve LLM "
|
||||||
|
"completions via HTTP.",
|
||||||
|
description=DESCRIPTION,
|
||||||
|
usage="vllm serve [model_tag] [options]",
|
||||||
)
|
)
|
||||||
|
|
||||||
serve_parser = make_arg_parser(serve_parser)
|
serve_parser = make_arg_parser(serve_parser)
|
||||||
|
|||||||
Reference in New Issue
Block a user