[cli] complete vllm cli help message (#31226)
Signed-off-by: Andy Xie <andy.xning@gmail.com>
This commit is contained in:
@@ -32,6 +32,7 @@ class BenchmarkSubcommand(CLISubcommand):
|
||||
) -> FlexibleArgumentParser:
|
||||
bench_parser = subparsers.add_parser(
|
||||
self.name,
|
||||
help=self.help,
|
||||
description=self.help,
|
||||
usage=f"vllm {self.name} <bench_type> [options]",
|
||||
)
|
||||
|
||||
@@ -66,7 +66,11 @@ class ServeSubcommand(CLISubcommand):
|
||||
self, subparsers: argparse._SubParsersAction
|
||||
) -> FlexibleArgumentParser:
|
||||
serve_parser = subparsers.add_parser(
|
||||
self.name, description=DESCRIPTION, usage="vllm serve [model_tag] [options]"
|
||||
self.name,
|
||||
help="Launch a local OpenAI-compatible API server to serve LLM "
|
||||
"completions via HTTP.",
|
||||
description=DESCRIPTION,
|
||||
usage="vllm serve [model_tag] [options]",
|
||||
)
|
||||
|
||||
serve_parser = make_arg_parser(serve_parser)
|
||||
|
||||
Reference in New Issue
Block a user