vllm.entrypoints.cli ¶
 Modules:
| Name | Description | 
|---|---|
benchmark |    |  
collect_env |    |  
main |    The CLI entrypoints of vLLM  |  
openai |    |  
run_batch |    |  
serve |    |  
types |    |  
  __all__  module-attribute  ¶
 __all__: list[str] = [
    "BenchmarkLatencySubcommand",
    "BenchmarkServingSubcommand",
    "BenchmarkSweepSubcommand",
    "BenchmarkThroughputSubcommand",
]
  BenchmarkLatencySubcommand ¶
  Bases: BenchmarkSubcommandBase
The latency subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/latency.py
   help  class-attribute instance-attribute  ¶
    add_cli_args  classmethod  ¶
 add_cli_args(parser: ArgumentParser) -> None
  BenchmarkServingSubcommand ¶
  Bases: BenchmarkSubcommandBase
The serve subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/serve.py
   add_cli_args  classmethod  ¶
 add_cli_args(parser: ArgumentParser) -> None
  BenchmarkSweepSubcommand ¶
  Bases: BenchmarkSubcommandBase
The sweep subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/sweep.py
   add_cli_args  classmethod  ¶
 add_cli_args(parser: ArgumentParser) -> None
  BenchmarkThroughputSubcommand ¶
  Bases: BenchmarkSubcommandBase
The throughput subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/throughput.py
   add_cli_args  classmethod  ¶
 add_cli_args(parser: ArgumentParser) -> None