vllm.entrypoints.cli ¶
Modules:
| Name | Description |
|---|---|
benchmark | |
collect_env | |
main | The CLI entrypoints of vLLM |
openai | |
run_batch | |
serve | |
types | |
BenchmarkLatencySubcommand ¶
Bases: BenchmarkSubcommandBase
The latency subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/latency.py
BenchmarkMMProcessorSubcommand ¶
Bases: BenchmarkSubcommandBase
The mm-processor subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/mm_processor.py
BenchmarkServingSubcommand ¶
Bases: BenchmarkSubcommandBase
The serve subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/serve.py
BenchmarkStartupSubcommand ¶
Bases: BenchmarkSubcommandBase
The startup subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/startup.py
BenchmarkSweepSubcommand ¶
Bases: BenchmarkSubcommandBase
The sweep subcommand for vllm bench.
Source code in vllm/entrypoints/cli/benchmark/sweep.py
BenchmarkThroughputSubcommand ¶
Bases: BenchmarkSubcommandBase
The throughput subcommand for vllm bench.