Files
vllm/examples/offline_inference/run_one_batch.py