Files
vllm/examples/offline_inference/batch_llm_inference.py