[TPU] Support multi-host inference (#7457)
This commit is contained in:
@@ -8,7 +8,7 @@ vLLM supports Google Cloud TPUs using PyTorch XLA.
|
||||
Requirements
|
||||
------------
|
||||
|
||||
* Google Cloud TPU VM (single host)
|
||||
* Google Cloud TPU VM (single & multi host)
|
||||
* TPU versions: v5e, v5p, v4
|
||||
* Python: 3.10
|
||||
|
||||
|
||||
Reference in New Issue
Block a user