Change the name to vLLM (#150)
This commit is contained in:
@@ -2,7 +2,7 @@ import random
|
||||
|
||||
import torch
|
||||
|
||||
from cacheflow import cache_ops
|
||||
from vllm import cache_ops
|
||||
|
||||
|
||||
@torch.inference_mode()
|
||||
|
||||
Reference in New Issue
Block a user