[Frontend] Remove custom_cache_manager (#13791)
Signed-off-by: fulvius31 <asangior@redhat.com>
This commit is contained in:
committed by
GitHub
parent
a4d83661d7
commit
374ee287d8
@@ -16,12 +16,8 @@ import torch
|
||||
|
||||
from vllm.config import VllmConfig
|
||||
from vllm.logger import init_logger
|
||||
from vllm.triton_utils.importing import HAS_TRITON
|
||||
from vllm.utils import _check_multiproc_method, get_mp_context, run_method
|
||||
|
||||
if HAS_TRITON:
|
||||
from vllm.triton_utils import maybe_set_triton_cache_manager
|
||||
|
||||
logger = init_logger(__name__)
|
||||
|
||||
T = TypeVar('T')
|
||||
@@ -314,7 +310,3 @@ def set_multiprocessing_worker_envs(parallel_config):
|
||||
current_parallelism, default_omp_num_threads)
|
||||
os.environ["OMP_NUM_THREADS"] = str(default_omp_num_threads)
|
||||
torch.set_num_threads(default_omp_num_threads)
|
||||
|
||||
# workaround for https://github.com/vllm-project/vllm/issues/6103
|
||||
if HAS_TRITON and parallel_config.world_size > 1:
|
||||
maybe_set_triton_cache_manager()
|
||||
|
||||
Reference in New Issue
Block a user