[BugFix] Fix get tokenizer when using ray (#3301)

This commit is contained in:
Roy
2024-03-11 10:17:16 +08:00
committed by GitHub
parent e4a28e5316
commit 9e8744a545
6 changed files with 23 additions and 7 deletions

View File

@@ -120,7 +120,8 @@ class TokenizerGroup:
def get_lora_tokenizer(
self,
lora_request: Optional[LoRARequest]) -> "PreTrainedTokenizer":
lora_request: Optional[LoRARequest] = None
) -> "PreTrainedTokenizer":
if not lora_request or not self.enable_lora:
return self.tokenizer
if lora_request.lora_int_id not in self.lora_tokenizers:
@@ -133,7 +134,8 @@ class TokenizerGroup:
async def get_lora_tokenizer_async(
self,
lora_request: Optional[LoRARequest]) -> "PreTrainedTokenizer":
lora_request: Optional[LoRARequest] = None
) -> "PreTrainedTokenizer":
if not lora_request or not self.enable_lora:
return self.tokenizer
if lora_request.lora_int_id not in self.lora_tokenizers: