[Bugfix] Fix usage of deprecated decorator (#11025)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
@@ -252,8 +252,21 @@ class LLM:
|
||||
else:
|
||||
tokenizer_group.tokenizer = get_cached_tokenizer(tokenizer)
|
||||
|
||||
@overload
|
||||
def generate(
|
||||
self,
|
||||
prompts: Union[PromptType, Sequence[PromptType]],
|
||||
/,
|
||||
*,
|
||||
sampling_params: Optional[Union[SamplingParams,
|
||||
Sequence[SamplingParams]]] = None,
|
||||
use_tqdm: bool = True,
|
||||
lora_request: Optional[Union[List[LoRARequest], LoRARequest]] = None,
|
||||
) -> List[RequestOutput]:
|
||||
...
|
||||
|
||||
@overload # LEGACY: single (prompt + optional token ids)
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def generate(
|
||||
self,
|
||||
prompts: str,
|
||||
@@ -266,7 +279,7 @@ class LLM:
|
||||
...
|
||||
|
||||
@overload # LEGACY: multi (prompt + optional token ids)
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def generate(
|
||||
self,
|
||||
prompts: List[str],
|
||||
@@ -279,7 +292,7 @@ class LLM:
|
||||
...
|
||||
|
||||
@overload # LEGACY: single (token ids + optional prompt)
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def generate(
|
||||
self,
|
||||
prompts: Optional[str] = None,
|
||||
@@ -293,7 +306,7 @@ class LLM:
|
||||
...
|
||||
|
||||
@overload # LEGACY: multi (token ids + optional prompt)
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def generate(
|
||||
self,
|
||||
prompts: Optional[List[str]] = None,
|
||||
@@ -307,7 +320,7 @@ class LLM:
|
||||
...
|
||||
|
||||
@overload # LEGACY: single or multi token ids [pos-only]
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def generate(
|
||||
self,
|
||||
prompts: None,
|
||||
@@ -318,19 +331,6 @@ class LLM:
|
||||
) -> List[RequestOutput]:
|
||||
...
|
||||
|
||||
@overload
|
||||
def generate(
|
||||
self,
|
||||
prompts: Union[PromptType, Sequence[PromptType]],
|
||||
/,
|
||||
*,
|
||||
sampling_params: Optional[Union[SamplingParams,
|
||||
Sequence[SamplingParams]]] = None,
|
||||
use_tqdm: bool = True,
|
||||
lora_request: Optional[Union[List[LoRARequest], LoRARequest]] = None,
|
||||
) -> List[RequestOutput]:
|
||||
...
|
||||
|
||||
@deprecate_kwargs(
|
||||
"prompt_token_ids",
|
||||
is_deprecated=lambda: LLM.DEPRECATE_LEGACY,
|
||||
@@ -672,8 +672,21 @@ class LLM:
|
||||
lora_request=lora_request,
|
||||
)
|
||||
|
||||
@overload
|
||||
def encode(
|
||||
self,
|
||||
prompts: Union[PromptType, Sequence[PromptType]],
|
||||
/,
|
||||
*,
|
||||
pooling_params: Optional[Union[PoolingParams,
|
||||
Sequence[PoolingParams]]] = None,
|
||||
use_tqdm: bool = True,
|
||||
lora_request: Optional[Union[List[LoRARequest], LoRARequest]] = None,
|
||||
) -> List[PoolingRequestOutput]:
|
||||
...
|
||||
|
||||
@overload # LEGACY: single (prompt + optional token ids)
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def encode(
|
||||
self,
|
||||
prompts: str,
|
||||
@@ -686,7 +699,7 @@ class LLM:
|
||||
...
|
||||
|
||||
@overload # LEGACY: multi (prompt + optional token ids)
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def encode(
|
||||
self,
|
||||
prompts: List[str],
|
||||
@@ -699,7 +712,7 @@ class LLM:
|
||||
...
|
||||
|
||||
@overload # LEGACY: single (token ids + optional prompt)
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def encode(
|
||||
self,
|
||||
prompts: Optional[str] = None,
|
||||
@@ -713,7 +726,7 @@ class LLM:
|
||||
...
|
||||
|
||||
@overload # LEGACY: multi (token ids + optional prompt)
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def encode(
|
||||
self,
|
||||
prompts: Optional[List[str]] = None,
|
||||
@@ -727,7 +740,7 @@ class LLM:
|
||||
...
|
||||
|
||||
@overload # LEGACY: single or multi token ids [pos-only]
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts")
|
||||
@deprecated("'prompt_token_ids' will become part of 'prompts'")
|
||||
def encode(
|
||||
self,
|
||||
prompts: None,
|
||||
@@ -738,19 +751,6 @@ class LLM:
|
||||
) -> List[PoolingRequestOutput]:
|
||||
...
|
||||
|
||||
@overload
|
||||
def encode(
|
||||
self,
|
||||
prompts: Union[PromptType, Sequence[PromptType]],
|
||||
/,
|
||||
*,
|
||||
pooling_params: Optional[Union[PoolingParams,
|
||||
Sequence[PoolingParams]]] = None,
|
||||
use_tqdm: bool = True,
|
||||
lora_request: Optional[Union[List[LoRARequest], LoRARequest]] = None,
|
||||
) -> List[PoolingRequestOutput]:
|
||||
...
|
||||
|
||||
@deprecate_kwargs(
|
||||
"prompt_token_ids",
|
||||
is_deprecated=lambda: LLM.DEPRECATE_LEGACY,
|
||||
|
||||
Reference in New Issue
Block a user