[Feature] Add load generation config from model (#11164)
Signed-off-by: liuyanyi <wolfsonliu@163.com> Signed-off-by: Yanyi Liu <wolfsonliu@163.com> Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk> Co-authored-by: Cyrus Leung <tlleungac@connect.ust.hk>
This commit is contained in:
@@ -91,6 +91,10 @@ class OpenAIServingChat(OpenAIServing):
|
||||
"been registered") from e
|
||||
|
||||
self.enable_prompt_tokens_details = enable_prompt_tokens_details
|
||||
diff_sampling_param = self.model_config.get_diff_sampling_param()
|
||||
if diff_sampling_param:
|
||||
logger.info("Overwriting default chat sampling param with: %s",
|
||||
diff_sampling_param)
|
||||
|
||||
async def create_chat_completion(
|
||||
self,
|
||||
@@ -191,13 +195,17 @@ class OpenAIServingChat(OpenAIServing):
|
||||
sampling_params: Union[SamplingParams, BeamSearchParams]
|
||||
default_max_tokens = self.max_model_len - len(
|
||||
engine_prompt["prompt_token_ids"])
|
||||
# Build default sampling params
|
||||
default_sampling_params = (
|
||||
self.model_config.get_diff_sampling_param())
|
||||
if request.use_beam_search:
|
||||
sampling_params = request.to_beam_search_params(
|
||||
default_max_tokens)
|
||||
default_max_tokens, default_sampling_params)
|
||||
else:
|
||||
sampling_params = request.to_sampling_params(
|
||||
default_max_tokens,
|
||||
self.model_config.logits_processor_pattern)
|
||||
self.model_config.logits_processor_pattern,
|
||||
default_sampling_params)
|
||||
|
||||
self._log_inputs(request_id,
|
||||
request_prompts[i],
|
||||
|
||||
Reference in New Issue
Block a user