[Model] Add mistral function calling format to all models loaded with "mistral" format (#8515)
Co-authored-by: Cyrus Leung <cyrus.tl.leung@gmail.com>
This commit is contained in:
committed by
GitHub
parent
9855b99502
commit
a54ed80249
@@ -1,5 +1,6 @@
|
||||
from contextlib import contextmanager
|
||||
from typing import ClassVar, List, Optional, Sequence, Union, cast, overload
|
||||
from typing import (Any, ClassVar, Dict, List, Optional, Sequence, Union, cast,
|
||||
overload)
|
||||
|
||||
from tqdm import tqdm
|
||||
|
||||
@@ -357,6 +358,7 @@ class LLM:
|
||||
lora_request: Optional[LoRARequest] = None,
|
||||
chat_template: Optional[str] = None,
|
||||
add_generation_prompt: bool = True,
|
||||
tools: Optional[List[Dict[str, Any]]] = None,
|
||||
) -> List[RequestOutput]:
|
||||
"""
|
||||
Generate responses for a chat conversation.
|
||||
@@ -401,6 +403,7 @@ class LLM:
|
||||
messages=messages,
|
||||
chat_template=chat_template,
|
||||
add_generation_prompt=add_generation_prompt,
|
||||
tools=tools,
|
||||
)
|
||||
else:
|
||||
prompt = apply_hf_chat_template(
|
||||
@@ -408,6 +411,7 @@ class LLM:
|
||||
conversation=conversation,
|
||||
chat_template=chat_template,
|
||||
add_generation_prompt=add_generation_prompt,
|
||||
tools=tools,
|
||||
)
|
||||
|
||||
inputs: PromptInputs
|
||||
|
||||
Reference in New Issue
Block a user