fix the issue that len(tokenizer(prompt)["input_ids"]) > prompt_len (#10524)

Signed-off-by: Wang, Yi A <yi.a.wang@intel.com>
This commit is contained in:
Wang, Yi
2024-11-21 19:15:36 +08:00
committed by GitHub
parent 1cfde82ffd
commit 8a93a598d9

View File

@@ -54,6 +54,7 @@ async def async_request_tgi(
"do_sample": True,
"temperature": 0.01, # TGI does not accept 0.0 temperature.
"top_p": 0.99, # TGI does not accept 1.0 top_p.
"truncate": request_func_input.prompt_len,
# TGI does not accept ignore_eos flag.
}
payload = {