[CI] Add PPL test for generation models (#24485)
Signed-off-by: wang.yuqi <noooop@126.com>
This commit is contained in:
@@ -59,7 +59,7 @@ def correctness_test_embed_models(hf_runner,
|
||||
|
||||
with hf_runner(
|
||||
model_info.name,
|
||||
dtype="float32",
|
||||
dtype=model_info.hf_dtype,
|
||||
is_sentence_transformer=True,
|
||||
) as hf_model:
|
||||
|
||||
|
||||
@@ -213,7 +213,7 @@ def mteb_test_embed_models(hf_runner,
|
||||
if model_info.mteb_score is None:
|
||||
with hf_runner(model_info.name,
|
||||
is_sentence_transformer=True,
|
||||
dtype="float32") as hf_model:
|
||||
dtype=model_info.hf_dtype) as hf_model:
|
||||
|
||||
# e.g. setting default parameters for the encode method of hf_runner
|
||||
if hf_model_callback is not None:
|
||||
@@ -278,9 +278,12 @@ def run_mteb_rerank(cross_encoder, tasks, languages):
|
||||
return main_score
|
||||
|
||||
|
||||
def mteb_test_rerank_models_hf(hf_runner, model_name, hf_model_callback=None):
|
||||
def mteb_test_rerank_models_hf(hf_runner,
|
||||
model_name,
|
||||
hf_dtype="float32",
|
||||
hf_model_callback=None):
|
||||
with hf_runner(model_name, is_cross_encoder=True,
|
||||
dtype="float32") as hf_model:
|
||||
dtype=hf_dtype) as hf_model:
|
||||
|
||||
original_predict = hf_model.predict
|
||||
|
||||
@@ -357,7 +360,7 @@ def mteb_test_rerank_models(hf_runner,
|
||||
# SentenceTransformers mteb score to a constant
|
||||
if model_info.mteb_score is None:
|
||||
st_main_score, st_dtype = mteb_test_rerank_models_hf(
|
||||
hf_runner, model_info.name, hf_model_callback)
|
||||
hf_runner, model_info.name, model_info.hf_dtype, hf_model_callback)
|
||||
else:
|
||||
st_main_score = model_info.mteb_score
|
||||
st_dtype = "Constant"
|
||||
|
||||
Reference in New Issue
Block a user