[Frontend][Doc][5/N] Improve all pooling task | Polish encode (pooling) api & Document. (#25524)

Signed-off-by: wang.yuqi <noooop@126.com>
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
Co-authored-by: Cyrus Leung <cyrus.tl.leung@gmail.com>
This commit is contained in:
wang.yuqi
2025-10-30 20:13:05 +08:00
committed by GitHub
parent 74374386e2
commit 4464723f22
27 changed files with 499 additions and 131 deletions

View File

@@ -24,7 +24,7 @@ def test_classify_models_using_activation(
model,
max_model_len=512,
dtype=dtype,
pooler_config=PoolerConfig(activation=False),
pooler_config=PoolerConfig(use_activation=False),
) as vllm_model:
wo_activation_out = vllm_model.classify(example_prompts)
@@ -32,7 +32,7 @@ def test_classify_models_using_activation(
model,
max_model_len=512,
dtype=dtype,
pooler_config=PoolerConfig(activation=True),
pooler_config=PoolerConfig(use_activation=True),
) as vllm_model:
w_activation_out = vllm_model.classify(example_prompts)
@@ -104,7 +104,7 @@ def test_reward_models_using_activation(
model,
max_model_len=1024,
dtype=dtype,
pooler_config=PoolerConfig(activation=False),
pooler_config=PoolerConfig(use_activation=False),
) as vllm_model:
wo_activation = vllm_model.reward(example_prompts)
@@ -112,7 +112,7 @@ def test_reward_models_using_activation(
model,
max_model_len=1024,
dtype=dtype,
pooler_config=PoolerConfig(activation=True),
pooler_config=PoolerConfig(use_activation=True),
) as vllm_model:
w_activation = vllm_model.reward(example_prompts)