[Misc] Remove unnecessary detokenization in multimodal processing (#12868)

This commit is contained in:
Cyrus Leung
2025-02-07 22:21:17 +08:00
committed by GitHub
parent 1918aa1b80
commit ce26b16268
4 changed files with 7 additions and 10 deletions

View File

@@ -93,5 +93,5 @@ async def test_image_embedding(server: RemoteOpenAIServer, model_name: str,
assert len(embeddings.data) == 1
assert len(embeddings.data[0].embedding) == 3072
assert embeddings.usage.completion_tokens == 0
assert embeddings.usage.prompt_tokens == 764
assert embeddings.usage.total_tokens == 764
assert embeddings.usage.prompt_tokens == 763
assert embeddings.usage.total_tokens == 763