Fix typo in tokenizer.py (#750)

conjuction -> conjunction
This commit is contained in:
Ikko Eltociear Ashimine
2023-08-15 14:26:36 +09:00
committed by GitHub
parent 1b151ed181
commit 805de738f6

View File

@@ -72,7 +72,7 @@ def detokenize_incrementally(
new_token_id: int,
skip_special_tokens: bool,
) -> Tuple[str, str]:
"""Detokenizes the new token in conjuction with the previous output tokens.
"""Detokenizes the new token in conjunction with the previous output tokens.
NOTE: This function does not update prev_output_tokens.