Implement presence and frequency penalties (#95)

This commit is contained in:
Woosuk Kwon
2023-05-10 23:39:12 -07:00
committed by GitHub
parent 9f88db35da
commit 55f8b0a5de
9 changed files with 215 additions and 82 deletions

View File

@@ -6,6 +6,8 @@ class SamplingParams:
def __init__(
self,
n: int,
presence_penalty: float,
frequency_penalty: float,
temperature: float,
top_p: float,
top_k: int,
@@ -16,6 +18,12 @@ class SamplingParams:
) -> None:
if n < 1:
raise ValueError(f"n must be at least 1, got {n}.")
if not -2.0 <= presence_penalty <= 2.0:
raise ValueError(
f"presence_penalty must be in [-2, 2], got {presence_penalty}.")
if not -2.0 <= frequency_penalty <= 2.0:
raise ValueError(
f"frequency_penalty must be in [-2, 2], got {frequency_penalty}.")
if temperature < 0.0:
raise ValueError(
f"temperature must be non-negative, got {temperature}.")
@@ -57,6 +65,8 @@ class SamplingParams:
"top_k must be -1 when using greedy sampling.")
self.n = n
self.presence_penalty = presence_penalty
self.frequency_penalty = frequency_penalty
self.temperature = temperature
self.top_p = top_p
self.top_k = top_k
@@ -67,6 +77,8 @@ class SamplingParams:
def __repr__(self) -> str:
return (f"SamplingParams(n={self.n}, "
f"presence_penalty={self.presence_penalty}, "
f"frequency_penalty={self.frequency_penalty}, "
f"temperature={self.temperature}, "
f"top_p={self.top_p}, "
f"top_k={self.top_k},"
@@ -77,13 +89,18 @@ class SamplingParams:
@classmethod
def from_dict(cls, d: Dict) -> "SamplingParams":
return cls(
n=d.get("n", 1),
temperature=d.get("temperature", 1.0),
top_p=d.get("top_p", 1.0),
top_k=d.get("top_k", -1),
use_beam_search=d.get("use_beam_search", False),
stop_token_ids=set(d.get("stop_token_ids", set())),
max_num_steps=d.get("max_num_steps", 16),
num_logprobs=d.get("num_logprobs", 0),
sampling_params = cls(
n=d.pop("n", 1),
presence_penalty=d.pop("presence_penalty", 0.0),
frequency_penalty=d.pop("frequency_penalty", 0.0),
temperature=d.pop("temperature", 1.0),
top_p=d.pop("top_p", 1.0),
top_k=d.pop("top_k", -1),
use_beam_search=d.pop("use_beam_search", False),
stop_token_ids=set(d.pop("stop_token_ids", set())),
max_num_steps=d.pop("max_num_steps", 16),
num_logprobs=d.pop("num_logprobs", 0),
)
if d:
raise ValueError(f"Unrecognized keys in dict: {d.keys()}")
return sampling_params