decoding.py -> sampling_params.py
This commit is contained in:
33
cacheflow/sampling_params.py
Normal file
33
cacheflow/sampling_params.py
Normal file
@@ -0,0 +1,33 @@
|
||||
from typing import Optional, Set
|
||||
|
||||
|
||||
class SamplingParams:
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
n: int = 1,
|
||||
temperature: float = 1.0,
|
||||
top_p: float = 1.0,
|
||||
use_beam_search: bool = False,
|
||||
stop_token_ids: Set[int] = [],
|
||||
max_context_len: Optional[int] = None,
|
||||
) -> None:
|
||||
assert n >= 1
|
||||
assert temperature >= 0.0
|
||||
assert 0.0 < top_p <= 1.0
|
||||
if use_beam_search:
|
||||
assert n > 1
|
||||
assert temperature > 0.0
|
||||
assert top_p == 1.0
|
||||
elif temperature == 0.0:
|
||||
# Zero temperature means greedy decoding.
|
||||
assert n == 1
|
||||
assert top_p == 1.0
|
||||
assert max_context_len is None or max_context_len >= 0
|
||||
|
||||
self.n = n
|
||||
self.temperature = temperature
|
||||
self.top_p = top_p
|
||||
self.use_beam_search = use_beam_search
|
||||
self.stop_token_ids = stop_token_ids
|
||||
self.max_context_len = max_context_len
|
||||
Reference in New Issue
Block a user