Provide default max model length (#1224)
This commit is contained in:
@@ -184,7 +184,7 @@ class EngineArgs:
|
||||
self.worker_use_ray)
|
||||
scheduler_config = SchedulerConfig(self.max_num_batched_tokens,
|
||||
self.max_num_seqs,
|
||||
model_config.get_max_model_len())
|
||||
model_config.max_model_len)
|
||||
return model_config, cache_config, parallel_config, scheduler_config
|
||||
|
||||
|
||||
|
||||
Reference in New Issue
Block a user