[Misc] Enable V1 LoRA by default (#15320)
Signed-off-by: Varun Sundar Rabindranath <varun@neuralmagic.com> Co-authored-by: Varun Sundar Rabindranath <varun@neuralmagic.com>
This commit is contained in:
committed by
GitHub
parent
30d6a015e0
commit
79455cf421
@@ -1512,10 +1512,6 @@ class EngineArgs:
|
||||
and _warn_or_fallback("Engine in background thread")):
|
||||
return False
|
||||
|
||||
# LoRA is supported on V1, but off by default for now.
|
||||
if self.enable_lora and _warn_or_fallback("LORA"):
|
||||
return False
|
||||
|
||||
# PP is supported on V1 with Ray distributed executor,
|
||||
# but off for MP distributed executor for now.
|
||||
if (self.pipeline_parallel_size > 1
|
||||
|
||||
Reference in New Issue
Block a user