|
@@ -92,8 +92,7 @@ def do_train(lora_name: str, micro_batch_size: int, batch_size: int, epochs: int
|
|
|
|
|
|
|
|
# == Input validation / processing ==
|
|
# == Input validation / processing ==
|
|
|
yield "Prepping..."
|
|
yield "Prepping..."
|
|
|
- # TODO: --lora-dir PR once pulled will need to be applied here
|
|
|
|
|
- lora_name = f"loras/{clean_path(None, lora_name)}"
|
|
|
|
|
|
|
+ lora_name = f"{shared.args.lora_dir}/{clean_path(None, lora_name)}"
|
|
|
if dataset is None:
|
|
if dataset is None:
|
|
|
return "**Missing dataset choice input, cannot continue.**"
|
|
return "**Missing dataset choice input, cannot continue.**"
|
|
|
if format is None:
|
|
if format is None:
|