|
@@ -130,17 +130,6 @@ def upload_soft_prompt(file):
|
|
|
|
|
|
|
|
return name
|
|
return name
|
|
|
|
|
|
|
|
-def create_model_and_preset_menus():
|
|
|
|
|
- with gr.Row():
|
|
|
|
|
- with gr.Column():
|
|
|
|
|
- with gr.Row():
|
|
|
|
|
- shared.gradio['model_menu'] = gr.Dropdown(choices=available_models, value=shared.model_name, label='Model')
|
|
|
|
|
- ui.create_refresh_button(shared.gradio['model_menu'], lambda : None, lambda : {'choices': get_available_models()}, 'refresh-button')
|
|
|
|
|
- with gr.Column():
|
|
|
|
|
- with gr.Row():
|
|
|
|
|
- shared.gradio['preset_menu'] = gr.Dropdown(choices=available_presets, value=default_preset if not shared.args.flexgen else 'Naive', label='Generation parameters preset')
|
|
|
|
|
- ui.create_refresh_button(shared.gradio['preset_menu'], lambda : None, lambda : {'choices': get_available_presets()}, 'refresh-button')
|
|
|
|
|
-
|
|
|
|
|
def save_prompt(text):
|
|
def save_prompt(text):
|
|
|
fname = f"{datetime.now().strftime('%Y-%m-%d-%H%M%S')}.txt"
|
|
fname = f"{datetime.now().strftime('%Y-%m-%d-%H%M%S')}.txt"
|
|
|
with open(Path(f'prompts/{fname}'), 'w', encoding='utf-8') as f:
|
|
with open(Path(f'prompts/{fname}'), 'w', encoding='utf-8') as f:
|
|
@@ -172,6 +161,20 @@ def create_prompt_menus():
|
|
|
shared.gradio['prompt_menu'].change(load_prompt, [shared.gradio['prompt_menu']], [shared.gradio['textbox']], show_progress=False)
|
|
shared.gradio['prompt_menu'].change(load_prompt, [shared.gradio['prompt_menu']], [shared.gradio['textbox']], show_progress=False)
|
|
|
shared.gradio['save_prompt'].click(save_prompt, [shared.gradio['textbox']], [shared.gradio['status']], show_progress=False)
|
|
shared.gradio['save_prompt'].click(save_prompt, [shared.gradio['textbox']], [shared.gradio['status']], show_progress=False)
|
|
|
|
|
|
|
|
|
|
+def create_model_menus():
|
|
|
|
|
+ with gr.Row():
|
|
|
|
|
+ with gr.Column():
|
|
|
|
|
+ with gr.Row():
|
|
|
|
|
+ shared.gradio['model_menu'] = gr.Dropdown(choices=available_models, value=shared.model_name, label='Model')
|
|
|
|
|
+ ui.create_refresh_button(shared.gradio['model_menu'], lambda : None, lambda : {'choices': get_available_models()}, 'refresh-button')
|
|
|
|
|
+ with gr.Column():
|
|
|
|
|
+ with gr.Row():
|
|
|
|
|
+ shared.gradio['lora_menu'] = gr.Dropdown(choices=available_loras, value=shared.lora_name, label='LoRA')
|
|
|
|
|
+ ui.create_refresh_button(shared.gradio['lora_menu'], lambda : None, lambda : {'choices': get_available_loras()}, 'refresh-button')
|
|
|
|
|
+
|
|
|
|
|
+ shared.gradio['model_menu'].change(load_model_wrapper, shared.gradio['model_menu'], shared.gradio['model_menu'], show_progress=True)
|
|
|
|
|
+ shared.gradio['lora_menu'].change(load_lora_wrapper, shared.gradio['lora_menu'], shared.gradio['lora_menu'], show_progress=True)
|
|
|
|
|
+
|
|
|
def create_settings_menus(default_preset):
|
|
def create_settings_menus(default_preset):
|
|
|
generate_params = load_preset_values(default_preset if not shared.args.flexgen else 'Naive', {}, return_dict=True)
|
|
generate_params = load_preset_values(default_preset if not shared.args.flexgen else 'Naive', {}, return_dict=True)
|
|
|
for k in ['max_new_tokens', 'seed', 'stop_at_newline', 'chat_prompt_size', 'chat_generation_attempts']:
|
|
for k in ['max_new_tokens', 'seed', 'stop_at_newline', 'chat_prompt_size', 'chat_generation_attempts']:
|
|
@@ -180,7 +183,9 @@ def create_settings_menus(default_preset):
|
|
|
|
|
|
|
|
with gr.Row():
|
|
with gr.Row():
|
|
|
with gr.Column():
|
|
with gr.Column():
|
|
|
- create_model_and_preset_menus()
|
|
|
|
|
|
|
+ with gr.Row():
|
|
|
|
|
+ shared.gradio['preset_menu'] = gr.Dropdown(choices=available_presets, value=default_preset if not shared.args.flexgen else 'Naive', label='Generation parameters preset')
|
|
|
|
|
+ ui.create_refresh_button(shared.gradio['preset_menu'], lambda : None, lambda : {'choices': get_available_presets()}, 'refresh-button')
|
|
|
with gr.Column():
|
|
with gr.Column():
|
|
|
shared.gradio['seed'] = gr.Number(value=shared.settings['seed'], label='Seed (-1 for random)')
|
|
shared.gradio['seed'] = gr.Number(value=shared.settings['seed'], label='Seed (-1 for random)')
|
|
|
|
|
|
|
@@ -214,9 +219,6 @@ def create_settings_menus(default_preset):
|
|
|
shared.gradio['length_penalty'] = gr.Slider(-5, 5, value=generate_params['length_penalty'], label='length_penalty')
|
|
shared.gradio['length_penalty'] = gr.Slider(-5, 5, value=generate_params['length_penalty'], label='length_penalty')
|
|
|
shared.gradio['early_stopping'] = gr.Checkbox(value=generate_params['early_stopping'], label='early_stopping')
|
|
shared.gradio['early_stopping'] = gr.Checkbox(value=generate_params['early_stopping'], label='early_stopping')
|
|
|
|
|
|
|
|
- with gr.Row():
|
|
|
|
|
- shared.gradio['lora_menu'] = gr.Dropdown(choices=available_loras, value=shared.lora_name, label='LoRA')
|
|
|
|
|
- ui.create_refresh_button(shared.gradio['lora_menu'], lambda : None, lambda : {'choices': get_available_loras()}, 'refresh-button')
|
|
|
|
|
|
|
|
|
|
with gr.Accordion('Soft prompt', open=False):
|
|
with gr.Accordion('Soft prompt', open=False):
|
|
|
with gr.Row():
|
|
with gr.Row():
|
|
@@ -227,9 +229,7 @@ def create_settings_menus(default_preset):
|
|
|
with gr.Row():
|
|
with gr.Row():
|
|
|
shared.gradio['upload_softprompt'] = gr.File(type='binary', file_types=['.zip'])
|
|
shared.gradio['upload_softprompt'] = gr.File(type='binary', file_types=['.zip'])
|
|
|
|
|
|
|
|
- shared.gradio['model_menu'].change(load_model_wrapper, shared.gradio['model_menu'], shared.gradio['model_menu'], show_progress=True)
|
|
|
|
|
shared.gradio['preset_menu'].change(load_preset_values, [shared.gradio[k] for k in ['preset_menu', 'generate_state']], [shared.gradio[k] for k in ['generate_state', 'do_sample', 'temperature', 'top_p', 'typical_p', 'repetition_penalty', 'encoder_repetition_penalty', 'top_k', 'min_length', 'no_repeat_ngram_size', 'num_beams', 'penalty_alpha', 'length_penalty', 'early_stopping']])
|
|
shared.gradio['preset_menu'].change(load_preset_values, [shared.gradio[k] for k in ['preset_menu', 'generate_state']], [shared.gradio[k] for k in ['generate_state', 'do_sample', 'temperature', 'top_p', 'typical_p', 'repetition_penalty', 'encoder_repetition_penalty', 'top_k', 'min_length', 'no_repeat_ngram_size', 'num_beams', 'penalty_alpha', 'length_penalty', 'early_stopping']])
|
|
|
- shared.gradio['lora_menu'].change(load_lora_wrapper, shared.gradio['lora_menu'], shared.gradio['lora_menu'], show_progress=True)
|
|
|
|
|
shared.gradio['softprompts_menu'].change(load_soft_prompt, shared.gradio['softprompts_menu'], shared.gradio['softprompts_menu'], show_progress=True)
|
|
shared.gradio['softprompts_menu'].change(load_soft_prompt, shared.gradio['softprompts_menu'], shared.gradio['softprompts_menu'], show_progress=True)
|
|
|
shared.gradio['upload_softprompt'].upload(upload_soft_prompt, shared.gradio['upload_softprompt'], shared.gradio['softprompts_menu'])
|
|
shared.gradio['upload_softprompt'].upload(upload_soft_prompt, shared.gradio['upload_softprompt'], shared.gradio['softprompts_menu'])
|
|
|
|
|
|
|
@@ -502,6 +502,9 @@ def create_interface():
|
|
|
shared.gradio['Stop'].click(stop_everything_event, [], [], queue=False, cancels=gen_events if shared.args.no_stream else None)
|
|
shared.gradio['Stop'].click(stop_everything_event, [], [], queue=False, cancels=gen_events if shared.args.no_stream else None)
|
|
|
shared.gradio['interface'].load(None, None, None, _js=f"() => {{{ui.main_js}}}")
|
|
shared.gradio['interface'].load(None, None, None, _js=f"() => {{{ui.main_js}}}")
|
|
|
|
|
|
|
|
|
|
+ with gr.Tab("Model", elem_id="model-tab"):
|
|
|
|
|
+ create_model_menus()
|
|
|
|
|
+
|
|
|
with gr.Tab("Training", elem_id="training-tab"):
|
|
with gr.Tab("Training", elem_id="training-tab"):
|
|
|
training.create_train_interface()
|
|
training.create_train_interface()
|
|
|
|
|
|