|
@@ -510,6 +510,10 @@ def train_process(
|
|
|
)
|
|
)
|
|
|
)
|
|
)
|
|
|
logger.info(project)
|
|
logger.info(project)
|
|
|
|
|
+
|
|
|
|
|
+ if llama_check_interval > llama_maxsteps:
|
|
|
|
|
+ llama_check_interval = llama_maxsteps
|
|
|
|
|
+
|
|
|
train_cmd = [
|
|
train_cmd = [
|
|
|
PYTHON,
|
|
PYTHON,
|
|
|
"fish_speech/train.py",
|
|
"fish_speech/train.py",
|
|
@@ -800,7 +804,7 @@ with gr.Blocks(
|
|
|
"Use LoRA can save GPU memory, but may reduce the quality of the model"
|
|
"Use LoRA can save GPU memory, but may reduce the quality of the model"
|
|
|
),
|
|
),
|
|
|
value=True,
|
|
value=True,
|
|
|
- interactive=False,
|
|
|
|
|
|
|
+ interactive=True,
|
|
|
)
|
|
)
|
|
|
llama_ckpt = gr.Dropdown(
|
|
llama_ckpt = gr.Dropdown(
|
|
|
label=i18n("Select LLAMA ckpt"),
|
|
label=i18n("Select LLAMA ckpt"),
|
|
@@ -816,19 +820,25 @@ with gr.Blocks(
|
|
|
with gr.Row(equal_height=False):
|
|
with gr.Row(equal_height=False):
|
|
|
llama_lr_slider = gr.Slider(
|
|
llama_lr_slider = gr.Slider(
|
|
|
label=i18n("Initial Learning Rate"),
|
|
label=i18n("Initial Learning Rate"),
|
|
|
|
|
+ info=i18n(
|
|
|
|
|
+ "lr smaller -> usually train slower but more stable"
|
|
|
|
|
+ ),
|
|
|
interactive=True,
|
|
interactive=True,
|
|
|
minimum=1e-5,
|
|
minimum=1e-5,
|
|
|
maximum=1e-4,
|
|
maximum=1e-4,
|
|
|
step=1e-5,
|
|
step=1e-5,
|
|
|
- value=init_llama_yml["model"]["optimizer"]["lr"],
|
|
|
|
|
|
|
+ value=5e-5,
|
|
|
)
|
|
)
|
|
|
llama_maxsteps_slider = gr.Slider(
|
|
llama_maxsteps_slider = gr.Slider(
|
|
|
label=i18n("Maximum Training Steps"),
|
|
label=i18n("Maximum Training Steps"),
|
|
|
|
|
+ info=i18n(
|
|
|
|
|
+ "recommend: max_steps = num_audios // batch_size * (2 to 5)"
|
|
|
|
|
+ ),
|
|
|
interactive=True,
|
|
interactive=True,
|
|
|
- minimum=50,
|
|
|
|
|
|
|
+ minimum=1,
|
|
|
maximum=10000,
|
|
maximum=10000,
|
|
|
- step=50,
|
|
|
|
|
- value=init_llama_yml["trainer"]["max_steps"],
|
|
|
|
|
|
|
+ step=1,
|
|
|
|
|
+ value=50,
|
|
|
)
|
|
)
|
|
|
with gr.Row(equal_height=False):
|
|
with gr.Row(equal_height=False):
|
|
|
llama_base_config = gr.Dropdown(
|
|
llama_base_config = gr.Dropdown(
|
|
@@ -841,13 +851,9 @@ with gr.Blocks(
|
|
|
llama_data_num_workers_slider = gr.Slider(
|
|
llama_data_num_workers_slider = gr.Slider(
|
|
|
label=i18n("Number of Workers"),
|
|
label=i18n("Number of Workers"),
|
|
|
minimum=1,
|
|
minimum=1,
|
|
|
- maximum=16,
|
|
|
|
|
|
|
+ maximum=32,
|
|
|
step=1,
|
|
step=1,
|
|
|
- value=(
|
|
|
|
|
- init_llama_yml["data"]["num_workers"]
|
|
|
|
|
- if sys.platform == "linux"
|
|
|
|
|
- else 1
|
|
|
|
|
- ),
|
|
|
|
|
|
|
+ value=4,
|
|
|
)
|
|
)
|
|
|
with gr.Row(equal_height=False):
|
|
with gr.Row(equal_height=False):
|
|
|
llama_data_batch_size_slider = gr.Slider(
|
|
llama_data_batch_size_slider = gr.Slider(
|
|
@@ -856,7 +862,7 @@ with gr.Blocks(
|
|
|
minimum=1,
|
|
minimum=1,
|
|
|
maximum=32,
|
|
maximum=32,
|
|
|
step=1,
|
|
step=1,
|
|
|
- value=init_llama_yml["data"]["batch_size"],
|
|
|
|
|
|
|
+ value=4,
|
|
|
)
|
|
)
|
|
|
llama_data_max_length_slider = gr.Slider(
|
|
llama_data_max_length_slider = gr.Slider(
|
|
|
label=i18n("Maximum Length per Sample"),
|
|
label=i18n("Maximum Length per Sample"),
|
|
@@ -864,7 +870,7 @@ with gr.Blocks(
|
|
|
minimum=1024,
|
|
minimum=1024,
|
|
|
maximum=4096,
|
|
maximum=4096,
|
|
|
step=128,
|
|
step=128,
|
|
|
- value=init_llama_yml["max_length"],
|
|
|
|
|
|
|
+ value=1024,
|
|
|
)
|
|
)
|
|
|
with gr.Row(equal_height=False):
|
|
with gr.Row(equal_height=False):
|
|
|
llama_precision_dropdown = gr.Dropdown(
|
|
llama_precision_dropdown = gr.Dropdown(
|
|
@@ -878,13 +884,14 @@ with gr.Blocks(
|
|
|
)
|
|
)
|
|
|
llama_check_interval_slider = gr.Slider(
|
|
llama_check_interval_slider = gr.Slider(
|
|
|
label=i18n("Save model every n steps"),
|
|
label=i18n("Save model every n steps"),
|
|
|
|
|
+ info=i18n(
|
|
|
|
|
+ "make sure that it's not greater than max_steps"
|
|
|
|
|
+ ),
|
|
|
interactive=True,
|
|
interactive=True,
|
|
|
- minimum=50,
|
|
|
|
|
|
|
+ minimum=1,
|
|
|
maximum=1000,
|
|
maximum=1000,
|
|
|
- step=50,
|
|
|
|
|
- value=init_llama_yml["trainer"][
|
|
|
|
|
- "val_check_interval"
|
|
|
|
|
- ],
|
|
|
|
|
|
|
+ step=1,
|
|
|
|
|
+ value=50,
|
|
|
)
|
|
)
|
|
|
with gr.Row(equal_height=False):
|
|
with gr.Row(equal_height=False):
|
|
|
llama_grad_batches = gr.Slider(
|
|
llama_grad_batches = gr.Slider(
|