Commit
·
d406160
1
Parent(s):
7b75dbf
Update loha_01/config_file.toml
Browse files- loha_01/config_file.toml +0 -5
loha_01/config_file.toml
CHANGED
@@ -1,8 +1,6 @@
|
|
1 |
[model_arguments]
|
2 |
-
project_name = project_name
|
3 |
v2 = false
|
4 |
v_parameterization = false
|
5 |
-
pretrained_model_name_or_path = pretrained_model_name_or_path
|
6 |
vae = "/content/VAE/VAE84EMA.vae.pt"
|
7 |
|
8 |
[additional_network_arguments]
|
@@ -31,8 +29,6 @@ debug_dataset = false
|
|
31 |
vae_batch_size = 1
|
32 |
|
33 |
[training_arguments]
|
34 |
-
output_dir = output_dir
|
35 |
-
output_name = project_name
|
36 |
save_precision = "fp16"
|
37 |
save_every_n_epochs = 10.0
|
38 |
train_batch_size = 6
|
@@ -47,7 +43,6 @@ gradient_accumulation_steps = 1
|
|
47 |
mixed_precision = "fp16"
|
48 |
clip_skip = 2
|
49 |
logging_dir = "/content/LoRA/log"
|
50 |
-
log_prefix = project_name
|
51 |
noise_offset = 0.01
|
52 |
lowram = false
|
53 |
|
|
|
1 |
[model_arguments]
|
|
|
2 |
v2 = false
|
3 |
v_parameterization = false
|
|
|
4 |
vae = "/content/VAE/VAE84EMA.vae.pt"
|
5 |
|
6 |
[additional_network_arguments]
|
|
|
29 |
vae_batch_size = 1
|
30 |
|
31 |
[training_arguments]
|
|
|
|
|
32 |
save_precision = "fp16"
|
33 |
save_every_n_epochs = 10.0
|
34 |
train_batch_size = 6
|
|
|
43 |
mixed_precision = "fp16"
|
44 |
clip_skip = 2
|
45 |
logging_dir = "/content/LoRA/log"
|
|
|
46 |
noise_offset = 0.01
|
47 |
lowram = false
|
48 |
|