Upload lora-scripts/config/lora.toml with huggingface_hub
Browse files
lora-scripts/config/lora.toml
ADDED
@@ -0,0 +1,62 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[model_arguments]
|
2 |
+
v2 = false
|
3 |
+
v_parameterization = false
|
4 |
+
pretrained_model_name_or_path = "./sd-models/model.ckpt"
|
5 |
+
|
6 |
+
[dataset_arguments]
|
7 |
+
train_data_dir = "./train/aki"
|
8 |
+
reg_data_dir = ""
|
9 |
+
resolution = "512,512"
|
10 |
+
prior_loss_weight = 1
|
11 |
+
|
12 |
+
[additional_network_arguments]
|
13 |
+
network_dim = 32
|
14 |
+
network_alpha = 16
|
15 |
+
network_train_unet_only = false
|
16 |
+
network_train_text_encoder_only = false
|
17 |
+
network_module = "networks.lora"
|
18 |
+
network_args = []
|
19 |
+
|
20 |
+
[optimizer_arguments]
|
21 |
+
unet_lr = 1e-4
|
22 |
+
text_encoder_lr = 1e-5
|
23 |
+
|
24 |
+
optimizer_type = "AdamW8bit"
|
25 |
+
lr_scheduler = "cosine_with_restarts"
|
26 |
+
lr_warmup_steps = 0
|
27 |
+
lr_restart_cycles = 1
|
28 |
+
|
29 |
+
[training_arguments]
|
30 |
+
train_batch_size = 1
|
31 |
+
noise_offset = 0.0
|
32 |
+
keep_tokens = 0
|
33 |
+
min_bucket_reso = 256
|
34 |
+
max_bucket_reso = 1024
|
35 |
+
caption_extension = ".txt"
|
36 |
+
max_token_length = 225
|
37 |
+
seed = 1337
|
38 |
+
xformers = true
|
39 |
+
lowram = false
|
40 |
+
max_train_epochs = 10
|
41 |
+
resolution = "512,512"
|
42 |
+
clip_skip = 2
|
43 |
+
mixed_precision = "fp16"
|
44 |
+
|
45 |
+
[sample_prompt_arguments]
|
46 |
+
sample_sampler = "euler_a"
|
47 |
+
sample_every_n_epochs = 5
|
48 |
+
|
49 |
+
[saving_arguments]
|
50 |
+
output_name = "output_name"
|
51 |
+
save_every_n_epochs = 1
|
52 |
+
save_state = false
|
53 |
+
save_model_as = "safetensors"
|
54 |
+
output_dir = "./output"
|
55 |
+
logging_dir = "./logs"
|
56 |
+
log_prefix = ""
|
57 |
+
save_precision = "fp16"
|
58 |
+
|
59 |
+
[others]
|
60 |
+
cache_latents = true
|
61 |
+
shuffle_caption = true
|
62 |
+
enable_bucket = true
|