khanon
commited on
Commit
•
7529c6f
0
Parent(s):
initial commit
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +34 -0
- .gitignore +1 -0
- README.md +3 -0
- atsuko/README.md +33 -0
- atsuko/lora_character_atsuko-v3_57i10r_768_batch3_2e-5_3epoch.json +43 -0
- atsuko/lora_character_atsuko-v4_57i11r_768_batch3_1e-5_3epoch.json +43 -0
- atsuko/lora_character_atsuko-v4_60i11r_768_batch3_1e-5_3epoch.json +43 -0
- atsuko/lora_character_atsuko-v5_60i10r_768_batch3_1e-5_4epoch.json +43 -0
- atsuko/lora_character_atsuko-v6_60i10r_768_batch3_2e-5_3epoch.json +43 -0
- atsuko/lora_character_atsuko_57i10r_768_batch3_slower3epoch.json +43 -0
- chise/README.md +31 -0
- chise/lora_chara_chise_v1_119i13r_832_batch3_5e-5text_2e-4unet_7epoch.json +50 -0
- hibiki/README.md +38 -0
- hibiki/lora_character_hibiki_119img9repeat_512.json +43 -0
- hibiki/lora_character_hibiki_119img9repeat_768.json +43 -0
- hibiki/lora_character_hibiki_split_65i6r-54i10r_768.json +43 -0
- hibiki/lora_character_hibiki_split_65i6r-54i10r_768_batch4_slower3epoch.json +43 -0
- hibiki/lora_character_hibiki_split_65i6r-54i10r_768_batch5_slower4epoch.json +43 -0
- hina/README.md +26 -0
- hina/lora_character_hina_106img10repeat_512.json +43 -0
- hina/lora_character_hina_106img10repeat_768.json +43 -0
- hina/lora_character_hina_106img8repeat_512.json +43 -0
- hina/lora_character_hina_106img8repeat_768.json +43 -0
- hina/lora_character_hina_106img9repeat_512.json +43 -0
- hina/lora_character_hina_106img9repeat_768.json +43 -0
- iroha/README.md +28 -0
- iroha/lora_character_iroha_95i7r_768_batch3_slower3epoch.json +43 -0
- iroha/lora_character_iroha_95i8r_768_batch3_slower3epoch.json +43 -0
- iroha/lora_character_iroha_95i9r_768_batch3_slower3epoch.json +43 -0
- iroha/lora_character_iroha_95i9r_768_batch4_slower3epoch.json +43 -0
- iroha/lora_character_irohav5_95i7r_768_batch3_midlr_3epoch.json +43 -0
- iroha/lora_character_irohav5mod_95i7r_768_batch3_midlr_3epoch.json +43 -0
- izuna/README.md +29 -0
- izuna/lora_character_izuna_67i10r_768_batch3_slower3epoch.json +43 -0
- kagura-nana/README.md +29 -0
- kagura-nana/lora_character_altnana-_105i20r_768_midlr_cosine.json +43 -0
- kagura-nana/lora_character_nana-v2_65i20r_768_midlr_cosine.json +43 -0
- kagura-nana/lora_character_nana-v3_65i20r_768_midlr_cosine-2e-5.json +43 -0
- kagura-nana/lora_character_nana-v4_split-61.7r-4.10r_768_cosine-1.5e-5.json +43 -0
- koharu/README.md +60 -0
- koharu/lora_character_koharu_v1_158i5r_768_batch3_5e-5text_1.5e-4unet_3epoch.json +43 -0
- koharu/lora_character_koharu_v2_180i6r-split_832_batch3_5e-5text_2e-4unet_3epoch.json +43 -0
- kokona/README.md +28 -0
- kokona/lora_character_kokona_95i7r_768_batch3_midlr_3epoch.json +43 -0
- mari/README.md +50 -0
- mari/lora_character_mari_v1_253i8r_832_batch3_5e-5text_2e-4unet_4epoch.json +48 -0
- mari/lora_character_mari_v2_253i6r_768-fixed-max-bucket-reso_batch4_5e-5text_5e-4unet_4epoch.json +48 -0
- mari/lora_character_mari_v3_253i8r_832_batch3_5e-5text_3e-4unet_4epoch.json +48 -0
- mari/lora_character_mari_v4_253i6r_832_batch3_5e-5text_2e-4unet_6epoch.json +48 -0
- mari/lora_character_mari_v4_253i6r_832_batch3_5e-5text_2e-4unet_7epoch.json +48 -0
.gitattributes
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
29 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
30 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
31 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
32 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
.gitignore
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
**/dataset/
|
README.md
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
---
|
atsuko/README.md
ADDED
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Hakari Atsuko (Blue Archive)
|
2 |
+
|
3 |
+
Note: her mask was not included in the training data. Stable Diffusion can't deal with them and a mixture of mask/no-mask Atsuko would fuck things up.
|
4 |
+
|
5 |
+
## Usage
|
6 |
+
Use any or all of these tags to summon Atsuko:
|
7 |
+
`1girl, halo, red eyes, pink hair, low twin braids`
|
8 |
+
Her hair is sometimes tagged pink, sometimes purple, even on Danbooru.
|
9 |
+
|
10 |
+
For her Arius outfit:
|
11 |
+
`hood, white dress, hair bow, black gloves, white kneehighs, frilled legwear`
|
12 |
+
|
13 |
+
For her leotard, just remove `white dress` and add `leotard` or `one-piece swimsuit` (WD unfortunately tagged it as a swimsuit more often than not).
|
14 |
+
|
15 |
+
The AI likes giving her fairly prominent lips/lipstick for some reason despite her mouth being basically just a line in most of the training data. It feels a little excessive so I just neg-prompt `lips` and it looks closer to correct.
|
16 |
+
|
17 |
+
You can add or ignore `atsuko, blue archive`; while they were in her captions, their effects aren't especially strong.
|
18 |
+
|
19 |
+
Weights from 0.9 - 1.2 work well depending on variant and model.
|
20 |
+
|
21 |
+
Variants:
|
22 |
+
- v3 -- 3 epochs, faster learning rate; slightly underfit but less likely to appear overcooked
|
23 |
+
- v5 -- 4 epochs, tweaked dataset, slower learning rate; does a great job with her outfit and halo but occasionally looks overcooked
|
24 |
+
- v6 -- v3 with v5's dataset; largely a wash, different but not necessarily better or worse.
|
25 |
+
Pick whichever, it doesn't really matter too much. Each one seems to have the chance of being the best for a given prompt so it's all RNG.
|
26 |
+
|
27 |
+
## Training
|
28 |
+
*All parameters are provided in the accompanying JSON files.*
|
29 |
+
- Trained on a set of 60 images initially provided by an /hdg/ anon, I pruned/added a few here and there.
|
30 |
+
- Dataset included a mixture of SFW and NSFW.
|
31 |
+
- Dataset was tagged with WD1.4 interrogator. Shuffling was enabled.
|
32 |
+
- `atsuko, blue archive` were added to the start of each caption. keep_tokens=2 was set to prevent shuffling those tokens.
|
33 |
+
- Three variations with different training params, check the JSON files if you care.
|
atsuko/lora_character_atsuko-v3_57i10r_768_batch3_2e-5_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/atsuko",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "6969",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "2e-5",
|
34 |
+
"unet_lr": "2e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "atsuko-v3-NAI-VAE-768px"
|
43 |
+
}
|
atsuko/lora_character_atsuko-v4_57i11r_768_batch3_1e-5_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/atsuko",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "6969",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "2e-5",
|
34 |
+
"unet_lr": "2e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "atsuko-v4-NAI-VAE-768px"
|
43 |
+
}
|
atsuko/lora_character_atsuko-v4_60i11r_768_batch3_1e-5_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/atsuko",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "6969",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "atsuko-v4-NAI-VAE-768px"
|
43 |
+
}
|
atsuko/lora_character_atsuko-v5_60i10r_768_batch3_1e-5_4epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/atsuko",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "4",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "6969",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "atsuko-v5-NAI-VAE-768px"
|
43 |
+
}
|
atsuko/lora_character_atsuko-v6_60i10r_768_batch3_2e-5_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/atsuko",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "6969",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "2e-5",
|
34 |
+
"unet_lr": "2e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "atsuko-v6-NAI-VAE-768px"
|
43 |
+
}
|
atsuko/lora_character_atsuko_57i10r_768_batch3_slower3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/atsuko",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1.2e-5",
|
34 |
+
"unet_lr": "1.2e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "atsuko-NAI-VAE-768px"
|
43 |
+
}
|
chise/README.md
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Waraku Chise (Blue Archive)
|
2 |
+
Came out pretty well I think. Smaller dataset than Mari, but otherwise very similar settings.
|
3 |
+
|
4 |
+
## Usage
|
5 |
+
Use any or all of these tags to summon Chise:
|
6 |
+
`chise, halo, red eyes, blue hair`
|
7 |
+
Hair and eyes are mostly optional if you describe a bit of her outfit as well.
|
8 |
+
She naturally likes to make her `:o` expression because most art features her doing that. However I also included images tagged with other expressions.
|
9 |
+
Use `open mouth`, `closed mouth`, and `parted lips` as necessary to get her to make whatever expressions you want.
|
10 |
+
|
11 |
+
For her normal outfit (add as many as necessary):
|
12 |
+
`braid, japanese clothes, detached sleeves, obi, tabi, geta, sailor collar, blue bow`
|
13 |
+
Her shoes are weird but they're tagged `geta` and the socks as `tabi`.
|
14 |
+
|
15 |
+
For her swimsuit outfit (add as many as necessary):
|
16 |
+
`side ponytail, swimsuit, striped bikini, see-through, sailor collar, side-tie bikini bottom`
|
17 |
+
You can also add `pointed ears`, which are usually only visible in her swimsuit outfit.
|
18 |
+
|
19 |
+
Weight 1 works fine. Also included epoch 6 in case you find the last epoch to be a bit stubborn with her outfits.
|
20 |
+
|
21 |
+
## Training
|
22 |
+
*All parameters are provided in the accompanying JSON files.*
|
23 |
+
- Trained on a set of 119 images split by outfit, repeated 10 times (swimsuit) or 14 times (uniform) for 7 epochs (119 images * 10 or 14 repeats / 3 batch size * 7 epochs = 3178 steps)
|
24 |
+
- Dataset included a mixture of SFW and NSFW.
|
25 |
+
- Initially tagged with WD1.4 VITv2 model, then performed heavy pruning and editing.
|
26 |
+
- Pruned implicit (`oni horns`) or redundant tags and simplified outfits so that they were always tagged with the same handful of tags
|
27 |
+
- Made sure important traits were present and consitently described, and traits like `halo` were consistent with actual visibility
|
28 |
+
- Added many facial expression, camera angle, and image composition hints
|
29 |
+
- Used network_dimension 128 (same as usual) and network_alpha 64 (new)
|
30 |
+
- This relies on the new alpha
|
31 |
+
- Trained without VAE.
|
chise/lora_chara_chise_v1_119i13r_832_batch3_5e-5text_2e-4unet_7epoch.json
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/chise",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/lora/trained/chara/chise",
|
9 |
+
"max_resolution": "832,832",
|
10 |
+
"learning_rate": "",
|
11 |
+
"lr_scheduler": "cosine_with_restarts",
|
12 |
+
"lr_warmup": "5",
|
13 |
+
"train_batch_size": 3,
|
14 |
+
"epoch": "7",
|
15 |
+
"save_every_n_epochs": "2",
|
16 |
+
"mixed_precision": "fp16",
|
17 |
+
"save_precision": "fp16",
|
18 |
+
"seed": "31337",
|
19 |
+
"num_cpu_threads_per_process": 32,
|
20 |
+
"cache_latents": true,
|
21 |
+
"caption_extension": "",
|
22 |
+
"enable_bucket": true,
|
23 |
+
"gradient_checkpointing": false,
|
24 |
+
"full_fp16": false,
|
25 |
+
"no_token_padding": false,
|
26 |
+
"stop_text_encoder_training": 0,
|
27 |
+
"use_8bit_adam": true,
|
28 |
+
"xformers": true,
|
29 |
+
"save_model_as": "safetensors",
|
30 |
+
"shuffle_caption": true,
|
31 |
+
"save_state": false,
|
32 |
+
"resume": "",
|
33 |
+
"prior_loss_weight": 1.0,
|
34 |
+
"text_encoder_lr": "5e-5",
|
35 |
+
"unet_lr": "2e-4",
|
36 |
+
"network_dim": 128,
|
37 |
+
"lora_network_weights": "",
|
38 |
+
"color_aug": false,
|
39 |
+
"flip_aug": false,
|
40 |
+
"clip_skip": 2,
|
41 |
+
"gradient_accumulation_steps": 1.0,
|
42 |
+
"mem_eff_attn": false,
|
43 |
+
"output_name": "chise-v1-NoVAE",
|
44 |
+
"model_list": "",
|
45 |
+
"max_token_length": "150",
|
46 |
+
"max_train_epochs": "",
|
47 |
+
"max_data_loader_n_workers": "",
|
48 |
+
"network_alpha": 64,
|
49 |
+
"training_comment": ""
|
50 |
+
}
|
hibiki/README.md
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Nekozuka Hibiki (Blue Archive)
|
2 |
+
|
3 |
+
## Usage
|
4 |
+
*Important:* This is a fairly temperamental LoRA due to the dataset, and it needs some wrangling to get good results. It won't look good with vanilla NAI and the standard NAI negative prompt, despite being trained on nai-animefull-final.
|
5 |
+
- Use a strong negative prompt. Consider using the bad_prompt_v2 embed at a reduced strength to dramatically improve things, though it does affect the style somewhat.
|
6 |
+
- Use a strong model. AbyssOrangeMix2 and nutmegmix work well with this LoRA.
|
7 |
+
- Use the negative prompt liberally to suppress cheerleader Hibiki if you don't want her, otherwise her traits tend to take over.
|
8 |
+
|
9 |
+
To summon Hibiki, use the following tags. Adjust strength as needed.
|
10 |
+
- `1girl, halo, black hair, blue eyes, bright pupils, animal ears, dog girl, tail, hair bobbles, goggles, eyewear on head, medium breasts`
|
11 |
+
|
12 |
+
For regular Hibiki, add the following tags. Adjust strength as needed.
|
13 |
+
- Prompt: `(black camisole, jacket, black shorts:1.2), (fishnet legwear:1.1)`
|
14 |
+
- Negative prompt: `(midriff, navel, skin tight, tight:1.4), (tattoo, arm tattoo, star sticker:1.30), ski goggles, wavy mouth, embarrassed`
|
15 |
+
- Cheerleader Hibiki tends to have a permanent embarrassed/wavy mouth expression unless you use negative tags to get rid of it.
|
16 |
+
|
17 |
+
For cheerleader Hibiki, add the following tags.
|
18 |
+
- Prompt: `cheerleader, midriff, embarrassed, wavy mouth, crop top, white pleated skirt`
|
19 |
+
- Negative prompt: `fishnets`
|
20 |
+
|
21 |
+
You can add or ignore `hibiki, blue archive`; while they were in her captions, they don't have an especially strong effect.
|
22 |
+
|
23 |
+
Adjust the weight as needed. Weights from 0.95 up to 1.25 work well (higher weights may summon Cheerleader Hibiki unintentionally).
|
24 |
+
|
25 |
+
## Training
|
26 |
+
*All parameters are provided in the accompanying JSON files.*
|
27 |
+
- Trained on 119 images, separated into two sets.
|
28 |
+
- Hibiki (regular) -- 54 images, 10 repeats
|
29 |
+
- Hibiki (cheerleader) -- 65 images, 6 repeats
|
30 |
+
- Dataset included a mixture of SFW and NSFW. Mostly SFW.
|
31 |
+
- As you can guess, her cheerleader alt makes up the vast majority of her art, and artists are more consistent when drawing her. Training them all together did not work well, so I had to split up the datasets.
|
32 |
+
- Dataset was tagged with WD1.4 interrogator. Shuffling was enabled.
|
33 |
+
- `hibiki, blue archive` were added to the start of each caption. keep_tokens=2 was set to prevent shuffling those tokens.
|
34 |
+
- Her cheerleader outfit is much more easily recognized by the tagger, leading to stronger tags. Even human artists can't seem to agree on what she's wearing in her normal outfit.
|
35 |
+
- Trained at 768px resolution. I stopped training a 512 variant because it was almost always worse and added to training time.
|
36 |
+
- Two variants included.
|
37 |
+
- v1: first attempt at splitting the dataset. Works well, but not as coherent in some ways (halo particularly) and still tends to blend details from her two outfits.
|
38 |
+
- v2: using the split dataset. Increased batch size slightly (3 >>> 4), reduced learning rate (3e-5 >>> 1e-5), increased epochs (1 >>> 3). Generally an improvement, though sometimes v1 is easier to wrangle.
|
hibiki/lora_character_hibiki_119img9repeat_512.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hibiki",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "512,512",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hibiki-NAI-VAE-512px-357steps"
|
43 |
+
}
|
hibiki/lora_character_hibiki_119img9repeat_768.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hibiki",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hibiki-NAI-VAE-768px-357steps"
|
43 |
+
}
|
hibiki/lora_character_hibiki_split_65i6r-54i10r_768.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hibiki",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hibiki-v2-NAI-VAE-768px-6.10split-310steps"
|
43 |
+
}
|
hibiki/lora_character_hibiki_split_65i6r-54i10r_768_batch4_slower3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hibiki",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 4,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hibiki-v3-NAI-VAE-768px-6.10split-310steps-batch4-slower"
|
43 |
+
}
|
hibiki/lora_character_hibiki_split_65i6r-54i10r_768_batch5_slower4epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hibiki",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 5,
|
13 |
+
"epoch": "4",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hibiki-v3-NAI-VAE-768px-6.10split-310steps-batch5-slower-4epoch"
|
43 |
+
}
|
hina/README.md
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Sorasaki Hina (Blue Archive)
|
2 |
+
|
3 |
+
## Training
|
4 |
+
*All parameters are provided in the accompanying JSON files.*
|
5 |
+
- Trained on 106 images curated by anon on /hdg/, repeated 10 times (1060 total images / 3 batchsize = ~354 iterations)
|
6 |
+
- Dataset included a mixture of SFW and NSFW
|
7 |
+
- Dataset included a lot of her swimsuit alt outfit, so it has a bit of a harder time producing her standard outfit (the military uniform)
|
8 |
+
- It does better with her more casual outfit ("sleeveless shirt, miniskirt") and obviously very well with her school swimsuit
|
9 |
+
- Dataset was tagged with WD1.4 interrogator. Shuffling was enabled.
|
10 |
+
- `hina, blue archive` were added to the start of each caption. keep_tokens=2 was set to prevent shuffling those tokens.
|
11 |
+
- Two resolution variants included (512px, 768px)
|
12 |
+
- The 768px version seems just generally better to my eyes but not overwhelmingly. They're definitely quite different so I've kept both.
|
13 |
+
|
14 |
+
## Usage
|
15 |
+
Hina needs a few tags to be summoned reliably. Some common tags in her dataset:
|
16 |
+
`1girl, halo, parted bangs, demon horns, long hair, white hair, purple eyes, low wings, small breasts`
|
17 |
+
|
18 |
+
You can add or ignore `hina, blue archive`; while they were in her captions, they don't seem to be particularly strong for some reason.
|
19 |
+
|
20 |
+
You can use the 512px or 768px variants. I want to say the 768px one is better, but it's hard to say definitively. Give both a shot and post your findings.
|
21 |
+
|
22 |
+
Weight 0.85-1.05 should work well depending on model.
|
23 |
+
|
24 |
+
## Dataset
|
25 |
+
Dataset curated by anon on /hdg/: https://mega.nz/folder/ayhWjJDA#xDGSB3T-yRnK6aRajsJ05g
|
26 |
+
/h/thread/7101537
|
hina/lora_character_hina_106img10repeat_512.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hina",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "512,512",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hina-NAI-VAE-512px-354steps"
|
43 |
+
}
|
hina/lora_character_hina_106img10repeat_768.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hina",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hina-NAI-VAE-noreg-768-10reps"
|
43 |
+
}
|
hina/lora_character_hina_106img8repeat_512.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hina",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "512,512",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hina-NAI-VAE-noreg-512"
|
43 |
+
}
|
hina/lora_character_hina_106img8repeat_768.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hina",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hina-NAI-VAE-noreg-768"
|
43 |
+
}
|
hina/lora_character_hina_106img9repeat_512.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hina",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "512,512",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hina-NAI-VAE-noreg-512-moresteps"
|
43 |
+
}
|
hina/lora_character_hina_106img9repeat_768.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/hina",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "3e-5",
|
34 |
+
"unet_lr": "3e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "hina-NAI-VAE-noreg-768-moresteps"
|
43 |
+
}
|
iroha/README.md
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Natsume Iroha (Blue Archive)
|
2 |
+
|
3 |
+
## Usage
|
4 |
+
Use any or all of these tags to summon Iroha:
|
5 |
+
`1girl, halo, red hair, very long hair, wavy hair, grey eyes, small breasts`
|
6 |
+
|
7 |
+
For her normal outfit:
|
8 |
+
`peacked cap, open jacket, collared shirt, partially unbuttoned, red necktie, armband, sleeves past wrists`
|
9 |
+
|
10 |
+
You can add or ignore `iroha, blue archive`; while they were in her captions, their effects aren't especially strong.
|
11 |
+
|
12 |
+
The network seems simultaneously undertrained and overfitted in some ways so weights are kind of all over the place, just play around with them. 0.8-1.2 work alright.
|
13 |
+
|
14 |
+
## Training
|
15 |
+
*All parameters are provided in the accompanying JSON files.*
|
16 |
+
- Trained on 95 images using a modified dataset curated by anon on /hdg/, repeated 7 times (665 total images / 3 batchsize = ~222 iterations per epoch, 3 epochs)
|
17 |
+
- Dataset included a mixture of SFW and NSFW
|
18 |
+
- Sightly increased learning rate from the last few (1e-5 >>> 1.5e-5)
|
19 |
+
- Seems a bit underfitted at times, overfitted others.
|
20 |
+
- Frequently forgets her jacket/shirt colors despite them being a very consistent part of her outfit, and her halo is pretty incoherent.
|
21 |
+
- She really likes doing a particular kind of toothy smirk that was too common in her dataset. When you don't want it, discourage it with `closed mouth` and negative `open mouth`.
|
22 |
+
- Dataset was tagged with WD1.4 interrogator. Shuffling was enabled.
|
23 |
+
- `iroha, blue archive` were added to the start of each caption. keep_tokens=2 was set to prevent shuffling those tokens.
|
24 |
+
- No variants. I trained 8 iterations of this LoRA and they all had issues of some kind, this latest one is the best of them.
|
25 |
+
|
26 |
+
## Dataset
|
27 |
+
Dataset curated by anon on /hdg/: https://mega.nz/folder/Hqg10AAI#9xsrFmuA8oqVBlhsj95y6w
|
28 |
+
/h/thread/7104198
|
iroha/lora_character_iroha_95i7r_768_batch3_slower3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/iroha",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "iroha-v1-NAI-VAE-768px"
|
43 |
+
}
|
iroha/lora_character_iroha_95i8r_768_batch3_slower3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/iroha",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "iroha-v2.2-NAI-VAE-768px"
|
43 |
+
}
|
iroha/lora_character_iroha_95i9r_768_batch3_slower3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/iroha",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "1234",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "iroha-v3-NAI-VAE-768px"
|
43 |
+
}
|
iroha/lora_character_iroha_95i9r_768_batch4_slower3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/iroha",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 4,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "1234",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "iroha-v3.1-NAI-VAE-768px"
|
43 |
+
}
|
iroha/lora_character_irohav5_95i7r_768_batch3_midlr_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/iroha",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1.5e-5",
|
34 |
+
"unet_lr": "1.5e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "iroha-v5-NAI-VAE-768px"
|
43 |
+
}
|
iroha/lora_character_irohav5mod_95i7r_768_batch3_midlr_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/iroha",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1.5e-5",
|
34 |
+
"unet_lr": "1.5e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "iroha-v5-mod-NAI-VAE-768px"
|
43 |
+
}
|
izuna/README.md
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Kuda Izuna (Blue Archive)
|
2 |
+
|
3 |
+
## Usage
|
4 |
+
Use any or all of these tags to summon Izuna:
|
5 |
+
`1girl, halo, yellow eyes, brown hair, animal ears, fox girl, fox tail, fox hair ornament, bangs, medium breasts, red eyeshadow`
|
6 |
+
|
7 |
+
For her normal outfit:
|
8 |
+
`blue sailor collar, red neckerchief, partially fingerless gloves, ninja, floral print, fishnets, single thighhigh`
|
9 |
+
|
10 |
+
For her swimsuit outfit:
|
11 |
+
`striped bikini, denim shorts, visor cap, highleg bikini, red neckerchief`
|
12 |
+
|
13 |
+
You can add or ignore `izuna, blue archive`; while they were in her captions, their effects aren't especially strong.
|
14 |
+
|
15 |
+
Weight 0.90-1.2 should work well depending on model.
|
16 |
+
|
17 |
+
## Training
|
18 |
+
*All parameters are provided in the accompanying JSON files.*
|
19 |
+
- Trained on 67 images curated by anon on /hdg/, repeated 10 times (670 total images / 3 batchsize = ~223 iterations per epoch)
|
20 |
+
- Dataset included a mixture of SFW and NSFW
|
21 |
+
- Dataset included a lot of her swimsuit alt outfit but it seems to be able to produce her normal outfit readily.
|
22 |
+
- Reduced learning rate (3e-5 >>> 1e-5) and increased epochs (1 >>> 3). I'll likely do this from now on since it seems to produce more consistent results.
|
23 |
+
- Dataset was tagged with WD1.4 interrogator. Shuffling was enabled.
|
24 |
+
- `izuna, blue archive` were added to the start of each caption. keep_tokens=2 was set to prevent shuffling those tokens.
|
25 |
+
- No variants, she was a good girl.
|
26 |
+
|
27 |
+
## Dataset
|
28 |
+
Dataset curated by anon on /hdg/: https://mega.nz/folder/Hqg10AAI#9xsrFmuA8oqVBlhsj95y6w
|
29 |
+
/h/thread/7104198
|
izuna/lora_character_izuna_67i10r_768_batch3_slower3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/izuna",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "23",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1e-5",
|
34 |
+
"unet_lr": "1e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "izuna-v1-NAI-VAE-768px"
|
43 |
+
}
|
kagura-nana/README.md
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Kagura Nana, alt outfit (Virtual Youtuber)
|
2 |
+
Request for anon on /hdg/. Sorta rough, not much art for her and especially NSFW art.
|
3 |
+
|
4 |
+
## Usage
|
5 |
+
|
6 |
+
Tags: `1girl, cat ears, tail, two side up, two-tone hair, blue eyes, small breasts`
|
7 |
+
|
8 |
+
Outfit: `sleeveless dress, off shoulder, jacket, choker, neck bell, striped bow, skirt`
|
9 |
+
|
10 |
+
The original version of this was extremely overfit due to bad training data, and was highly resistant towards nudity because there was literally none of it in the dataset. I did some pruning, generated/found a couple of nude images, and cleaned up tags -- now her outfit is manually tagged and exclusively described using the tags above so it is generally more composable (ie. you can prompt for the `jacket` but negative prompt the `sleeveless dress` and it will work fairly reliably). I included `skirt` in images where the flared, ruffled bottom of her dress was visible even though it's not really a skirt but it may help with outfit composition.
|
11 |
+
|
12 |
+
It's still slightly overtrained so turning down the weights a bit (0.85 - 1) is recommended.
|
13 |
+
|
14 |
+
Three variations:
|
15 |
+
- v1: the original, in the `old` folder. Not really recommended unless you want it to hyperfocus on her details and outfit at the cost of NSFW capability. Might be okay for mixing?
|
16 |
+
- v3: new cleaned up dataset, somewhat overcooked
|
17 |
+
- v4: same as v3 with reduced steps, except for the small number of NSFW images which were boosted
|
18 |
+
|
19 |
+
I did something different with this one and tagged the four lewd images in her dataset with `nsfw`. I can't really tell if it helped or hurt, but you may as well prompt for that if you want porn.
|
20 |
+
|
21 |
+
## Training
|
22 |
+
*All parameters are provided in the accompanying JSON files.*
|
23 |
+
- Trained on a set of 67 images initially provided by an /hdg/ anon, heavily curated and re-tagged.
|
24 |
+
- Dataset included a mixture of SFW and NSFW.
|
25 |
+
- Some NSFW images were outputs from v1 of the model. Does not appear to have caused down syndrome Nanas from model inbreeding.
|
26 |
+
- Dataset was initially tagged with WD1.4 interrogator and then heavily pruned and manually cleaned up.
|
27 |
+
- `kagura nana, virtual youtuber` were added to the start of each caption. keep_tokens=2 was set to prevent shuffling those tokens.
|
28 |
+
- All ambiguities or alternate tags for her outfit were removed and replaced with the standard set above. Prompt for `sleeveless dress` not `sleeveless shirt`, and `jacket` not `black jacket` for the highest correlation.
|
29 |
+
- Three variations with different training params, check the JSON files if you care.
|
kagura-nana/lora_character_altnana-_105i20r_768_midlr_cosine.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/kagura-nana",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "cosine_with_restarts",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1.5e-5",
|
34 |
+
"unet_lr": "1.5e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "nana-v1-NAI-VAE-768px"
|
43 |
+
}
|
kagura-nana/lora_character_nana-v2_65i20r_768_midlr_cosine.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/kagura-nana",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "cosine_with_restarts",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1.5e-5",
|
34 |
+
"unet_lr": "1.5e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "nana-v2-NAI-VAE-768px"
|
43 |
+
}
|
kagura-nana/lora_character_nana-v3_65i20r_768_midlr_cosine-2e-5.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/kagura-nana",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "cosine_with_restarts",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "1",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "2e-5",
|
34 |
+
"unet_lr": "2e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "nana-v3-NAI-VAE-768px"
|
43 |
+
}
|
kagura-nana/lora_character_nana-v4_split-61.7r-4.10r_768_cosine-1.5e-5.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/kagura-nana",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "cosine_with_restarts",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1.5e-5",
|
34 |
+
"unet_lr": "1.5e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "nana-v4-NAI-VAE-768px"
|
43 |
+
}
|
koharu/README.md
ADDED
@@ -0,0 +1,60 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Shimoe Koharu (Blue Archive)
|
2 |
+
|
3 |
+
Changed training methodology around for Koharu. It took way more time and effort due to the degree of manual tagging involved, but it turned out pretty well.
|
4 |
+
|
5 |
+
I'll probably return to this one later to make further improvements now that I've got a much better handle on the impact of tagging and how to get the most out of larger datasets. I don't expect to manual tag every future student, though.
|
6 |
+
|
7 |
+
## Usage
|
8 |
+
Use any or all of these tags to summon Koharu:
|
9 |
+
`koharu, 1girl, halo, pink eyes, ringed eyes, head wings, low wings, pink hair`
|
10 |
+
Unlike previous LoRAs, the character's name does help this one somewhat. You can probably omit her hair to save tokens.
|
11 |
+
|
12 |
+
The vertical line running down her body appears consistently, but may not always reach past her chest because artists are inconsistent in how they draw it. You can try to describe it literally: "vertical black line running past navel" or whatever. Don't try `tattoo` unless you want womb tattoos.
|
13 |
+
|
14 |
+
It does a decent, but not perfect job with her eyes. Adding some combination of `embarrassed`, `open mouth`, `swirly eyes` with varying degrees of emphasis can draw out her characteristic horny retard look.
|
15 |
+
|
16 |
+
I tried to add the slit pupils expression and the model sorta gets it, but not very well. You can prompt it with `slit pupils` and `flustered` but it generally creates abominations.
|
17 |
+
|
18 |
+
For her normal Trinity outfit:
|
19 |
+
`school uniform, off shoulder, hat, skirt`
|
20 |
+
|
21 |
+
Some of her swimsuits are in there too.
|
22 |
+
|
23 |
+
Weights from 0.8 - 1.05 should work well.
|
24 |
+
|
25 |
+
### Important
|
26 |
+
This LoRA may be more aggressive than others in forcing a close-up/portrait camera. I believe this is because I scraped Booru tags for this one, and WD1.4 more reliably tags camera angles and image composition than human taggers. You can mitigate this by always prompting for an angle or composition tag, like `above waist` or `cowboy shot` or `from above`. You can combine them, too.
|
27 |
+
|
28 |
+
Trying to prompt Koharu from behind or the side generally doesn't work very well -- it can render her back if you use `from behind` and `back focus`, but her wings will be attached to her stomach and her halo will be flipped,because the AI doesn't know how to generalize those traits to different angles and there's not enough training data for them.
|
29 |
+
|
30 |
+
## Training
|
31 |
+
*All parameters are provided in the accompanying JSON files.*
|
32 |
+
|
33 |
+
Koharu's training was handled substantially differently.
|
34 |
+
|
35 |
+
- Trained on a heavily curated set of 183 images, most repeated 6 times. 1150 total steps.
|
36 |
+
- Dataset included a mixture of SFW and NSFW.
|
37 |
+
- Doubled the number of steps because the dataset was larger than usual. I typically target 450 - 650.
|
38 |
+
- New tagging methodology. No WD1.4 tags; instead I scraped tags from Sankaku Complex using Hydrus and manually cleaned them up.
|
39 |
+
- Removed tons of shit tags
|
40 |
+
- Made sure important traits were present and consitently described, and traits like `halo` were consistent with actual visibility
|
41 |
+
- Pruned lots of redundant tags and simplified outfits. There is no `black serafuku, long sleeves`, only Koharu's `school uniform`.
|
42 |
+
- Added camera angles and image composition hints
|
43 |
+
- Added facial expressions (particularly `embarrassed`) and unusual pupils when present
|
44 |
+
- Different learning rate than usual.
|
45 |
+
- 5e-5 text encoder (typically 1e-5 ~ 2e-5)
|
46 |
+
- 2e-4 UNet (typically one order of magnitude faster than text)
|
47 |
+
- This was experimental -- human tags tend to be more varied, allowing for more expressiveness (WD1.4 did not do a good job with her) but potentially requiring more training. The dataset was also larger.
|
48 |
+
- VAE removed. I usually train the dataset on the NAI VAE but after some tests, I think this was leading to oversaturated outputs and it does not play nicely with alternative VAEs.
|
49 |
+
- May offer a No VAE and a WD1.4 VAE in the future as these seem to present the best results across many configurations
|
50 |
+
|
51 |
+
While I think the experimental things I tried out with this dataset worked out well enough to be called a success, tag cleanup took literal hours and I will probably not be able to put nearly so much effort into every character. I just really like Koharu. I will probably retrain some old ones with at least the new hyperparameter methodologies, though.
|
52 |
+
|
53 |
+
## To-do
|
54 |
+
- More consistently tag NSFW/SFW/nudity
|
55 |
+
- Add more image composition/camera angle tags
|
56 |
+
- Find additional images with prominent swirly eyes
|
57 |
+
- Improve tags for socks/shoes
|
58 |
+
- Remove `halo` tag from images where it is just barely visible to force camera to pull further away
|
59 |
+
- Un-fuck wings from side angle (folded wings tag?)
|
60 |
+
- Add `looking away` / `facing away` to applicable images because it is impossible
|
koharu/lora_character_koharu_v1_158i5r_768_batch3_5e-5text_1.5e-4unet_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/koharu",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "cosine_with_restarts",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "5e-5",
|
34 |
+
"unet_lr": "1.5e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "koharu-v1-NoVAE"
|
43 |
+
}
|
koharu/lora_character_koharu_v2_180i6r-split_832_batch3_5e-5text_2e-4unet_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/koharu",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "832,832",
|
10 |
+
"lr_scheduler": "cosine_with_restarts",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "5e-5",
|
34 |
+
"unet_lr": "2e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "koharu-v2-NoVAE"
|
43 |
+
}
|
kokona/README.md
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Sunohara Kokona (Blue Archive)
|
2 |
+
|
3 |
+
## Usage
|
4 |
+
Use any or all of these tags to summon Kokona:
|
5 |
+
`1girl, halo, animal ears, orange eyes, long hair, grey hair`
|
6 |
+
She really doesn't need many, just `1girl, halo` is enough to summon an instantly recognizable version of her. The model may be a little overtrained but it looks good.
|
7 |
+
|
8 |
+
For her normal outfit:
|
9 |
+
`china dress, vertical-striped dress, white skirt, frilled skirt, off-shoulder, jacket`
|
10 |
+
|
11 |
+
Somewhat frequently, a strange, mysterious man in a black suit likes to appear out-of-frame beside her. You can shoo him away with `1boy` in the negative prompt. Not sure why this happens as there was nothing like it in the training data.
|
12 |
+
|
13 |
+
The AI is generally pretty good at picturing her, but it really struggles with her pout. Try various combinations of `pout`, `:t`, `pouting` and negative prompt `wavy mouth` to try to get it to appear correctly, or just use img2img without the LoRA. It seems to be related to the LoRA since vanilla NAI can do pouting just fine.
|
14 |
+
|
15 |
+
Her stamp is in the training data and it sorta knows what it is but is probably too small to be recognized by the AI.
|
16 |
+
|
17 |
+
You can add or ignore `kokona, blue archive`; while they were in her captions, their effects aren't especially strong.
|
18 |
+
|
19 |
+
Weights close to 1 work well with her, she's very cooperative.
|
20 |
+
|
21 |
+
## Training
|
22 |
+
*All parameters are provided in the accompanying JSON files.*
|
23 |
+
- Trained on 95 curated images, repeated 7 times (665 total images / 3 batchsize = ~222 iterations per epoch, 3 epochs)
|
24 |
+
- Dataset included a mixture of SFW and NSFW.
|
25 |
+
- 1.5e-5 learning rate, same as Iroha.
|
26 |
+
- Dataset was tagged with WD1.4 interrogator. Shuffling was enabled.
|
27 |
+
- `kokona, blue archive` were added to the start of each caption. keep_tokens=2 was set to prevent shuffling those tokens.
|
28 |
+
- No variants, she's well-behaved.
|
kokona/lora_character_kokona_95i7r_768_batch3_midlr_3epoch.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/kokona",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"lr_scheduler": "constant_with_warmup",
|
11 |
+
"lr_warmup": "5",
|
12 |
+
"train_batch_size": 3,
|
13 |
+
"epoch": "3",
|
14 |
+
"save_every_n_epochs": "1",
|
15 |
+
"mixed_precision": "fp16",
|
16 |
+
"save_precision": "fp16",
|
17 |
+
"seed": "31337",
|
18 |
+
"num_cpu_threads_per_process": 32,
|
19 |
+
"cache_latent": true,
|
20 |
+
"caption_extention": ".txt",
|
21 |
+
"enable_bucket": true,
|
22 |
+
"gradient_checkpointing": false,
|
23 |
+
"full_fp16": false,
|
24 |
+
"no_token_padding": false,
|
25 |
+
"stop_text_encoder_training": 0,
|
26 |
+
"use_8bit_adam": true,
|
27 |
+
"xformers": true,
|
28 |
+
"save_model_as": "safetensors",
|
29 |
+
"shuffle_caption": true,
|
30 |
+
"save_state": false,
|
31 |
+
"resume": "",
|
32 |
+
"prior_loss_weight": 1.0,
|
33 |
+
"text_encoder_lr": "1.5e-5",
|
34 |
+
"unet_lr": "1.5e-4",
|
35 |
+
"network_dim": 128,
|
36 |
+
"lora_network_weights": "",
|
37 |
+
"color_aug": false,
|
38 |
+
"flip_aug": false,
|
39 |
+
"clip_skip": 2,
|
40 |
+
"gradient_accumulation_steps": 1.0,
|
41 |
+
"mem_eff_attn": false,
|
42 |
+
"output_name": "kokona-NAI-VAE-768px"
|
43 |
+
}
|
mari/README.md
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Iochi Mari (Blue Archive)
|
2 |
+
I used a bigger dataset than normal since Mari has a lot of good art (253 images).
|
3 |
+
|
4 |
+
This was a tricky LoRA to train since she has two very distinct outfits (and her gym outfit has as much art without her jacket as with). The LoRA recalls both outfits, and even her swimsuit, but you may need to negative prompt elements of whichever outfit you don't want.
|
5 |
+
|
6 |
+
## Usage
|
7 |
+
Use any or all of these tags to summon Mari:
|
8 |
+
`mari, halo, 1girl, blue eyes`
|
9 |
+
`cat ears` was only tagged on images where her cowl was not present, in which case `animal ear headwear` is the tag you want.
|
10 |
+
Hair and eyes are optional.
|
11 |
+
|
12 |
+
For her normal outfit (you don't usually need all of these):
|
13 |
+
`single braid, animal ear headwear, nun, habit, puffy long sleeves, blue neckerchief, white sailor collar`
|
14 |
+
Her shoes were tagged `mary janes`.
|
15 |
+
The sleeves of her outfit leak into other outfits/nudity so you may need to negative prompt them (`puffy long sleeves`)
|
16 |
+
|
17 |
+
For her gym uniform outfit (you don't usually need all of these):
|
18 |
+
`low ponytail, sportswear, track jacket, shorts, gym shirt, ponytail, id card, bottle`
|
19 |
+
Images with her jacket visible were tagged `track jacket`. If the jacket was open, I added `open jacket`.
|
20 |
+
Images with her gym shirt visible (so not those where her jacket was closed) were tagged `gym shirt`. Adding `white` can help.
|
21 |
+
`wet, wet clothes` was common in the trianing data and works well.
|
22 |
+
|
23 |
+
For her swimsuit (only a handful of images in the training data, so use emphasis):
|
24 |
+
`black one-piece swimsuit, frilled swimsuit, frills, paw print`
|
25 |
+
Negative prompt: `competition swimsuit`
|
26 |
+
|
27 |
+
Included three epochs. I did way more steps than normal in pursuit of perfect halo gens,
|
28 |
+
- epoch 7 (3542 steps) = nails her expressions and outfits pretty well but outfit details can leak over. best halo but still not great
|
29 |
+
- epoch 6 (~3000 steps) = pretty good balance between the others
|
30 |
+
- epoch 4 (~2000 steps) = less overfit on outfits but halo and expression consistency is not as good
|
31 |
+
|
32 |
+
Weight 1 works fine for the most part. Epoch 7 may be somewhat stubborn/overfit on some parts of her outfit, in which case you may have better luck dropping down to epoch 6 or 4.
|
33 |
+
|
34 |
+
|
35 |
+
## Training
|
36 |
+
*All parameters are provided in the accompanying JSON files.*
|
37 |
+
- Trained on a set of 253 images repeated 6 times, 7 epochs (253 images * 6 repeats / 3 batch size * 7 epochs = 3542 steps)
|
38 |
+
- Dataset included a mixture of SFW and NSFW.
|
39 |
+
- Same params as Koharu, but with almost double the amount of steps, since the dataset was large and I wasn't initially satisfied with halo/facial expression consistency
|
40 |
+
- Initially tagged with WD1.4, then performed heavy pruning and editing.
|
41 |
+
- Pruned redundant tags and simplified outfits so that they were always tagged with the same handful of tags
|
42 |
+
- Made sure important traits were present and consitently described, and traits like `halo` were consistent with actual visibility
|
43 |
+
- Added many facial expression, camera angle, and image composition hints
|
44 |
+
- Made a hard-coded tweak to bmaltais's LoRA GUI script.
|
45 |
+
- Previously increasing the `resolution` values (512 >> 768 >> 832) would not increase `max_bucket_reso`, so buckets were limited to 1024 (the default) in either dimension.
|
46 |
+
- Per the kohya_ss README, when increasing `resolution` to 768, it is recommended to increase `max_bucket_reso` to 1280 to keep the bucket aspect ratios the same as 512 resolution
|
47 |
+
- I adjusted the script to set `max_bucket_reso` based on `resolution`, using the formula:
|
48 |
+
- `max_bucket_reso = resolution * (1 + (512 / resolution))` which causes bucket aspect ratios to be the same as 512 for any given resolution.
|
49 |
+
- In practice, this only affected a very small number of very tall or very wide images so I don't think it would have made much of a difference
|
50 |
+
- Trained without VAE.
|
mari/lora_character_mari_v1_253i8r_832_batch3_5e-5text_2e-4unet_4epoch.json
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/mari",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "832,832",
|
10 |
+
"learning_rate": "",
|
11 |
+
"lr_scheduler": "cosine_with_restarts",
|
12 |
+
"lr_warmup": "5",
|
13 |
+
"train_batch_size": 3,
|
14 |
+
"epoch": "6",
|
15 |
+
"save_every_n_epochs": "2",
|
16 |
+
"mixed_precision": "fp16",
|
17 |
+
"save_precision": "fp16",
|
18 |
+
"seed": "31337",
|
19 |
+
"num_cpu_threads_per_process": 32,
|
20 |
+
"cache_latents": true,
|
21 |
+
"caption_extension": "",
|
22 |
+
"enable_bucket": true,
|
23 |
+
"gradient_checkpointing": false,
|
24 |
+
"full_fp16": false,
|
25 |
+
"no_token_padding": false,
|
26 |
+
"stop_text_encoder_training": 0,
|
27 |
+
"use_8bit_adam": true,
|
28 |
+
"xformers": true,
|
29 |
+
"save_model_as": "safetensors",
|
30 |
+
"shuffle_caption": true,
|
31 |
+
"save_state": false,
|
32 |
+
"resume": "",
|
33 |
+
"prior_loss_weight": 1.0,
|
34 |
+
"text_encoder_lr": "5e-5",
|
35 |
+
"unet_lr": "2e-4",
|
36 |
+
"network_dim": 128,
|
37 |
+
"lora_network_weights": "",
|
38 |
+
"color_aug": false,
|
39 |
+
"flip_aug": false,
|
40 |
+
"clip_skip": 2,
|
41 |
+
"gradient_accumulation_steps": 1.0,
|
42 |
+
"mem_eff_attn": false,
|
43 |
+
"output_name": "mari-v3-NoVAE",
|
44 |
+
"model_list": "",
|
45 |
+
"max_token_length": "150",
|
46 |
+
"max_train_epochs": "",
|
47 |
+
"max_data_loader_n_workers": ""
|
48 |
+
}
|
mari/lora_character_mari_v2_253i6r_768-fixed-max-bucket-reso_batch4_5e-5text_5e-4unet_4epoch.json
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/mari",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "768,768",
|
10 |
+
"learning_rate": "",
|
11 |
+
"lr_scheduler": "cosine_with_restarts",
|
12 |
+
"lr_warmup": "5",
|
13 |
+
"train_batch_size": 4,
|
14 |
+
"epoch": "5",
|
15 |
+
"save_every_n_epochs": "1",
|
16 |
+
"mixed_precision": "fp16",
|
17 |
+
"save_precision": "fp16",
|
18 |
+
"seed": "31337",
|
19 |
+
"num_cpu_threads_per_process": 32,
|
20 |
+
"cache_latents": true,
|
21 |
+
"caption_extension": "",
|
22 |
+
"enable_bucket": true,
|
23 |
+
"gradient_checkpointing": false,
|
24 |
+
"full_fp16": false,
|
25 |
+
"no_token_padding": false,
|
26 |
+
"stop_text_encoder_training": 0,
|
27 |
+
"use_8bit_adam": true,
|
28 |
+
"xformers": true,
|
29 |
+
"save_model_as": "safetensors",
|
30 |
+
"shuffle_caption": true,
|
31 |
+
"save_state": false,
|
32 |
+
"resume": "",
|
33 |
+
"prior_loss_weight": 1.0,
|
34 |
+
"text_encoder_lr": "5e-5",
|
35 |
+
"unet_lr": "5e-4",
|
36 |
+
"network_dim": 128,
|
37 |
+
"lora_network_weights": "",
|
38 |
+
"color_aug": false,
|
39 |
+
"flip_aug": false,
|
40 |
+
"clip_skip": 2,
|
41 |
+
"gradient_accumulation_steps": 1.0,
|
42 |
+
"mem_eff_attn": false,
|
43 |
+
"output_name": "mari-v2-NoVAE",
|
44 |
+
"model_list": "",
|
45 |
+
"max_token_length": "150",
|
46 |
+
"max_train_epochs": "",
|
47 |
+
"max_data_loader_n_workers": ""
|
48 |
+
}
|
mari/lora_character_mari_v3_253i8r_832_batch3_5e-5text_3e-4unet_4epoch.json
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/mari",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "832,832",
|
10 |
+
"learning_rate": "",
|
11 |
+
"lr_scheduler": "cosine_with_restarts",
|
12 |
+
"lr_warmup": "5",
|
13 |
+
"train_batch_size": 3,
|
14 |
+
"epoch": "4",
|
15 |
+
"save_every_n_epochs": "3",
|
16 |
+
"mixed_precision": "fp16",
|
17 |
+
"save_precision": "fp16",
|
18 |
+
"seed": "23",
|
19 |
+
"num_cpu_threads_per_process": 32,
|
20 |
+
"cache_latents": true,
|
21 |
+
"caption_extension": "",
|
22 |
+
"enable_bucket": true,
|
23 |
+
"gradient_checkpointing": false,
|
24 |
+
"full_fp16": false,
|
25 |
+
"no_token_padding": false,
|
26 |
+
"stop_text_encoder_training": 0,
|
27 |
+
"use_8bit_adam": true,
|
28 |
+
"xformers": true,
|
29 |
+
"save_model_as": "safetensors",
|
30 |
+
"shuffle_caption": true,
|
31 |
+
"save_state": false,
|
32 |
+
"resume": "",
|
33 |
+
"prior_loss_weight": 1.0,
|
34 |
+
"text_encoder_lr": "5e-5",
|
35 |
+
"unet_lr": "3e-4",
|
36 |
+
"network_dim": 128,
|
37 |
+
"lora_network_weights": "",
|
38 |
+
"color_aug": false,
|
39 |
+
"flip_aug": false,
|
40 |
+
"clip_skip": 2,
|
41 |
+
"gradient_accumulation_steps": 1.0,
|
42 |
+
"mem_eff_attn": false,
|
43 |
+
"output_name": "mari-v3-NoVAE",
|
44 |
+
"model_list": "",
|
45 |
+
"max_token_length": "150",
|
46 |
+
"max_train_epochs": "",
|
47 |
+
"max_data_loader_n_workers": ""
|
48 |
+
}
|
mari/lora_character_mari_v4_253i6r_832_batch3_5e-5text_2e-4unet_6epoch.json
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/mari",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "832,832",
|
10 |
+
"learning_rate": "",
|
11 |
+
"lr_scheduler": "cosine_with_restarts",
|
12 |
+
"lr_warmup": "5",
|
13 |
+
"train_batch_size": 3,
|
14 |
+
"epoch": "6",
|
15 |
+
"save_every_n_epochs": "2",
|
16 |
+
"mixed_precision": "fp16",
|
17 |
+
"save_precision": "fp16",
|
18 |
+
"seed": "31337",
|
19 |
+
"num_cpu_threads_per_process": 32,
|
20 |
+
"cache_latents": true,
|
21 |
+
"caption_extension": "",
|
22 |
+
"enable_bucket": true,
|
23 |
+
"gradient_checkpointing": false,
|
24 |
+
"full_fp16": false,
|
25 |
+
"no_token_padding": false,
|
26 |
+
"stop_text_encoder_training": 0,
|
27 |
+
"use_8bit_adam": true,
|
28 |
+
"xformers": true,
|
29 |
+
"save_model_as": "safetensors",
|
30 |
+
"shuffle_caption": true,
|
31 |
+
"save_state": false,
|
32 |
+
"resume": "",
|
33 |
+
"prior_loss_weight": 1.0,
|
34 |
+
"text_encoder_lr": "5e-5",
|
35 |
+
"unet_lr": "2e-4",
|
36 |
+
"network_dim": 128,
|
37 |
+
"lora_network_weights": "",
|
38 |
+
"color_aug": false,
|
39 |
+
"flip_aug": false,
|
40 |
+
"clip_skip": 2,
|
41 |
+
"gradient_accumulation_steps": 1.0,
|
42 |
+
"mem_eff_attn": false,
|
43 |
+
"output_name": "mari-v4-NoVAE",
|
44 |
+
"model_list": "",
|
45 |
+
"max_token_length": "150",
|
46 |
+
"max_train_epochs": "",
|
47 |
+
"max_data_loader_n_workers": ""
|
48 |
+
}
|
mari/lora_character_mari_v4_253i6r_832_batch3_5e-5text_2e-4unet_7epoch.json
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "G:/sd/repo/models/Stable-diffusion/nai-animefull-final-pruned.safetensors",
|
3 |
+
"v2": false,
|
4 |
+
"v_parameterization": false,
|
5 |
+
"logging_dir": "",
|
6 |
+
"train_data_dir": "G:/sd/training/datasets/mari",
|
7 |
+
"reg_data_dir": "G:/sd/training/datasets/regempty",
|
8 |
+
"output_dir": "G:/sd/repo/extensions/sd-webui-additional-networks/models/lora",
|
9 |
+
"max_resolution": "832,832",
|
10 |
+
"learning_rate": "",
|
11 |
+
"lr_scheduler": "cosine_with_restarts",
|
12 |
+
"lr_warmup": "5",
|
13 |
+
"train_batch_size": 3,
|
14 |
+
"epoch": "7",
|
15 |
+
"save_every_n_epochs": "2",
|
16 |
+
"mixed_precision": "fp16",
|
17 |
+
"save_precision": "fp16",
|
18 |
+
"seed": "31337",
|
19 |
+
"num_cpu_threads_per_process": 32,
|
20 |
+
"cache_latents": true,
|
21 |
+
"caption_extension": "",
|
22 |
+
"enable_bucket": true,
|
23 |
+
"gradient_checkpointing": false,
|
24 |
+
"full_fp16": false,
|
25 |
+
"no_token_padding": false,
|
26 |
+
"stop_text_encoder_training": 0,
|
27 |
+
"use_8bit_adam": true,
|
28 |
+
"xformers": true,
|
29 |
+
"save_model_as": "safetensors",
|
30 |
+
"shuffle_caption": true,
|
31 |
+
"save_state": false,
|
32 |
+
"resume": "",
|
33 |
+
"prior_loss_weight": 1.0,
|
34 |
+
"text_encoder_lr": "5e-5",
|
35 |
+
"unet_lr": "2e-4",
|
36 |
+
"network_dim": 128,
|
37 |
+
"lora_network_weights": "",
|
38 |
+
"color_aug": false,
|
39 |
+
"flip_aug": false,
|
40 |
+
"clip_skip": 2,
|
41 |
+
"gradient_accumulation_steps": 1.0,
|
42 |
+
"mem_eff_attn": false,
|
43 |
+
"output_name": "mari-v4-NoVAE",
|
44 |
+
"model_list": "",
|
45 |
+
"max_token_length": "150",
|
46 |
+
"max_train_epochs": "",
|
47 |
+
"max_data_loader_n_workers": ""
|
48 |
+
}
|