josdirksen commited on
Commit
b8691aa
1 Parent(s): e7d29a9

Scheduled Commit

Browse files
Files changed (1) hide show
  1. ai-toolkit.log +142 -0
ai-toolkit.log CHANGED
@@ -1,3 +1,145 @@
1
  The cache for model files in Transformers v4.22.0 has been updated. Migrating your old cache. This is a one-time only operation. You can interrupt this and resume the migration later on by calling `transformers.utils.move_cache()`.
2
  Running 1 job
3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  0%| | 0/8 [00:00<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
5
  0%| | 0/8 [00:00<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
6
  0%| | 0/8 [00:00<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
1
  The cache for model files in Transformers v4.22.0 has been updated. Migrating your old cache. This is a one-time only operation. You can interrupt this and resume the migration later on by calling `transformers.utils.move_cache()`.
2
  Running 1 job
3
 
4
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/mediapipe_face/mediapipe_face_common.py:7: UserWarning: The module 'mediapipe' is not installed. The package will have limited functionality. Please install it using the command: pip install 'mediapipe'
5
+ warnings.warn(
6
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_5m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_5m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
7
+ return register_model(fn_wrapper)
8
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_11m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_11m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
9
+ return register_model(fn_wrapper)
10
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
11
+ return register_model(fn_wrapper)
12
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_384 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_384. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
13
+ return register_model(fn_wrapper)
14
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_512 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_512. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
15
+ return register_model(fn_wrapper)
16
+ You set `add_prefix_space`. The tokenizer needs to be converted from the slow tokenizers
17
+ {
18
+ "type": "sd_trainer",
19
+ "training_folder": "output",
20
+ "device": "cuda:0",
21
+ "network": {
22
+ "type": "lora",
23
+ "linear": 16,
24
+ "linear_alpha": 16
25
+ },
26
+ "save": {
27
+ "dtype": "float16",
28
+ "save_every": 500,
29
+ "max_step_saves_to_keep": 4,
30
+ "push_to_hub": false
31
+ },
32
+ "datasets": [
33
+ {
34
+ "folder_path": "/workspace/ai-toolkit/images",
35
+ "caption_ext": "txt",
36
+ "caption_dropout_rate": 0.05,
37
+ "shuffle_tokens": false,
38
+ "cache_latents_to_disk": true,
39
+ "resolution": [
40
+ 512,
41
+ 768,
42
+ 1024
43
+ ]
44
+ }
45
+ ],
46
+ "train": {
47
+ "batch_size": 1,
48
+ "steps": 1000,
49
+ "gradient_accumulation_steps": 1,
50
+ "train_unet": true,
51
+ "train_text_encoder": false,
52
+ "gradient_checkpointing": true,
53
+ "noise_scheduler": "flowmatch",
54
+ "optimizer": "adamw8bit",
55
+ "lr": 0.0004,
56
+ "ema_config": {
57
+ "use_ema": true,
58
+ "ema_decay": 0.99
59
+ },
60
+ "dtype": "bf16"
61
+ },
62
+ "model": {
63
+ "name_or_path": "black-forest-labs/FLUX.1-dev",
64
+ "is_flux": true,
65
+ "quantize": true
66
+ },
67
+ "sample": {
68
+ "sampler": "flowmatch",
69
+ "sample_every": 500,
70
+ "width": 1024,
71
+ "height": 1024,
72
+ "prompts": [
73
+ "Photo of joslora holding a sign that says 'I LOVE PROMPTS!'",
74
+ "Professional headshot of joslora in a business suit.",
75
+ "A happy pilot joslora of a Boeing 747.",
76
+ "A doctor joslora talking to a patient.",
77
+ "A chef joslora in the middle of a bustling kitchen, plating a beautifully arranged dish.",
78
+ "A young joslora with a big grin, holding a large ice cream cone in front of an old-fashioned ice cream parlor.",
79
+ "A person joslora in a tuxedo, looking directly into the camera with a confident smile, standing on a red carpet at a gala event.",
80
+ "Person joslora with a bitchin' 80's mullet hairstyle leaning out the window of a pontiac firebird"
81
+ ],
82
+ "neg": "",
83
+ "seed": 42,
84
+ "walk_seed": true,
85
+ "guidance_scale": 4,
86
+ "sample_steps": 20
87
+ },
88
+ "trigger_word": "joslora"
89
+ }
90
+ Using EMA
91
+
92
+ #############################################
93
+ # Running job: my_first_flux_lora_v1
94
+ #############################################
95
+
96
+
97
+ Running 1 process
98
+ Loading Flux model
99
+ Loading transformer
100
+ Quantizing transformer
101
+ Loading vae
102
+ Loading t5
103
+
104
+
105
+ Quantizing T5
106
+ Loading clip
107
+ making pipe
108
+ preparing
109
+ create LoRA network. base dim (rank): 16, alpha: 16
110
+ neuron dropout: p=None, rank dropout: p=None, module dropout: p=None
111
+ create LoRA for Text Encoder: 0 modules.
112
+ create LoRA for U-Net: 494 modules.
113
+ enable LoRA for U-Net
114
+ Dataset: /workspace/ai-toolkit/images
115
+ - Preprocessing image dimensions
116
+
117
  0%| | 0/8 [00:00<?, ?it/s]
118
+ - Found 8 images
119
+ Bucket sizes for /workspace/ai-toolkit/images:
120
+ 576x384: 8 files
121
+ 1 buckets made
122
+ Caching latents for /workspace/ai-toolkit/images
123
+ - Saving latents to disk
124
+
125
+ Dataset: /workspace/ai-toolkit/images
126
+ - Preprocessing image dimensions
127
+
128
  0%| | 0/8 [00:00<?, ?it/s]
129
+ - Found 8 images
130
+ Bucket sizes for /workspace/ai-toolkit/images:
131
+ 832x576: 8 files
132
+ 1 buckets made
133
+ Caching latents for /workspace/ai-toolkit/images
134
+ - Saving latents to disk
135
+
136
+ Dataset: /workspace/ai-toolkit/images
137
+ - Preprocessing image dimensions
138
+
139
  0%| | 0/8 [00:00<?, ?it/s]
140
+ - Found 8 images
141
+ Bucket sizes for /workspace/ai-toolkit/images:
142
+ 1216x832: 8 files
143
+ 1 buckets made
144
+ Caching latents for /workspace/ai-toolkit/images
145
+ - Saving latents to disk
146
+
147
+ Generating baseline samples before training
148
+