josdirksen
commited on
Commit
•
989d864
1
Parent(s):
d9a66a3
Scheduled Commit
Browse files- ai-toolkit.log +147 -0
ai-toolkit.log
CHANGED
@@ -0,0 +1,147 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
0 |
0%| | 0/7 [00:00<?, ?it/s]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
0%| | 0/7 [00:00<?, ?it/s]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
0%| | 0/7 [00:00<?, ?it/s]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
The cache for model files in Transformers v4.22.0 has been updated. Migrating your old cache. This is a one-time only operation. You can interrupt this and resume the migration later on by calling `transformers.utils.move_cache()`.
|
2 |
+
Running 1 job
|
3 |
+
|
4 |
+
/usr/local/lib/python3.10/dist-packages/albumentations/__init__.py:13: UserWarning: A new version of Albumentations is available: 1.4.16 (you have 1.4.15). Upgrade using: pip install -U albumentations. To disable automatic update checks, set the environment variable NO_ALBUMENTATIONS_UPDATE to 1.
|
5 |
+
check_for_updates()
|
6 |
+
/usr/local/lib/python3.10/dist-packages/controlnet_aux/mediapipe_face/mediapipe_face_common.py:7: UserWarning: The module 'mediapipe' is not installed. The package will have limited functionality. Please install it using the command: pip install 'mediapipe'
|
7 |
+
warnings.warn(
|
8 |
+
/usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_5m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_5m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
|
9 |
+
return register_model(fn_wrapper)
|
10 |
+
/usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_11m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_11m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
|
11 |
+
return register_model(fn_wrapper)
|
12 |
+
/usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
|
13 |
+
return register_model(fn_wrapper)
|
14 |
+
/usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_384 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_384. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
|
15 |
+
return register_model(fn_wrapper)
|
16 |
+
/usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_512 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_512. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
|
17 |
+
return register_model(fn_wrapper)
|
18 |
+
You set `add_prefix_space`. The tokenizer needs to be converted from the slow tokenizers
|
19 |
+
{
|
20 |
+
"type": "sd_trainer",
|
21 |
+
"training_folder": "output",
|
22 |
+
"device": "cuda:0",
|
23 |
+
"network": {
|
24 |
+
"type": "lora",
|
25 |
+
"linear": 16,
|
26 |
+
"linear_alpha": 16
|
27 |
+
},
|
28 |
+
"save": {
|
29 |
+
"dtype": "float16",
|
30 |
+
"save_every": 500,
|
31 |
+
"max_step_saves_to_keep": 4,
|
32 |
+
"push_to_hub": false
|
33 |
+
},
|
34 |
+
"datasets": [
|
35 |
+
{
|
36 |
+
"folder_path": "/workspace/ai-toolkit/images",
|
37 |
+
"caption_ext": "txt",
|
38 |
+
"caption_dropout_rate": 0.05,
|
39 |
+
"shuffle_tokens": false,
|
40 |
+
"cache_latents_to_disk": true,
|
41 |
+
"resolution": [
|
42 |
+
512,
|
43 |
+
768,
|
44 |
+
1024
|
45 |
+
]
|
46 |
+
}
|
47 |
+
],
|
48 |
+
"train": {
|
49 |
+
"batch_size": 1,
|
50 |
+
"steps": 1000,
|
51 |
+
"gradient_accumulation_steps": 1,
|
52 |
+
"train_unet": true,
|
53 |
+
"train_text_encoder": false,
|
54 |
+
"gradient_checkpointing": true,
|
55 |
+
"noise_scheduler": "flowmatch",
|
56 |
+
"optimizer": "adamw8bit",
|
57 |
+
"lr": 0.0004,
|
58 |
+
"ema_config": {
|
59 |
+
"use_ema": true,
|
60 |
+
"ema_decay": 0.99
|
61 |
+
},
|
62 |
+
"dtype": "bf16"
|
63 |
+
},
|
64 |
+
"model": {
|
65 |
+
"name_or_path": "black-forest-labs/FLUX.1-dev",
|
66 |
+
"is_flux": true,
|
67 |
+
"quantize": true
|
68 |
+
},
|
69 |
+
"sample": {
|
70 |
+
"sampler": "flowmatch",
|
71 |
+
"sample_every": 500,
|
72 |
+
"width": 1024,
|
73 |
+
"height": 1024,
|
74 |
+
"prompts": [
|
75 |
+
"Photo of elinelora holding a sign that says 'I LOVE PROMPTS!'",
|
76 |
+
"Professional headshot of elinelora in a business suit.",
|
77 |
+
"A happy pilot elinelora of a Boeing 747.",
|
78 |
+
"A doctor elinelora talking to a patient.",
|
79 |
+
"A chef elinelora in the middle of a bustling kitchen, plating a beautifully arranged dish.",
|
80 |
+
"A young elinelora with a big grin, holding a large ice cream cone in front of an old-fashioned ice cream parlor.",
|
81 |
+
"A person elinelora in a tuxedo, looking directly into the camera with a confident smile, standing on a red carpet at a gala event.",
|
82 |
+
"Person elinelora with a bitchin' 80's mullet hairstyle leaning out the window of a pontiac firebird"
|
83 |
+
],
|
84 |
+
"neg": "",
|
85 |
+
"seed": 42,
|
86 |
+
"walk_seed": true,
|
87 |
+
"guidance_scale": 4,
|
88 |
+
"sample_steps": 20
|
89 |
+
},
|
90 |
+
"trigger_word": "elinelora"
|
91 |
+
}
|
92 |
+
Using EMA
|
93 |
+
|
94 |
+
#############################################
|
95 |
+
# Running job: my_first_flux_lora_v1
|
96 |
+
#############################################
|
97 |
+
|
98 |
+
|
99 |
+
Running 1 process
|
100 |
+
Loading Flux model
|
101 |
+
Loading transformer
|
102 |
+
Quantizing transformer
|
103 |
+
Loading vae
|
104 |
+
Loading t5
|
105 |
+
|
106 |
+
|
107 |
+
Quantizing T5
|
108 |
+
Loading clip
|
109 |
+
making pipe
|
110 |
+
preparing
|
111 |
+
create LoRA network. base dim (rank): 16, alpha: 16
|
112 |
+
neuron dropout: p=None, rank dropout: p=None, module dropout: p=None
|
113 |
+
create LoRA for Text Encoder: 0 modules.
|
114 |
+
create LoRA for U-Net: 494 modules.
|
115 |
+
enable LoRA for U-Net
|
116 |
+
Dataset: /workspace/ai-toolkit/images
|
117 |
+
- Preprocessing image dimensions
|
118 |
+
|
119 |
0%| | 0/7 [00:00<?, ?it/s]
|
120 |
+
- Found 7 images
|
121 |
+
Bucket sizes for /workspace/ai-toolkit/images:
|
122 |
+
448x576: 7 files
|
123 |
+
1 buckets made
|
124 |
+
Caching latents for /workspace/ai-toolkit/images
|
125 |
+
- Saving latents to disk
|
126 |
+
|
127 |
+
Dataset: /workspace/ai-toolkit/images
|
128 |
+
- Preprocessing image dimensions
|
129 |
+
|
130 |
0%| | 0/7 [00:00<?, ?it/s]
|
131 |
+
- Found 7 images
|
132 |
+
Bucket sizes for /workspace/ai-toolkit/images:
|
133 |
+
640x832: 7 files
|
134 |
+
1 buckets made
|
135 |
+
Caching latents for /workspace/ai-toolkit/images
|
136 |
+
- Saving latents to disk
|
137 |
+
|
138 |
+
Dataset: /workspace/ai-toolkit/images
|
139 |
+
- Preprocessing image dimensions
|
140 |
+
|
141 |
0%| | 0/7 [00:00<?, ?it/s]
|
142 |
+
- Found 7 images
|
143 |
+
Bucket sizes for /workspace/ai-toolkit/images:
|
144 |
+
832x1152: 7 files
|
145 |
+
1 buckets made
|
146 |
+
Caching latents for /workspace/ai-toolkit/images
|
147 |
+
- Saving latents to disk
|
148 |
+
|
149 |
+
Generating baseline samples before training
|
150 |
+
|