josdirksen commited on
Commit
989d864
1 Parent(s): d9a66a3

Scheduled Commit

Browse files
Files changed (1) hide show
  1. ai-toolkit.log +147 -0
ai-toolkit.log CHANGED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/7 [00:00<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
1
  0%| | 0/7 [00:00<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
2
  0%| | 0/7 [00:00<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
1
+ The cache for model files in Transformers v4.22.0 has been updated. Migrating your old cache. This is a one-time only operation. You can interrupt this and resume the migration later on by calling `transformers.utils.move_cache()`.
2
+ Running 1 job
3
+
4
+ /usr/local/lib/python3.10/dist-packages/albumentations/__init__.py:13: UserWarning: A new version of Albumentations is available: 1.4.16 (you have 1.4.15). Upgrade using: pip install -U albumentations. To disable automatic update checks, set the environment variable NO_ALBUMENTATIONS_UPDATE to 1.
5
+ check_for_updates()
6
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/mediapipe_face/mediapipe_face_common.py:7: UserWarning: The module 'mediapipe' is not installed. The package will have limited functionality. Please install it using the command: pip install 'mediapipe'
7
+ warnings.warn(
8
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_5m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_5m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
9
+ return register_model(fn_wrapper)
10
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_11m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_11m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
11
+ return register_model(fn_wrapper)
12
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_224 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_224. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
13
+ return register_model(fn_wrapper)
14
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_384 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_384. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
15
+ return register_model(fn_wrapper)
16
+ /usr/local/lib/python3.10/dist-packages/controlnet_aux/segment_anything/modeling/tiny_vit_sam.py:654: UserWarning: Overwriting tiny_vit_21m_512 in registry with controlnet_aux.segment_anything.modeling.tiny_vit_sam.tiny_vit_21m_512. This is because the name being registered conflicts with an existing name. Please check if this is not expected.
17
+ return register_model(fn_wrapper)
18
+ You set `add_prefix_space`. The tokenizer needs to be converted from the slow tokenizers
19
+ {
20
+ "type": "sd_trainer",
21
+ "training_folder": "output",
22
+ "device": "cuda:0",
23
+ "network": {
24
+ "type": "lora",
25
+ "linear": 16,
26
+ "linear_alpha": 16
27
+ },
28
+ "save": {
29
+ "dtype": "float16",
30
+ "save_every": 500,
31
+ "max_step_saves_to_keep": 4,
32
+ "push_to_hub": false
33
+ },
34
+ "datasets": [
35
+ {
36
+ "folder_path": "/workspace/ai-toolkit/images",
37
+ "caption_ext": "txt",
38
+ "caption_dropout_rate": 0.05,
39
+ "shuffle_tokens": false,
40
+ "cache_latents_to_disk": true,
41
+ "resolution": [
42
+ 512,
43
+ 768,
44
+ 1024
45
+ ]
46
+ }
47
+ ],
48
+ "train": {
49
+ "batch_size": 1,
50
+ "steps": 1000,
51
+ "gradient_accumulation_steps": 1,
52
+ "train_unet": true,
53
+ "train_text_encoder": false,
54
+ "gradient_checkpointing": true,
55
+ "noise_scheduler": "flowmatch",
56
+ "optimizer": "adamw8bit",
57
+ "lr": 0.0004,
58
+ "ema_config": {
59
+ "use_ema": true,
60
+ "ema_decay": 0.99
61
+ },
62
+ "dtype": "bf16"
63
+ },
64
+ "model": {
65
+ "name_or_path": "black-forest-labs/FLUX.1-dev",
66
+ "is_flux": true,
67
+ "quantize": true
68
+ },
69
+ "sample": {
70
+ "sampler": "flowmatch",
71
+ "sample_every": 500,
72
+ "width": 1024,
73
+ "height": 1024,
74
+ "prompts": [
75
+ "Photo of elinelora holding a sign that says 'I LOVE PROMPTS!'",
76
+ "Professional headshot of elinelora in a business suit.",
77
+ "A happy pilot elinelora of a Boeing 747.",
78
+ "A doctor elinelora talking to a patient.",
79
+ "A chef elinelora in the middle of a bustling kitchen, plating a beautifully arranged dish.",
80
+ "A young elinelora with a big grin, holding a large ice cream cone in front of an old-fashioned ice cream parlor.",
81
+ "A person elinelora in a tuxedo, looking directly into the camera with a confident smile, standing on a red carpet at a gala event.",
82
+ "Person elinelora with a bitchin' 80's mullet hairstyle leaning out the window of a pontiac firebird"
83
+ ],
84
+ "neg": "",
85
+ "seed": 42,
86
+ "walk_seed": true,
87
+ "guidance_scale": 4,
88
+ "sample_steps": 20
89
+ },
90
+ "trigger_word": "elinelora"
91
+ }
92
+ Using EMA
93
+
94
+ #############################################
95
+ # Running job: my_first_flux_lora_v1
96
+ #############################################
97
+
98
+
99
+ Running 1 process
100
+ Loading Flux model
101
+ Loading transformer
102
+ Quantizing transformer
103
+ Loading vae
104
+ Loading t5
105
+
106
+
107
+ Quantizing T5
108
+ Loading clip
109
+ making pipe
110
+ preparing
111
+ create LoRA network. base dim (rank): 16, alpha: 16
112
+ neuron dropout: p=None, rank dropout: p=None, module dropout: p=None
113
+ create LoRA for Text Encoder: 0 modules.
114
+ create LoRA for U-Net: 494 modules.
115
+ enable LoRA for U-Net
116
+ Dataset: /workspace/ai-toolkit/images
117
+ - Preprocessing image dimensions
118
+
119
  0%| | 0/7 [00:00<?, ?it/s]
120
+ - Found 7 images
121
+ Bucket sizes for /workspace/ai-toolkit/images:
122
+ 448x576: 7 files
123
+ 1 buckets made
124
+ Caching latents for /workspace/ai-toolkit/images
125
+ - Saving latents to disk
126
+
127
+ Dataset: /workspace/ai-toolkit/images
128
+ - Preprocessing image dimensions
129
+
130
  0%| | 0/7 [00:00<?, ?it/s]
131
+ - Found 7 images
132
+ Bucket sizes for /workspace/ai-toolkit/images:
133
+ 640x832: 7 files
134
+ 1 buckets made
135
+ Caching latents for /workspace/ai-toolkit/images
136
+ - Saving latents to disk
137
+
138
+ Dataset: /workspace/ai-toolkit/images
139
+ - Preprocessing image dimensions
140
+
141
  0%| | 0/7 [00:00<?, ?it/s]
142
+ - Found 7 images
143
+ Bucket sizes for /workspace/ai-toolkit/images:
144
+ 832x1152: 7 files
145
+ 1 buckets made
146
+ Caching latents for /workspace/ai-toolkit/images
147
+ - Saving latents to disk
148
+
149
+ Generating baseline samples before training
150
+