Text-to-Image
Diffusers
Safetensors
StableDiffusionPipeline
stable-diffusion
Inference Endpoints

[Tutorial] How To Do Stable Diffusion LORA Training By Using Web UI On Different Models - Tested SD 1.5, SD 2.1

#19
by MonsterMMORPG - opened

I hope this video gets added to the FAQ, wiki and stickies.

Appreciate very much.

Full Stable Diffusion related playlist that explains both on google colab and PC by Automatic1111 Web UI :
➡️ https://www.youtube.com/playlist?list=PL_pbwdIyffsmclLl0O144nQRnezKlNdx3

🎦 https://youtu.be/mfaqqL5yOO4

image.png

content of the video

0:00 Introduction speech
1:07 How to install the LoRA extension to the Stable Diffusion Web UI
2:36 Preparation of training set images by properly sized cropping
2:54 How to crop images using Paint .NET, an open-source image editing software
5:02 What is Low-Rank Adaptation (LoRA)
5:35 Starting preparation for training using the DreamBooth tab - LoRA
6:50 Explanation of all training parameters, settings, and options
8:27 How many training steps equal one epoch
9:09 Save checkpoints frequency
9:48 Save a preview of training images after certain steps or epochs
10:04 What is batch size in training settings
11:56 Where to set LoRA training in SD Web UI
13:45 Explanation of Concepts tab in training section of SD Web UI
14:00 How to set the path for training images
14:28 Classification Dataset Directory
15:22 Training prompt - how to set what to teach the model
15:55 What is Class and Sample Image Prompt in SD training
17:57 What is Image Generation settings and why we need classification image generation in SD training
19:40 Starting the training process
21:03 How and why to tune your Class Prompt (generating generic training images)
22:39 Why we generate regularization generic images by class prompt
23:27 Recap of the setting up process for training parameters, options, and settings
29:23 How much GPU, CPU, and RAM the class regularization image generation uses
29:57 Training process starts after class image generation has been completed
30:04 Displaying the generated class regularization images folder for SD 2.1
30:31 The speed of the training process - how many seconds per iteration on an RTX 3060 GPU
31:19 Where LoRA training checkpoints (weights) are saved
32:36 Where training preview images are saved and our first training preview image
33:10 When we will decide to stop training
34:09 How to resume training after training has crashed or you close it down
36:49 Lifetime vs. session training steps
37:54 After 30 epochs, resembling images start to appear in the preview folder
38:19 The command line printed messages are incorrect in some cases
39:05 Training step speed, a certain number of seconds per iteration (IT)
39:25 Results after 5600 steps (350 epochs) - it was sufficient for SD 2.1
39:44 How I'm picking a checkpoint to generate a full model .ckpt file
40:23 How to generate a full model .ckpt file from a LoRA checkpoint .pt file
41:17 Generated/saved file name is incorrect, but it is generated from the correct selected .pt file
42:01 Doing inference (generating new images) using the text2img tab with our newly trained and generated model
42:47 The results of SD 2.1 Version 768 pixel model after training with the LoRA method and teaching a human face
44:38 Setting up the training parameters/options for SD version 1.5 this time
48:35 Re-generating class regularization images since SD 1.5 uses 512 pixel resolution
49:11 Displaying the generated class regularization images folder for SD 1.5
50:16 Training of Stable Diffusion 1.5 using the LoRA methodology and teaching a face has been completed and the results are displayed
51:09 The inference (text2img) results with SD 1.5 training
51:19 You have to do more inference with LoRA since it has less precision than DreamBooth
51:39 How to give more attention/emphasis to certain keywords in the SD Web UI
52:51 How to generate more than 100 images using the script section of the Web UI
54:46 How to check PNG info to see used prompts and settings
55:24 How to upscale using AI models
56:12 Fixing face image quality, especially eyes, with GFPGAN visibility
56:32 How to batch post-process
57:00 Where batch-generated images are saved
57:18 Conclusion and ending speech

MonsterMMORPG changed discussion title from How To Do Stable Diffusion LORA Training By Using Web UI On Different Models - Tested SD 1.5, SD 2.1 to [Tutorial] How To Do Stable Diffusion LORA Training By Using Web UI On Different Models - Tested SD 1.5, SD 2.1
MonsterMMORPG changed discussion status to closed
deleted
This comment has been hidden

Nice Tutorial

thank you so much for comment and feedback

Where can I find a google colab with Lora training ?

Where can I find a google colab with Lora training ?

Unfortunately I dont have tutorial for that

Why don't you do dreambooth on Colab?

Here my 2 excellent tutorials

1
https://youtu.be/mnCY8uM7E50

Then 2
https://youtu.be/kIyqAdd_i10

playlist of tutorials for those who wonders : https://www.youtube.com/watch?v=mnCY8uM7E50&list=PL_pbwdIyffsmclLl0O144nQRnezKlNdx3

Would love to see a text writeup of this. ADHD and videos don't mesh well lol

Would love to see a text writeup of this. ADHD and videos don't mesh well lol

necroresurrecting the post because i am ADHD too and here you have the transcription done with happy scribe (i had some hours left so dw)

https://www.happyscribe.com/transcriptions/e263fe1aede940c4b7f2dbe12f12d73c/edit_v2

Would love to see a text writeup of this. ADHD and videos don't mesh well lol

necroresurrecting the post because i am ADHD too and here you have the transcription done with happy scribe (i had some hours left so dw)

https://www.happyscribe.com/transcriptions/e263fe1aede940c4b7f2dbe12f12d73c/edit_v2

I already have manually fixed subtitle you can use it :d

Sign up or log in to comment