Zero to Hero Stable Diffusion 3 Tutorial with Amazing SwarmUI SD Web UI that Utilizes ComfyUI

preview_player
Показать описание
Do not skip any part of this tutorial to master how to use Stable Diffusion 3 (SD3) with the most advanced generative AI open source APP SwarmUI. Automatic1111 SD Web UI or Fooocus are not supporting the #SD3 yet. Therefore, I am starting to make tutorials for SwarmUI as well. #StableSwarmUI is officially developed by the StabilityAI and your mind will be blown after you watch this tutorial and learn its amazing features. StableSwarmUI uses #ComfyUI as the back end thus it has all the good features of ComfyUI and it brings you easy to use features of Automatic1111 #StableDiffusion Web UI with them. I really liked SwarmUI and planning to do more tutorials for it.


0:00 Introduction to the Stable Diffusion 3 (SD3) and SwarmUI and what is in the tutorial
4:12 Architecture and features of SD3
5:05 What each different model files of Stable Diffusion 3 means
6:26 How to download and install SwarmUI on Windows for SD3 and all other Stable Diffusion models
8:42 What kind of folder path you should use when installing SwarmUI
10:28 If you get installation error how to notice and fix it
11:49 Installation has been completed and now how to start using SwarmUI
12:29 Which settings I change before start using SwarmUI and how to change your theme like dark, white, gray
12:56 How to make SwarmUI save generated images as PNG
13:08 How to find description of each settings and configuration
13:28 How to download SD3 model and start using on Windows
13:38 How to use model downloader utility of SwarmUI
14:17 How to set models folder paths and link your existing models folders in SwarmUI
14:35 Explanation of Root folder path in SwarmUI
14:52 VAE of SD3 do we need to download?
15:25 Generate and model section of the SwarmUI to generate images and how to select your base model
16:02 Setting up parameters and what they do to generate images
17:06 Which sampling method is best for SD3
17:22 Information about SD3 text encoders and their comparison
18:14 First time generating an image with SD3
19:36 How to regenerate same image
20:17 How to see image generation speed and step speed and more information
20:29 Stable Diffusion 3 it per second speed on RTX 3090 TI
20:39 How to see VRAM usage on Windows 10
22:08 And testing and comparing different text encoders for SD3
22:36 How to use FP16 version of T5 XXL text encoder instead of default FP8 version
25:27 The image generation speed when using best config for SD3
26:37 Why VAE of the SD3 is many times better than previous Stable Diffusion models, 4 vs 8 vs 16 vs 32 channels VAE
27:40 How to and where to download best AI upscaler models
29:10 How to use refiner and upscaler models to improve and upscale generated images
29:21 How to restart and start SwarmUI
32:01 The folders where the generated images are saved
32:13 Image history feature of SwarmUI
33:10 Upscaled image comparison
34:01 How to download all upscaler models at once
34:34 Presets feature in depth
36:55 How to generate forever / infinite times
37:13 Non-tiled upscale caused issues
38:36 How to compare tiled vs non-tiled upscale and decide best
39:05 275 SwarmUI presets (cloned from Fooocus) I prepared and the scripts I coded to prepare them and how to import those presets
42:10 Model browser feature
43:25 How to generate TensorRT engine for huge speed up
43:47 How to update SwarmUI
44:27 Prompt syntax and advanced features
45:35 How to use Wildcards (random prompts) feature
46:47 How to see full details / metadata of generated images
47:13 Full guide for extremely powerful grid image generation (like X/Y/Z plot)
47:35 How to put all downloaded upscalers from zip file
51:37 How to see what is happening at the server logs
53:04 How to continue grid generation process after interruption
54:32 How to open grid generation after it has been completed and how to use it
56:13 Example of tiled upscaling seaming problem
1:00:30 Full guide for image history
1:02:22 How to directly delete images and star them
1:03:20 How to use SD 1.5 and SDXL models and LoRAs
1:06:24 Which sampler method is best
1:06:43 How to use image to image
1:08:43 How to use edit image / inpainting
1:10:38 How to use amazing segmentation feature to automatically inpaint any part of images
1:15:55 How to use segmentation on existing images for inpainting and get perfect results with different seeds
1:18:19 More detailed information regarding upscaling and tiling and SD3
1:20:08 Seams perfect explanation and example and how to fix it
1:21:09 How to use queue system
1:21:23 How to use multiple GPUs with adding more backends
1:24:38 Loading model in low VRAM mode
1:25:10 How to fix colors over saturation
1:27:00 Best image generation configuration for SD3
1:27:44 How to apply upscale to your older generated images quickly via preset
1:28:39 Other amazing features of SwarmUI
1:28:49 Clip tokenization and rare token OHWX
Рекомендации по теме
Комментарии
Автор

Video chapters and manually fixed captions added. RunPod instructions added to the above link. Our Massed Compute VM now also has pre-installed SwarmUI.

SECourses
Автор

I thank you very much for all your effort and time you put together with all this detailed explanation.

TomiTom
Автор

A bit overwhelming at first but way easier than Comfy on its own ^^
Always great Tutorials here :)
Let's hope that we can launch SD to the next level and that Open Ai won't strangulate itself and us with it :D
Cheers

frzenisshadowbanned
Автор

🔥🔥🔥 I know this must have taken you a long time to put together. This was extremely helpful. More short tutorials would be so helpful

marioa
Автор

Nice, but I think I need license just to watch this video 😂

pastuh
Автор

Hi I am a checkpoint creator. For now it's better to stick with SDXL. It's far better right now, it is supported better by UIs, its uncensored, You can use it commercial without paying 20 dollars a month. There are a lot of Loras (small style additions mostly) available. At the moment I don't would use SD3. It's maybe something for the future. Wait a couple of weeks and then look at it again.

Afrmanpeace
Автор

12GB+ VRAM usage in txt2img with no control nets enabled? Is this a common reference or does it just use more VRAM, if it is made available to it?

divye.ruhela
Автор

Very good video. It must have taken you so long to create this. Thank you.

gohan
Автор

You are always do amazing work thank you. Eid Mubarak 😘😘😘👍👍👍

ihsasss
Автор

I took the time and tested it today.
Sadly I am very underwhelmed with SWARM.
The Upscaling introduces these shimmering Color Spots, and without Upscaling the "Restore Face" Alternative (<segment:face, 0.6, 0.5>)produces visible Overlay residue... sooo it's not good :(
I never had luck with anything ComfyUi ^^ Missing Nodes - Incompatibilities - bad Updates... But at least *Forge* works great :)
Why can't they just add Restore Face to Comfy??? That would have helped with 50% of my Comfy Problems already.

frzenisshadowbanned
Автор

Haha 8:45 i see you have like myself about 7 different OS on partitions! I am currently in Windows 10 on a 2012 Mac Mini with a 16 gb ram core i7 cpu (8 cores) and weak nVidia 4000 graphics, because 2012 mac mini will not update past mac os Catalina, and apps like gpt4all are not usable or installable or supported, but on windows the hardware is sitll supported, so i partitioned 1 tb of my 2 tb disk ssd as ntfs and installed windows ten from my lenovo yoga (serial number), which is also unable to upgrade to windows 11, whereas the mac mini hardware seen through the eyes of windows, is still valid for 2024 usage. GPT4All runs a bit slow, but it runs. I would like to get this up and running tomorrow from your tutorial as i already have most of the dependencies installed, but will be careful to take it a step at a time and make sure a perfect install on first time happens (hopefully) although as they say 'breaking things is one of the best ways to learn how to fix them'

HermesFibonacci
Автор

waiting for a local fine tuning with my 3090!! I hope that 24gb will be enough and also I wonder if will ned to train the new powerfull text encoder to get the most of it

dlarva
Автор

Çok yayarlı bir eğitim oldu hocam teşekkür ederim. Bu arayüzde comfyui da olduğu gibi faceswap, renklendirme gibi extrem işlemleri gerçekleştirebilirmiyiz acaba. Galiba comfyui dan buna geçiş yapıcaz gibi.

cemilhaci
Автор

thanks very much for this you are great man!

nomorejustice
Автор

I cant make it use my image and work with that as input

slookify
Автор

can a macbook air do this? or do any of your tutorials? I feel trapped as a beginner

shawnho
Автор

As a complete noob, what's the difference between stable diffusion 3 and something like SDXL?

VencoBg
Автор

i’m going to buy a computer specifically for this could you please give me the most recommended specifications for a laptop or even a desktop PC for three to $4000? GPUCPU how many cores? I don’t need to know anything else just those three the CPU, the GPU, and how many cores. I’m thinking about 32 GB ram and a core nine processor with two of those 40 something or other series Nvidia gpus

GPTHackstation
Автор

The diffusers have updated train_dreambooth_lora_sd3. Can you release a new course for this? Thank you

squallzy
Автор

The question should not be HOW, but WHY heh.
But on a serious note, good tutorial.

azradun