filmov
tv
Training Midjourney Level Style And Yourself Into The SD 1.5 Model via DreamBooth Stable Diffusion

Показать описание
Playlist of #StableDiffusion Tutorials, Automatic1111 and Google Colab Guides, #DreamBooth, Textual Inversion / Embedding, LoRA, AI Upscaling, Pix2Pix, Img2Img:
Used training dataset in the video:
Style Trained model .safetensors (not includes myself - based on SD 1.5 pruned ckpt):
2400 Photo Of Man classification images:
0:00 Midjourney level style for free
0:35 Most important part of teaching a style into Stable Diffusion
1:23 I trained myself into the model and got amazing styled myself
1:30 First part of animation generation just like Corridor Crew did in their anime video
1:37 Used DreamBooth extension of Web UI on RunPod for training
1:57 Why training dataset is 1024x1024 pixels
2:25 Which rare token and class token are chosen why
2:54 Which dataset I have used to train myself
3:05 What kind of training dataset you need to generate consistent animation like Corridor Crew
3:27 A better way to connect your RunPod web UI instance
3:43 Which DreamBooth settings I have used to train myself into the base model
4:10 A good explanation of max resolution settings
4:31 Advanced tab settings of DreamBooth extension
5:15 Concepts tab of DreamBooth training
5:35 FileWords - image captions explanation
7:31 Where to see source checkpoint used for training
7:49 Why do seperate training instead of multiple concepts training
8:08 Style training used settings
9:20 Analysis of after style training upon myself trained model
10:25 x/y/z plot testing for Brad Pitt face to see overtraining effect
11:03 Castle in a forest test to verify not overtrained 1 more time
11:32 I had to do another training of my face
12:03 Final x/y/z plot comparison to decide best checkpoint
13:05 Analysis of final x/y/z plot
14:48 What you can do by using this methodology I explained
15:05 How to generate good quality, good face distant shots
15:10 very important parts of selecting good face training dataset
15:25 Why Stable Diffusion can't produce good face distant shots
15:33 How to do inpainting to fix your face in distant shots
15:48 What settings to use for inpainting to fix faces
16:17 How to upscale your image
16:30 GFPGAN to further improve face
The Midjourney level style provides an excellent starting point for creators looking to develop unique AI-generated animations. The most crucial aspect of teaching a style into Stable Diffusion is ensuring a comprehensive training dataset. To achieve this, the creator utilized a dataset of 1024x1024 pixels, which offers sufficient resolution for high-quality animation generation.
The choice of dataset is critical for generating consistent animations like Corridor Crew. To ensure success, the creator selected a rare token and class token that best suited their needs. The training dataset should be carefully curated to include a diverse range of images and styles to generate the desired outcome.
DreamBooth Extension Settings and Features
The DreamBooth extension offers a variety of settings to optimize the training process. The creator used specific settings in the max resolution and advanced tab settings to fine-tune the training process. Furthermore, they used the concepts tab and FileWords feature to add image captions and enhance the quality of the output.
Training and Analysis
Separate training for different concepts is recommended to achieve the best results. After style training, the creator analyzed the model and performed x/y/z plot testing for Brad Pitt's face to detect any overtraining effects. They also conducted a castle in a forest test to further verify that the model was not overtrained.
Improving Quality and Fixing Issues
One challenge faced in AI-generated animation is producing high-quality, distant face shots. Stable Diffusion may struggle with these shots, but inpainting can be employed to fix faces in distant shots. The creator used specific settings for inpainting and utilized GFPGAN to upscale and further improve facial images.
Conclusion
The methodology outlined in this article offers a comprehensive approach to mastering Midjourney level style and Stable Diffusion for AI-generated animation. By carefully selecting the training dataset, optimizing settings in DreamBooth, and employing advanced techniques such as inpainting and GFPGAN, creators can generate high-quality animations and images that captivate audiences.
Комментарии