This new Outpainting Technique is INSANE - ControlNet 1.1.
TLDRIn this video, the creator introduces an advanced outpainting technique using ControlNet 1.1, ideal for generating hyper-realistic images. The video walks viewers through installing ControlNet, configuring settings, and using specific models like inpant plus Llama for enhanced results. Key steps include setting up image resolutions, denoising strength, and working with Stable Diffusion models. The creator also compares techniques like 'inpaint only' versus 'inpaint plus Llama,' showing how the latter produces cleaner, more realistic outcomes. Finally, tips on scaling images and fixing common errors are provided.
Takeaways
- 🎨 Outpainting is a technique for extending images, and ControlNet 1.1 is key to achieving high-quality results.
- ⚙️ Quick installation: Use the 'Extensions' tab in Stable Diffusion, search for ControlNet, install, and restart Automatic 1111 for proper setup.
- 📂 Download the appropriate .pth and .yaml model files and place them in the ControlNet folder for successful integration.
- 🏰 Start with an image featuring a distinct subject, like a person or landmark, to make outpainting more effective.
- 🤖 For optimal results, use 'inpaint only + llama' in ControlNet for a cleaner and more refined output.
- 🖼️ Adjust image resolution to 1024x768 and set a high denoising strength (0.75-1) to introduce substantial changes during outpainting.
- 🔧 Resize and fill options in ControlNet provide better control over image composition compared to other methods.
- 🔍 After initial outpainting, you can upscale images using a simple workflow by setting the 'denoising strength' to 0.2 and using the DPM++ 2M Karras sampler.
- 📈 ControlNet allows outpainting on images not originally created in Stable Diffusion, giving more flexibility for expanding existing visuals.
- 💡 Experiment with different models and settings, such as 'inpaint only' vs. 'inpaint + llama,' to find the style that best suits your preferences.
Q & A
What is the focus of the video?
-The video focuses on a new outpainting technique using ControlNet 1.1, which the creator believes is the best for outpainting in AI art.
What are the initial steps to install ControlNet?
-First, click on Extensions in the interface, then click Available, Load from, search for ControlNet, click Install, and then click Installed and Apply. After restarting, download the models (both the .pth and .yaml files) and place them in the stable diffusion folder under models/controlnet.
Which model is recommended for outpainting in the video?
-The video recommends using the 'several realistic 3.1' model, which is known for creating hyper-realistic art.
What is the technique discussed for outpainting?
-The video suggests using a combined technique called 'inpaint only plus llama,' which merges two models, providing cleaner and more concise results for outpainting.
What are the recommended settings for the best results in outpainting?
-For the image resolution, the recommended settings are 1024x768. The denoising strength should be set high, between 0.75 and 1, to introduce significant changes. The control net should be enabled, and the same image should be loaded into image-to-image.
How does the creator improve outpainting results compared to using the 'Poor Man's outpainting' script?
-The creator suggests avoiding the 'Poor Man's outpainting' script because its results are mediocre. Instead, using the combined 'inpaint only plus llama' technique yields better, cleaner outcomes.
What is the impact of the ‘resize and fill’ option in ControlNet?
-The 'resize and fill' option ensures that the image is filled out evenly without cropping or distorting it. This setting is preferred over just resizing or cropping and resizing, as these can negatively affect the image's appearance.
What changes are made for upscaling in the workflow?
-For upscaling, ControlNet is disabled, the denoising strength is set to 0.2, and the sampler is switched to DPM++ 2M Karras. The resize factor is set to 2.
Can this outpainting technique be used on images not created in Stable Diffusion?
-Yes, the outpainting technique can be applied to images not generated in Stable Diffusion. You just need to load the image from your drive and specify its resolution to properly adjust it for outpainting.
What is the main advantage of using the 'inpaint plus llama' method over 'inpaint only'?
-The 'inpaint plus llama' method provides cleaner, more realistic results compared to 'inpaint only,' although users are encouraged to experiment with both to see which they prefer.
Outlines
🎨 Introduction to Outpainting Techniques and Control Net Setup
This paragraph introduces the concept of outpainting, a technique for expanding images using AI. The creator explains that they will demonstrate their preferred technique, which involves using Control Net. The installation process for Control Net is outlined in a quick 40-second tutorial. The steps include downloading the extension, installing it, and applying the restart. After installation, users are directed to download specific model files, the .pth and .yaml, and place them in the appropriate folder. The creator wraps up the introduction by stating that they will be using the 'Realistic 3.1' model for generating hyper-realistic art.
🖼️ Starting the Outpainting Process and Control Net Setup
The focus of this paragraph is on the steps required to start the outpainting process. It begins by recommending that users select an initial image, preferably one with a landmark or central focus. The creator advises using the 'Poor Man’s Outpainting' script as a starting point, but notes that it offers mediocre results. The paragraph continues with specific instructions on setting up Control Net, choosing the 'inpaint' option, and combining it with another AI model called Llama for better quality. The creator emphasizes the importance of selecting 'resize and fill' to avoid distorting the image, and setting resolution and denoising strength to optimize the outpainting effect.
🔧 Rendering and Fine-Tuning the Image
Here, the focus shifts to rendering the image with the outpainting technique. The creator suggests setting the resolution to 1024x768 and keeping the sampler unchanged for now. They also recommend setting the denoising strength between 0.75 and 1, with 0.9 being ideal to introduce significant changes. The Control Net must be enabled, and the same image used in 'image-to-image' should be loaded again. After rendering, the creator expresses satisfaction with the results. Additionally, there is a brief intermission in which the creator asks viewers to subscribe, explaining that it helps them gauge the value of their content and audience engagement.
📈 Upscaling and Further Adjustments
In this segment, the creator discusses the next steps for enhancing the image through upscaling. They reference a previous video and demonstrate a simple method by using the 'image-to-image' feature. Settings such as disabling Control Net and lowering the denoising strength to 0.2 are recommended. The sampler is set to 'DPM++ 2M Karras' for the best results. The creator then compares two different results from this method, commenting on their impressive quality. The paragraph ends with a teaser for the next part of the tutorial, which will cover using the outpainting technique on images not generated by Stable Diffusion.
🖼️ Outpainting with External Images
This section explains how to use outpainting on images that were not originally created within Stable Diffusion. The creator walks through the process of loading an external image using Control Net, selecting 'inpaint' and 'resize and refill' options. They stress the importance of knowing the resolution of the image being used to avoid distortions. The example used here is a 600x900 image, which is resized to 1024x900. The rendering process is showcased, with different results being presented using both 'inpaint only' and 'inpaint plus Llama'. The creator notes that the images rendered with the combined models appear cleaner and more realistic, while encouraging viewers to experiment with their own preferences.
🌕 Creating a Custom Scene with Inpainting
In this part, the creator demonstrates how to craft a specific scene using inpainting. A prompt is used to generate an image of a 'girl in a medieval city with a full moon and heavy rain,' which the creator praises as magnificent. They also address a mistake from a previous video, where they incorrectly stated that the whole image would be downsampled during inpainting. The correction clarifies that downsampling only occurs when not using a mask. By using a mask, users can retain the original resolution and increase details in specific areas. This correction ties back to their ongoing tutorial series, providing useful insights for viewers struggling with AI art workflows.
📺 Conclusion and Recommendations
The final paragraph wraps up the video by directing viewers to the creator's earlier basic workflow tutorial, which covers essential AI art techniques. The creator encourages anyone struggling with their AI art process to refer to that video for additional guidance and improved workflow strategies.
Mindmap
Keywords
💡Outpainting
💡ControlNet
💡Stable Diffusion
💡Inpainting
💡Prompt Engineering
💡LLaMA
💡Denoising Strength
💡Sampler
💡Resize and Fill
💡Image Resolution
Highlights
New outpainting technique using ControlNet 1.1 is considered by the creator to be the best for achieving high-quality results.
Quick 40-second installation guide for ControlNet: Extensions > Available > Load from > Search for ControlNet > Install > Restart Automatic 1111.
Reminder to download the correct pth and yaml files for ControlNet and place them in the stable diffusion models folder.
Using the model 'Severally Realistic 3.1' for creating hyper-realistic art as a strong recommendation from the creator.
First, prompt engineer an image with a landmark or central figure before starting the outpainting process.
The ‘Poor Man’s Outpainting’ script gives mediocre results, so it is advised not to use it.
Using 'Inpaint Only' combined with the 'Llama' model in ControlNet results in cleaner and more concise images.
Important settings: Choose ‘ControlNet is more important’ and ‘resize and fill’ to avoid issues with the image appearance.
For best results, set the resolution to 1024x768 and denoising strength between 0.75 and 1 (recommended 0.9).
After setting everything up in ControlNet, remember to reload the image in image-to-image for consistent rendering.
Demonstrates how to upscale images by resizing in image-to-image with a scale factor of 2 and modifying denoising strength.
Inpainting without using prompts also yields effective results; however, Inpaint + Llama produces cleaner, more realistic images.
Explains how to outpaint even with images not generated in Stable Diffusion by adjusting resolution settings in ControlNet.
ControlNet’s ‘mask only’ option allows for upscaling without downsampling the entire image, improving inpainted details.
Users are encouraged to experiment with different models and approaches, as preferences for Inpaint + Llama or Inpaint Only are subjective.