9 May 202310:06

TLDRThis video tutorial delves into the process of refining illustrations generated by AI to alter clothing. It revisits the use of the image tab's paint function and introduces ControlNet for significant modifications. The method requires a stable diffusion web environment and ControlNet integration. The video guides through changing hair color without affecting other aspects and altering outfits using external tools like Paint. It highlights the versatility of stable diffusion web UI and ControlNet, enabling precise adjustments and generating customized images.


  • 🎨 The video discusses methods for modifying illustrations created by image generation AI.
  • 🖌️ The process involves using the 'Image' tab's 'Inpaint' feature to make partial changes to the image.
  • 🔄 The video introduces using ControlNet in conjunction with the 'Inpaint' function for significant alterations.
  • 🌐 The method requires a stable diffusion web environment and ControlNet integration.
  • 🔍 The video provides a link in the description for viewers interested in learning more about web3 and AI.
  • 🖼️ The demonstration begins by uploading an image to the 'Image' tab and using the 'Inpaint' tool to modify the paper color without changing the hairstyle.
  • 🎨 The ControlNet's 'Carry' function is highlighted, allowing for changing hair color while maintaining the original image's appearance.
  • 👗 The video also covers changing clothing in the image using external tools, specifically mentioning Windows Paint for removing the background.
  • 👚 An example is given where the AI adjusts the image to reflect a school uniform, despite minor issues arising from over-editing.
  • 📏 The video emphasizes the importance of adjusting parameters such as 'Inpainting Strength' and ControlNet's 'Control Weight' for optimal results.
  • 👗 The versatility of the method is showcased by changing the outfit to a maid uniform and then to a swimsuit.
  • 🔄 The video concludes by encouraging viewers to subscribe to the channel for more content on Stable Diffusion WEBUI and ControlNet.

Q & A

  • What is the main focus of the video?

    -The main focus of the video is explaining how to modify illustrations created by an AI image generator, specifically changing the clothes of the characters.

  • What method was mentioned for making partial changes to AI-generated illustrations?

    -The video mentions using the inpainting feature of the image tool tab for making partial changes to AI-generated illustrations.

  • What new approach does the video propose for modifying AI-generated images?

    -The new approach involves using both the inpainting feature and a tool called ControlNet to make significant modifications to AI-generated images.

  • What is necessary to implement the method described in the video?

    -Implementing the described method requires setting up a local environment for Stable Diffusion Web UI and installing ControlNet.

  • What does the video suggest for people who haven't built extensions?

    -For people who haven't built extensions, the video suggests referring to other videos for guidance.

  • How does the video suggest changing the color of an element without altering its shape?

    -The video suggests using the carry feature of ControlNet to change the color of an element, like hair, without altering its shape.

  • What workaround is provided for changing clothes in the illustration?

    -To change clothes, the video recommends using an external drawing tool or Paint, which comes standard with Windows, to erase the clothes before having the AI redraw them.

  • How does the video address potential errors in AI modifications?

    -The video suggests adjusting parameters such as 'Inpainting Strength' or 'ControlNet Weight' to correct any errors or unwanted changes made by the AI.

  • What types of content does the video's channel focus on?

    -The channel focuses on explanatory videos about Web 2.0, Web 3.0, and AI technologies.

  • What does the video say about the evolution of AI and its tools?

    -The video mentions that AI and its tools, like Stable Diffusion Web UI and ControlNet, are rapidly evolving, expanding the possibilities for image modification and creation.



🎨 Improving AI-Generated Illustrations with Clothes Modification

This video introduces techniques for modifying AI-generated illustrations, specifically focusing on changing outfits using the inpainting feature and ControlNet within the Stable Diffusion WebUI environment. It highlights the necessity of local setup for Stable Diffusion WebUI and ControlNet implementation. The video continues from previous tutorials on partial modifications using inpainting, aiming for significant changes this time. It guides viewers through uploading an image, using inpainting to change specific parts like hair color without altering the hairstyle, and emphasizes the limitations related to color prompt ranges primarily in English. The video also suggests referring to other tutorials for detailed parameters and installation guides, encouraging viewers to explore more about Web 2, Web 3, and AI on the channel.


👗 Changing Outfits on AI-Generated Characters

This segment walks viewers through the process of changing outfits on AI-generated characters using external tools like paint software and Stable Diffusion's WebUI along with ControlNet. It demonstrates how to remove unwanted lines to prepare for outfit changes, how to adjust parameters to compensate for over-erased parts, and how AI can automatically correct minor mistakes. The tutorial covers various outfits, including school uniforms and maid costumes, and acknowledges the fine line of appropriateness for YouTube content. It emphasizes the expanded capabilities and lowered barriers for creating high-quality images with Stable Diffusion WebUI and ControlNet, despite rapid updates that challenge content creation. The video concludes with an invitation to subscribe and check out additional resources for mastering prompt crafting and staying updated with AI advancements.


✨ A Glimpse into New Worlds

This brief closing segment metaphorically reflects on the endless possibilities and new worlds unveiled through the creative journey with AI and technology. It hints at the wonder and potential unlocked by exploring and mastering these tools, inviting viewers to anticipate and navigate these new horizons together.



💡Image Generation AI

Image Generation AI refers to artificial intelligence systems designed to create visual content, such as illustrations or images, based on input data or prompts. In the context of the video, it is used to generate initial illustrations that are then modified to change outfits and hairstyles, demonstrating the AI's capability to produce detailed and customizable visual outputs.

💡Image Tab and Paint Function

The Image Tab and Paint Function are features within an imaging tool that allows users to upload and edit images partially. The Image Tab is where the image is displayed and worked on, while the Paint Function provides brushes and other tools to manually edit specific areas of the image, such as changing the paper color without affecting the hairstyle.

💡Control Net

Control Net is a feature that enables users to make significant changes to an image while maintaining the overall structure and composition of the original image. It works by allowing users to upload a reference image and then applying adjustments based on that reference, ensuring that major elements like the subject's pose or facial structure remain consistent with the original.

💡Stable Diffusion WEBUI

Stable Diffusion WEBUI is a user interface for the Stable Diffusion model, which is an AI-based image generation and editing tool. The WEBUI provides a platform for users to interact with the Stable Diffusion model directly from a web browser, enabling them to generate and modify images without the need for extensive coding knowledge.

💡Local Environment

A local environment refers to a setup where software or applications are installed and run on a user's personal computer or device, as opposed to a remote or cloud-based server. In the context of the video, setting up a local environment for Stable Diffusion WEBUI and Control Net is necessary for the image editing process.

💡Image Upload

Image Upload is the process of transferring an image file from a local storage device to a software application or online platform. In the video, image upload is a crucial step to import the original image into the imaging tool for further editing and modification.

💡Parameter Adjustment

Parameter Adjustment refers to the process of changing the settings or values within a software application to achieve a desired outcome. In the context of the video, adjusting parameters can involve modifying the image's color, size, or other attributes to match the user's vision.

💡Paint Software

Paint Software, such as the Windows standard Paint application, is a digital imaging program that allows users to create and edit images through various tools like brushes, erasers, and color palettes. In the video, paint software is used to remove the background and make adjustments to the line drawing of the image.


A prompt is a text input or command given to an AI system to guide its output. In the context of image generation AI, prompts often describe the desired visual elements or characteristics that the AI should incorporate into the generated image.


Inpainting is a digital image editing technique that involves filling in or repairing missing or damaged parts of an image. In the context of the video, inpainting is used to automatically fill in areas where lines were erased, allowing the AI to generate a coherent image even when manual edits remove parts of the original drawing.

💡Image Quality

Image Quality refers to the clarity, resolution, and overall visual appeal of an image. High-quality images are sharp, detailed, and free from distortion or noise. In the video, improving image quality is a goal as the user seeks to create visually impressive and accurate representations of their desired scenes or subjects.


The video discusses methods for modifying illustrations created by image generation AI.

It covers how to correct images generated by AI and change the clothing depicted in them.

The presenter has previously taught how to use the Paint function in the Image tab to make partial changes.

This time, the video introduces using ControlNet in conjunction with the Intense Image Tab's Intense Paint function for more significant alterations.

The method requires a Stable Diffusion WEBUI environment and the introduction of ControlNet.

The video provides a link in the description for viewers to learn more about the methods discussed.

The process begins by opening the Image tab and uploading the image to be corrected.

The Paint function is used to make desired changes, such as altering paper color without affecting the hairstyle.

ControlNet's Carry function is highlighted as a way to change hair color while maintaining the original image's appearance.

The video demonstrates changing hair color from blonde to black while keeping the shape and style intact.

The presenter notes the challenge of color naming in different languages and the AI's ability to understand color names primarily in English.

The video then moves on to changing clothing using external tools, such as Paint software.

It shows how to remove the background and unwanted lines from the image using basic image editing techniques.

The presenter explains how to re-upload the edited line drawing into the Stable Diffusion WEBUI and use ControlNet to change the clothing.

An example of changing a school uniform to a maid outfit is given, with a note on the limitations and adjustments needed.

The video emphasizes the versatility of Stable Diffusion WEBUI and ControlNet, allowing for precise modifications and creative control over generated images.

The presenter mentions the continuous evolution of AI and the increasing capabilities of tools like Stable Diffusion WEBUI and ControlNet.

The video concludes with an invitation for viewers to subscribe to the channel for more updates and tutorials on AI and related technologies.