Run Stable Diffusion 3 Locally! | ComfyUI Tutorial
TLDRThis tutorial guides viewers on running Stable Diffusion 3 Medium locally using ComfyUI. The process begins with downloading the necessary files from Hugging Face, including the sd3 medium safe tensors and text encoders. After updating ComfyUI, the models are installed, and the tutorial demonstrates generating an image with a prompt. The video highlights the impressive results and encourages users to address licensing issues with Stability AI, ending with a community call to action.
Takeaways
- 🌐 Visit Hugging Face to access the Stable Diffusion 3 medium model, which requires filling out a form and agreeing to access the repository.
- 📁 Download essential files including 'sd3 medium.safetensors', 'clip G clip L', 'T5 XXL', and 'fp16' from the Hugging Face repository.
- 🔄 Update ComfyUI by navigating to its directory and running the 'update_comfy_ui.bat' script to ensure compatibility with the new models.
- 📂 Organize downloaded models by placing them in the appropriate folders within the ComfyUI directory, such as the 'clip' and 'checkpoints' folders.
- 🚀 Prepare to run ComfyUI by executing the 'Nvidia GPU dobat' script to start the application with the necessary GPU support.
- 🔍 Load the 'sd3 medium.safetensors' checkpoint in ComfyUI to integrate the Stable Diffusion 3 medium model for image generation.
- 📝 Use natural language prompts in ComfyUI for better response and image generation, as demonstrated by the example prompt about a female character with northern lights-like hair.
- 🎨 Witness the generation of high-quality images with the Stable Diffusion 3 model, showcasing its impressive capabilities.
- 📝 Note the licensing issue mentioned in the video; consider opening an issue or contacting Stability AI to address the licensing concerns.
- 🔧 The video suggests that the community should work together to help update the license for the model's proper use and distribution.
- 👋 The tutorial concludes with a reminder to enjoy the capabilities of the newly released Stable Diffusion 3 model and to have a great day.
Q & A
What is the main topic of the video tutorial?
-The main topic of the video tutorial is how to use Stable Diffusion 3 Medium and ComfyUI locally.
Where should one go to access the Stable Diffusion 3 model?
-To access the Stable Diffusion 3 model, one should go to Hugging Face and fill out the form to gain access to the repository.
What files does the user need to download from Hugging Face for Stable Diffusion 3 Medium?
-The user needs to download the SD3 Medium safe tensors, text encoders including CLIP G, CLIP L, and T5 XXL, all in fp16 format.
What is the purpose of updating ComfyUI before installing new models?
-Updating ComfyUI ensures that the software is compatible with the new models and provides the latest features and bug fixes.
How does one update ComfyUI according to the tutorial?
-To update ComfyUI, one should go to the ComfyUI directory, navigate to the 'update' folder, and run the 'update_comfy_ui.bat' file.
What is the recommended workflow to use with the Stable Diffusion 3 Medium model?
-The tutorial recommends using the 'basic inference workflow' with the Stable Diffusion 3 Medium model.
Where should the downloaded models and checkpoints be placed within the ComfyUI directory structure?
-The downloaded models should be placed in the 'clip' folder under 'models', and the checkpoints should be placed in the 'checkpoints' folder, preferably in a new 'sd3' folder.
What is the significance of the 'Q prompt' in the ComfyUI workflow?
-The 'Q prompt' is used to input a description or prompt for the model to generate an image based on the provided text.
What type of image is generated in the example provided in the script?
-The example generates an image of a female character with long flowing hair made of ethereal swirling patterns resembling the Northern Lights or Aurora Borealis.
What issue is mentioned regarding the licensing of the Stable Diffusion 3 model?
-The issue mentioned is that the licensing is a bit unclear or 'messed up', and the community is encouraged to open an issue or contact Stability AI to update the license.
How does the tutorial describe the generated image quality of the Stable Diffusion 3 model?
-The tutorial describes the generated image quality as 'really amazing' and expresses excitement about the release of the model's weights for free.
Outlines
🎨 Introduction to Using Stable Diffusion 3 Medium
The video begins with an introduction to the Stable Diffusion 3 Medium model, which has just been released. The host guides viewers on how to access and use this gated model by visiting Hugging Face, filling out a form, and agreeing to access the repository. The process includes downloading necessary files such as the 'sd3 medium.safetensors', text encoders like 'clip G clip L and T5 xx fp16', and the 'comfy UI workflows'. The host also explains the need to close and update Comfy UI before proceeding with the installation of the new models.
🔄 Updating Comfy UI and Installing Models
This paragraph details the steps to update Comfy UI and install new models. The host instructs viewers to navigate to the Comfy UI directory and run the 'update comfy ui.bat' file to ensure they have the latest version. After updating, the host guides viewers to install the CLIP models by placing the downloaded files into the appropriate folders within the Comfy UI directory. Additionally, a new 'sd3' folder is created for the 'sd3 medium.safetensors' file, which is then added to the checkpoints folder, preparing the system for the next steps.
🚀 Starting Comfy UI with the New Model
The host demonstrates how to start using the new Stable Diffusion 3 Medium model with Comfy UI. After ensuring the Nvidia GPU 'dobat' is running, the host switches to another machine to load the 'sd3 medium.safetensors' checkpoint and the CLIP files. The workflow is set up to use a natural language prompt, which is different from the traditional 'boru tag' style, and the host uses an example prompt provided by the model developers to generate an image of a female character with hair resembling the northern lights. The host expresses excitement about the model's capabilities and encourages the community to help address licensing issues by opening issues or contacting Stability AI.
Mindmap
Keywords
💡Stable Diffusion 3
💡ComfyUI
💡Hugging Face
💡Text Encoders
💡CLIP Models
💡Safe Tensors
💡Checkpoints
💡Nvidia GPU
💡Workflow
💡Q Prompt
💡Ethereal
💡Aurora Borealis
Highlights
Introduction to using Stable Diffusion 3 Medium and ComfyUI.
Accessing the gated model on Hugging Face and filling out the form to gain access.
Downloading required files such as sd3 medium safe tensors and text encoders.
Instructions on updating ComfyUI to the latest version.
Installing CLIP models into the ComfyUI directory.
Creating a new folder for sd3 medium safe tensors in the checkpoints directory.
Starting ComfyUI with the Nvidia GPU dobat.
Loading the checkpoint sd3 medium safe tensors in ComfyUI.
Using the example prompt for generating an image with natural language.
Explanation of the model's response to the prompt featuring a female character with northern lights hair.
Observation of the model's incredible image generation capabilities.
Discussion on the model's licensing issues and the need for community involvement.
Encouragement to open an issue with Stability AI regarding the license.
Final thoughts on the tutorial and a reminder to have a great day.