Get Creative With Image to Image & Inpainting in Playground AI

Playground AI
2 Mar 202310:05

TLDRThe video script offers a step-by-step guide on utilizing image-to-image techniques in AI playground to create and refine compositions. It demonstrates how to set up parameters, use prompts, and apply filters like 'storybook' for a watercolor effect. The process includes creating variations of an image, masking, and drawing tools to craft landscapes, with the goal of achieving a desired composition and aesthetic. The video emphasizes the creative potential of AI in image generation and editing.


  • 🎨 Utilize image to image functionality in AI playgrounds to refine compositions and reduce the number of image generations needed.
  • πŸ–ŒοΈ Set up your parameters, such as stable diffusion version, width, height, and prompt, to begin the image generation process.
  • 🦝 Use specific and descriptive prompts, like 'anthropomorphic raccoon wearing a suit in Top Hat', to guide the AI in creating desired images.
  • 🚫 Include negative prompts to exclude undesired elements from the generated images.
  • 🎬 Choose a style, such as Pixar, to give the generated images a particular aesthetic.
  • πŸ“Έ Generate multiple images and select the ones that best match your vision, deleting those that do not.
  • 🌟 Use the 'create variations' feature to make slight adjustments to the chosen image, utilizing image strength to control the degree of variation.
  • πŸ–ΌοΈ Mask out unwanted parts of an image, such as the background, to isolate the subject and prepare for adding new elements.
  • πŸ™οΈ Add typical environmental elements, like cars and people, to create a more realistic and complete scene.
  • πŸ–ŒοΈ Experiment with the drawing tool to manually create landscapes or other scenes, then generate an image based on the drawing.
  • 🎨 Apply filters, like storybook, to achieve different artistic styles in the final image.
  • πŸ”„ Use the upscaling feature to enhance the resolution of the image if desired, for sharing or further use.

Q & A

  • What is the main topic of the video?

    -The main topic of the video is using image to image in playground AI to set composition and generate images with specific characteristics.

  • Which AI model is used in the video for image generation?

    -The AI model used in the video is Stable Diffusion 1.5.

  • What are the dimensions set for the image generation in the video?

    -The dimensions set for the image generation are width at 512 and height at 768.

  • What is the initial prompt used in the video to generate the images?

    -The initial prompt used is 'cute and adorable raccoon wearing a suit in Top Hat' with the word 'anthropomorphic' to give the raccoon human-like features.

  • How does the video demonstrate the use of negative prompts?

    -The video demonstrates the use of negative prompts by including them in a specific area to refine the image generation process and avoid unwanted elements in the output images.

  • What is the purpose of using 'standing on the sidewalk' in the prompt?

    -Using 'standing on the sidewalk' in the prompt helps to describe the desired action for the character, which aids in generating images that show the full body and context of the character.

  • How does the video explain the use of image strength in image to image variations?

    -The video explains that image strength is related to the creativity of the AI in generating variations. A higher image strength makes the generated image more similar to the original, while a lower image strength allows for more creative variations.

  • What is the masking tool used for in the video?

    -The masking tool is used to isolate specific areas of the image for changes, such as painting around the character to remove the background or modify certain parts of the image.

  • How does the video utilize the drawing tool in the image to image section?

    -The drawing tool is utilized to create a custom landscape image by painting different elements like the sky, water, mountains, and grass directly onto the canvas before generating the final image.

  • What filter is applied to the landscape image in the video, and why?

    -The storybook filter is applied to the landscape image to give it a watercolor style, moving away from a photorealistic look and adding a creative touch to the final image.

  • What is the final step shown in the video for enhancing the image?

    -The final step shown is upscaling the image by four times using the actions in the platform, which improves the resolution and detail of the generated image.



🎨 Image-to-Image Composition and Variations

The paragraph introduces the concept of using image-to-image techniques in playground AI to enhance the composition process. It explains how this method can reduce the number of image generations needed. The example provided involves setting up parameters such as columns, stable diffusion version, width, height, and prompt guidance. The tutorial demonstrates creating an image of an anthropomorphic raccoon wearing a top hat, emphasizing the use of specific prompts and negative prompts to refine the results. It also discusses the process of regenerating images to achieve a desired composition and highlights the benefits of describing actions within the prompt. The use of the 'create variations' feature is explored, showing how image strength affects the AI's creativity and adherence to the original image. The paragraph concludes with a discussion on masking the background and adding new elements to the image to enhance its context and realism.


πŸ–ŒοΈ Utilizing the Drawing Tool and Filters for Image-to-Image

This section delves into the use of the drawing tool in the image-to-image process, which is often overlooked. It begins by explaining how to set up the workspace for a landscape-style image and selecting appropriate parameters. The tutorial then guides the user through the process of drawing elements such as a sky, clouds, water, and mountains using different colors and brush sizes. The importance of working backwards, starting with the sky and moving to the foreground, is emphasized. After drawing, the user is instructed to describe the landscape in the prompt and generate an image based on the drawing. The paragraph continues with the application of filters, such as the storybook filter, to achieve a desired artistic style. The process of refining the image through additional generations and adjustments is discussed, highlighting the flexibility and creativity offered by the image-to-image feature.


πŸ‘‹ Conclusion and Future Video Suggestions

The final paragraph wraps up the video script by inviting viewers to share their thoughts and suggestions for future content in the comments section. It encourages interaction with the community and expresses a willingness to explore topics of interest to the audience. The closing statement is a friendly and engaging reminder that the video series aims to be both informative and enjoyable for viewers of all ages.



πŸ’‘Image to Image

Image to Image is a technique discussed in the video that involves using an existing image as a reference or base to generate a new image. This process is part of AI-based image generation and manipulation, where the AI model, in this case, stable diffusion 1.5, takes cues from the original image to create variations or new compositions. The video demonstrates how this can be used to refine and improve the generated images by adjusting the image strength and using the masking and drawing tools.

πŸ’‘Stable Diffusion 1.5

Stable Diffusion 1.5 is an AI model mentioned in the video that is used for generating images. It is a type of deep learning algorithm that can take a text prompt and produce an image based on that description. The model is capable of understanding and applying various stylistic elements and compositional improvements to the generated images as directed by the user.


In the context of the video, a 'prompt' is a text description or a set of keywords that guide the AI in generating an image. The prompt serves as the input for the AI model, which then interprets the text to create a visual representation. The effectiveness of the generated image often depends on the clarity and specificity of the prompt.


Anthropomorphic refers to the attribution of human traits, characteristics, or features to non-human entities, such as animals or objects. In the video, the term is used to describe the desired appearance of the raccoon character, indicating that the AI should generate an image of a raccoon with human-like features, such as wearing a suit and standing on a sidewalk.


Masking is a technique used in image editing where certain parts of an image are isolated and manipulated while the rest of the image remains unchanged. In the video, the creator uses a masking tool to remove the background of the generated image, allowing for the addition of a new, more fitting background.

πŸ’‘Drawing Tool

The drawing tool is an interface within the AI image generation platform that allows users to manually create or modify images by drawing shapes and lines. This tool can be used to add elements to a generated image or to refine details that the AI may not have captured perfectly.

πŸ’‘Image Strength

Image strength is a parameter in the AI image generation process that determines the degree to which the generated image adheres to the reference image. A higher image strength means the AI will produce an image closer to the original, while a lower image strength allows for more creative variations.

πŸ’‘In Paint

In Paint refers to a feature within the AI image generation platform that allows users to make manual adjustments or edits to the generated images. This can include fixing details, adding filters, or making other aesthetic changes to enhance the final output.


A filter in image editing is a function that alters the appearance of an image, often to achieve a specific aesthetic or visual effect. Filters can be used to change colors, add textures, or apply stylistic enhancements to the image.


Composition refers to the arrangement of visual elements within an image to create a harmonious and balanced whole. It is an essential aspect of art and design, and good composition can make an image more engaging and aesthetically pleasing.


In the context of the video, 'community' refers to the group of individuals who share a common interest in AI-generated images and the platform where they interact, share their creations, and provide feedback. The community aspect encourages collaboration, learning, and the exchange of ideas among its members.


Introduction to using image to image in playground AI for composition setting.

Utilizing stable diffusion 1.5 for image generation with specific width and height settings.

The importance of prompt guidance, quality, and details in the image generation process.

Selecting the Euler, ancestral sampler for image generation.

Using a combination of prompts to create an anthropomorphic raccoon character.

The inclusion of negative prompts to refine image generation.

The strategy of regenerating images to achieve desired composition.

The concept of full body and action descriptions in prompts to guide character positioning.

The use of the 'play tune' option for a Pixar style image.

Creating variations of an image using the 'create variations' feature.

Adjusting image strength for different levels of creativity and adherence to the original image.

Masking out the background to customize the scene.

Adding elements to the background to enhance the image's context.

Using the drawing tool to create landscapes and other scenes from scratch.

The process of painting and erasing in the drawing tool for image creation.

Describing the drawn landscape in the prompt for image generation.

Applying filters like storybook to achieve a specific artistic style.

Upscaling the final image for higher resolution and detail.