img2txt stable diffusion. The Stable Diffusion 2. img2txt stable diffusion

 
The Stable Diffusion 2img2txt stable diffusion In this quick episode we do a simple workflow where we upload an image into our SDXL graph inside of ComfyUI and add additional noise to produce an altered i

Don't use other versions unless you are looking for trouble. 4/5 generated image and get the prompt to replicate that image/style. . Linux: run the command webui-user. creates original designs within seconds. This model card gives an overview of all available model checkpoints. If you don't like the results, you can generate new designs an infinite number of times until you find a logo you absolutely love! Watch It In Action. Abstract. Goals. Deforum Stable Diffusion Prompts. The weights were ported from the original implementation. I was using one but it does not work anymore since yesterday. You've already forked stable-diffusion-webui 0 Code Issues Packages Projects Releases Wiki ActivityWe present a dataset of 5,85 billion CLIP-filtered image-text pairs, 14x bigger than LAION-400M, previously the biggest openly accessible image-text dataset in the world - see also our NeurIPS2022 paper. With those sorts of specs, you. 0. Discover amazing ML apps made by the communityPosition the 'Generation Frame' in the right place. A checkpoint (such as CompVis/stable-diffusion-v1-4 or runwayml/stable-diffusion-v1-5) may also be used for more than one task, like text-to-image or image-to-image. GitHub. AIイラストに衣装を着せたときの衣装の状態に関する呪文(プロンプト)についてまとめました。 七海が実際にStable Diffusionで生成したキャラクターを使って検証した衣装の状態に関する呪文をご紹介します。 ※このページから初めて、SThis tutorial shows how to fine-tune a Stable Diffusion model on a custom dataset of {image, caption} pairs. Repeat the process until you achieve the desired outcome. マイクロソフトは DirectML を最適化し、Stable Diffusion で使用されているトランスフォーマーと拡散モデルを高速化することで、Windows ハードウェア・エコシステム全体でより優れた動作を実現しました。 AMD は、Olive のプレリリースに見られるように. 5 it/s (The default software) tensorRT: 8 it/s. ago. See the complete guide for prompt building for a tutorial. Stable Diffusion Hub. txt2txt + img2img + heavy Photoshop. The latest stability ai release is 2. And now Stable Diffusion runs on the Xbox Series X and S! r/StableDiffusion •. 5, ControlNet Linear/OpenPose, DeFlicker Resolve. img2imgの基本的な使い方を解説します。img2imgはStable Diffusionの入力に画像を追加したものです。画像をプロンプトで別の画像に改変できます. 1 I use this = oversaturated, ugly, 3d, render, cartoon, grain, low-res, kitsch, black and white. Use the resulting prompts with text-to-image models like Stable Diffusion to create cool art! Public. Once finished, scroll back up to the top of the page and click Run Prompt Now to generate your AI. Put this in the prompt text box. 前提:Stable. If you click the Option s icon in the prompt box, you can go a little deeper: For Style, you can choose between Anime, Photographic, Digital Art, Comic Book. Generate high-resolution realistic images with AI. By default this will display the “Stable Diffusion Checkpoint” drop down box which can be used to select the different models which you have saved in the “stable-diffusion-webuimodelsStable-diffusion” directory. The model bridges the gap between vision and natural. File "C:\Users\Gros2\stable-diffusion-webui\ldm\models\blip. Intro to ComfyUI. If you put your picture in, would Stable Diffusion start roasting you with tags?. However, at the time he installed it only one . September 14, 2022 AI/ML. 아래 링크를 클릭하면 exe 실행 파일이 다운로드. Get an approximate text prompt, with style, matching an image. /webui. Authors: Christoph Schuhmann, Richard Vencu, Romain Beaumont, Theo Coombes, Cade Gordon, Aarush Katta, Robert Kaczmarczyk, Jenia JitsevFirst, choose a diffusion model on promptoMANIA and put down your prompt or the subject of your image. We tested 45 different GPUs in total — everything that has. 0. img2txt huggingface. Overview Stable Diffusion V3 APIs Text2Image API generates an image from a text prompt. ago Stable diffusion uses openai clip for img2txt and it works pretty well. 이제 부터 Stable Diffusion은 줄여서 SD로 표기하겠습니다. 4. No VAE compared to NAI Blessed. With fp16 it runs at more than 1 it/s but I had problems. The results from the Stable Diffusion and Kandinsky models vary due to their architecture differences and training process; you can generally expect SDXL to produce higher quality images than Stable Diffusion v1. This version of Stable Diffusion creates a server on your local PC that is accessible via its own IP address, but only if you connect through the correct port: 7860. I have been using Stable Diffusion for about 2 weeks now. This model can follow a two-stage model process (though each model can also be used alone); the base model generates an image, and a refiner model takes that image and further enhances its details and quality. Our conditional diffusion model, InstructPix2Pix, is trained on our generated data, and generalizes to real images and. C:stable-diffusion-uimodelsstable-diffusion)Option 1: Every time you generate an image, this text block is generated below your image. I. r/StableDiffusion •. Embeddings (aka textual inversion) are specially trained keywords to enhance images generated using Stable Diffusion. Stable Diffusion is a diffusion model, meaning it learns to generate images by gradually removing noise from a very noisy image. The script outputs an image file based on the model's interpretation of the prompt. Set sampling steps to 20 and sampling method to DPM++ 2M Karras. (You can also experiment with other models. Stable Diffusion img2img support comes to Photoshop. The StableDiffusionPipeline is capable of generating photorealistic images given any text input. Use your browser to go to the Stable Diffusion Online site and click the button that says Get started for free. This is a repo providing same stable diffusion experiments, regarding textual inversion task and captioning task pytorch clip captioning-images img2txt caption-generation caption-generator huggingface latent-diffusion stable-diffusion huggingface-diffusers latent-diffusion-models textual-inversion VGG16 Guided Stable Diffusion. josemuanespinto. 10. ckpt for using v1. Pak jsem si řekl, že zkusím img2txt a ten vytvořil. These encoders are trained to maximize the similarity of (image, text) pairs via a contrastive loss. 5. Training or anything else that needs captioning. lupaspirit. safetensors format. fixとは?. $0. Enter the following commands in the terminal, followed by the enter key, to. 20. r/StableDiffusion. Img2Prompt. 4 (v1. Improving image generation at different aspect ratios using conditional masking during training. 手順2:「gui. Initialize the DSD environment with run all, as described just above. 打开stable-diffusion-webuimodelsstable-diffusion目录,此处为各种模型的存放处。 需要预先存放一个模型才能正常使用。 3. exe, follow instructions. It’s easy to overfit and run into issues like catastrophic forgetting. As we work on our next generation of open-source generative AI models and expand into new modalities, we are excited to. Img2Txt. A decoder, which turns the final 64x64 latent patch into a higher-resolution 512x512 image. 0 的过程,包括下载必要的模型以及如何将它们安装到. Predictions typically complete within 27 seconds. 04 through 22. See the SDXL guide for an alternative setup with SD. env. The idea is to gradually reinterpret the data as the original image gets upscaled, making for better hand/finger structure and facial clarity for even full-body compositions, as well as extremely detailed skin. NMKD Stable Diffusion GUI, perfect for lazy peoples and beginners : Not a WEBui but a software pretty stable self install python / model easy to use face correction + upscale. Negative prompting influences the generation process by acting as a high-dimension anchor,. Base models: stable_diffusion_1. 0) Watch on. 4 Overview. Just go to this address and you will see and learn: Fine-tune Your AI Images With These Simple Prompting Techniques - Stable Diffusion Art (stable-diffusion-art. 手順3:学習を行う. 上記2つの検証を行います。. Diffusers dreambooth runs fine with --gradent_checkpointing and adam8bit, 0. This model runs on Nvidia T4 GPU hardware. Discover amazing ML apps made by the communityThe Stable-Diffusion-v1-5 NSFW REALISM checkpoint was initialized with the weights of the Stable-Diffusion-v1-2 checkpoint and subsequently fine-tuned on 595k steps at resolution 512x512 on "laion-aesthetics v2 5+" and 10% dropping of the text-conditioning to improve classifier-free guidance sampling. ) Come up with a prompt that describe your final picture as accurately as possible. AIArtstable-diffusion-webuimodelsStable-diffusion768-v-ema. Stable DiffusionはNovelAIやMidjourneyとはどう違うの? Stable Diffusionを簡単に使えるツールは結局どれを使えばいいの? 画像生成用のグラフィックボードを買うならどれがオススメ? モデルのckptとsafetensorsって何が違うの? モデルのfp16・fp32・prunedって何?本教程需要一些AI绘画基础,并不是面对0基础人员,如果你没有学习过stable diffusion的基本操作或者对Controlnet插件毫无了解,可以先看看秋葉aaaki等up的教程,做到会存放大模型,会安装插件并且有基本的视频剪辑能力。-----一、准备工作This issue is a workaround for a security vulnerability. The default value is set to 2. com) r/StableDiffusion. First, your text prompt gets projected into a latent vector space by the. Contents. Also you can transform PDF file into images, on output you will get. Go to extensions tab; Click "Install from URL" sub tabtry going to an image editor like photoshop or gimp, find a picture of crumpled up paper, something that has some textures in it and use it as a background, add your logo on the top layer and apply some small amount of noise to the whole thing, make sure to have a good amount of contrast between the background and foreground (if your background. Stable diffusion image-to-text (SDIT) is an advanced image captioning model based on the GPT architecture and uses a diffusion-based training algorithm to improve stability and. information gathering ; txt2img ; img2txt ; stable diffusion ; Stable Diffusion is a tool to create pictures with keywords. Select interrogation types. ChatGPT page. Interrogation: Attempts to generate a list of words and confidence levels that describe an image. 220 and it is a. Using a model is an easy way to achieve a certain style. img2txt github. 使用anaconda进行webui的创建. conda create -n 522-project python=3. For more details on how this dataset was scraped, see Midjourney User. Type and ye shall receive. 9): 0. To shed light on these questions, we present an inference benchmark of Stable Diffusion on different GPUs and CPUs. You can run open-source models, or deploy your own models. This endpoint generates and returns an image from a text passed in the request. Unlike Midjourney, which is a paid and proprietary model, Stable Diffusion is a. flickr30k. Mikromobilita. You will get the same image as if you didn’t put anything. Image: The Verge via Lexica. Stable Diffusion consists of three parts: A text encoder, which turns your prompt into a latent vector. 5);. Now use this as a negative prompt: [the: (ear:1. Images generated by Stable Diffusion based on the prompt we’ve. 5. like 233. Text prompt with description of the things you want in the image to be generated. 5 it/s. novelai用了下,故意挑了些涩图tag,效果还可以 基于stable diffusion,操作和sd类似 他们的介绍文档 价格主要是订阅那一下有点贵,要10刀,送1000token 一张图5token(512*768),细化什么的额外消耗token 这方面倒还好,就是买算力了… 充值token 10刀10000左右,其实还行Model Type. ckpt checkpoint was downloaded), run the following: Technical details regarding Stable Diffusion samplers, confirmed by Katherine: - DDIM and PLMS are originally the Latent Diffusion repo DDIM was implemented by CompVis group and was default (slightly different update rule than the samplers below, eqn 15 in DDIM paper is the update rule vs solving eqn 14's ODE directly) While Stable Diffusion doesn't have a native Image-Variation task, the authors recreated the effects of their Image-Variation script using the Stable Diffusion v1-4 checkpoint. Credit Cost. . NSFW: Attempts to predict if a given image is NSFW. Below are some of the key features: – User-friendly interface, easy to use right in the browser – Supports various image generation options like size, amount, mode,. img2txt online. Scroll to the bottom of the notebook to the Prompts section near the very bottom of the notebook. A dmg file should be downloaded. This parameter controls the number of these denoising steps. The backbone. Still another tool lets people see how attaching different adjectives to a prompt changes the images the AI model spits out. CLIP Interrogator extension for Stable Diffusion WebUI. It’s a fun and creative way to give a unique twist to my images. The goal of this article is to get you up to speed on stable diffusion. 0. To use img2txt stable diffusion, all you need to do is provide the path or URL of the image you want to convert. Height. 152. You need one of these models to use stable diffusion and generally want to chose the latest one that fits your needs. This will allow for the entire image to be seen during training instead of center cropped images, which. Beyond 256². The Stable Diffusion model was created by researchers and engineers from CompVis, Stability AI, Runway, and LAION. ckpt file was a choice. 【画像生成2022】Stable Diffusion第3回 〜日本語のテキストから画像生成(txt2img)を試してみる〜. Features. Download Link. For certain inputs, simply running the model in a convolutional fashion on larger features than it was trained on can sometimes result in interesting results. A negative prompt is a way to use Stable Diffusion in a way that allows the user to specify what he doesn’t want to see, without any extra input. The Payload config is central to everything that Payload does. Installing. More awesome work from Christian Cantrell in his free plugin. 1 (diffusion, upscaling and inpainting checkpoints) 🆕 Now available as a Stable Diffusion Web UI Extension! 🆕. ps1」を実行して設定を行う. Step 1: Go to DiffusionBee’s download page and download the installer for MacOS – Apple Silicon. . 手順1:教師データ等を準備する. The domain img2txt. SDXL,也称为Stable Diffusion XL,是一种备受期待的开源生成式AI模型,最近由StabilityAI向公众发布。它是 SD 之前版本(如 1. Just two. Check it out: Stable Diffusion Photoshop Plugin (0. 仕組みを簡単に説明すると、Upscalerで指定した倍率の解像度に対して. This version is optimized for 8gb of VRAM. The StableDiffusionImg2ImgPipeline uses the diffusion-denoising mechanism proposed in SDEdit: Guided Image Synthesis and Editing with Stochastic Differential Equations by. information gathering ; txt2img ; img2txt ; stable diffusion ; Stable Diffusion is a tool to create pictures with keywords. 5 it/s. Windows: double-click webui-user. I built the easiest-to-use desktop application for running Stable Diffusion on your PC - and it's free for all of you. AI不仅能够自动用文字生成画面,还能够对制定的图片扩展画面意外的内容,也就是根据图片扩展画面内容。这个视频是介绍如何使用stable diffusion中的outpainting(局部重绘)功能来补充图片以外画面,结合PS的粗略处理,可以得到一个完美画面。让AI成为画手的一个得力工具。, 视频播放量 14221、弹幕. In your stable-diffusion-webui folder, create a sub-folder called hypernetworks. PromptMateIO • 7 mo. Get an approximate text prompt, with style, matching an image. Stable Diffusion 2. Does anyone know of any extensions for A1111, that allow you to insert a picture, and it can give you a prompt? I tried a feature like it on my. Already up to date. We follow the original repository and provide basic inference scripts to sample from the models. Others are delightfully strange. Some types of picture include digital illustration, oil painting (usually good results), matte painting, 3d render, medieval map. ; Mind you, the file is over 8GB so while you wait for the download. First, your text prompt gets projected into a latent vector space by the. Doing this on a loop takes advantage of the imprecision in using CLIP latent space walk - fixed seed but two different prompts. 5 is a latent diffusion model initialized from an earlier checkpoint, and further finetuned for 595K steps on 512x512 images. CLIP Interrogator extension for Stable Diffusion WebUI. 667 messages. You can create your own model with a unique style if you want. To differentiate what task you want to use the checkpoint for, you have to load it directly with its corresponding task-specific pipeline class:La manera más sencilla de utilizar Stable Diffusion es registrarte en un editor de imágenes por IA llamado Dream Studio. As of June 2023, Midjourney also gained inpainting and outpainting via the Zoom Out button. The generated image will be named img2img-out. Note: This repo aims to provide a Ready-to-Go setup with TensorFlow environment for Image Captioning Inference using pre-trained model. Get an approximate text prompt, with style, matching an image. Type cmd. 它是一種 潛在 ( 英语 : Latent variable model ) 擴散模型,由慕尼黑大學的CompVis研究團體開發的各. com. To use this pipeline for image-to-image, you’ll need to prepare an initial image to pass to the pipeline. run. 9 conda activate 522-project # install torch 2. 手順3:PowerShellでコマンドを打ち込み、環境を構築する. Output. A decoder, which turns the final 64x64 latent patch into a higher-resolution 512x512 image. Check out the img2img. Stable diffusion is a critical aspect of obtaining high-quality image transformations using Img2Img. When using the "Send to txt2img" or "Send to img2txt" options, the seed and denoising are set, but the "Extras" checkbox is not set so the variation seed settings aren't applied. This model runs on Nvidia A40 (Large) GPU hardware. Inside your subject folder, create yet another subfolder and call it output. These are our findings: Many consumer grade GPUs can do a fine job, since stable diffusion only needs about 5 seconds and 5 GB of VRAM to run. 103. Stable Doodle. Pipeline for text-to-image generation using Stable Diffusion. r/StableDiffusion •. Want to see examples of what you can build with Replicate? Check out our showcase. I do think that your approach will struggle by the fact it's a similar training method on the already limited faceset you have - so if it's not good enough to work already in DFL for producing those missing angles I'm not sure stable-diffusion will let you. for examples:"Logo of a pirate","logo of a sunglass with girl" or something complex like "logo of a ice-cream with snake" etc. Introduction. SD教程•重磅更新!. Help & Questions Megathread! Howdy! u/SandCheezy here again! We just saw another influx of new users. A method to fine tune weights for CLIP and Unet, the language model and the actual image de-noiser used by Stable Diffusion, generously donated to the world by our friends at Novel AI in autumn 2022. ” img2img ” diffusion) can be a powerful technique for creating AI art. 5 model. Copy the prompt, paste it to the Stable Diffusion and press Generate to see generated images. Stable Diffusion is a concealed text-to-image diffusion model, capable of generating photorealistic images from any textual input, fosters independent flexibility in producing remarkable visuals. . 前回、画像生成AI「Stable Diffusion WEB UI」の基本機能を色々試してみました。 ai-china. 0 (SDXL 1. Predictions typically complete within 14 seconds. . All stylized images in this section is generated from the original image below with zero examples. Public. Generated in -4480634. 1M runs. 2. Under the Generate button there is an Interrogate CLIP which when clicked will download the CLIP for reasoning about the Prompt of the image in the current image box and filling it to the prompt. 🖊️ sd-2. Put this in the prompt text box. 以下方式部署的stable diffusion ui仅会使用CPU进行计算,在没有gpu加速的情况下,ai绘图会占用 非常高(几乎全部)的CPU资源 ,并且绘制单张图片的 时间会比较长 ,仅建议CPU性能足够强的情况下使用(作为对比参考,我的使用环境为笔记本平台的5900HX,在默认参数. World of Warcraft? Návrat ke kostce, a vyšel neuvěřitelně. 以 google. 5를 그대로 사용하며, img2txt. py file for more options, including the number of steps. A diffusion model, which repeatedly "denoises" a 64x64 latent image patch. In the 'General Defaults' area, change the width and height to "768". Let's dive in deep and learn how to generate beautiful AI Art based on prom. I’ll go into greater depth on this later in the article. Stable Diffusion web UIをインストールして使えるようにしておく。 Stable Diffusion web UI用のControlNet拡張機能もインストールしておく。 この2つについては下記の記事でやり方等を丁寧にご説明していますので、まだ準備ができていないよという方はそちらも併せて. The easiest way to try it out is to use one of the Colab notebooks: ; GPU Colab ; GPU Colab Img2Img ; GPU Colab Inpainting ; GPU Colab - Tile / Texture generation ; GPU Colab - Loading. Press the Window key (It should be on the left of the space bar on your keyboard), and a search window should appear. txt2img OR "imaging" is mathematically divergent operation, from less bits to more bits, even ARM or RISC-V can do that. Learn the importance, workings, and benefits of using Kiwi Prompt's chat GPT & Google Bard prompts to enhance your stable diffusion writing. Mockup generator (bags, t-shirts, mugs, billboard etc) using Stable Diffusion in-painting. Starting from a random noise, the picture is enhanced several times and the final result is supposed to be as close as possible to the keywords. At least that is what he says. 1M runs. Shortly after the release of Stable Diffusion 2. We would like to show you a description here but the site won’t allow us. Another experimental VAE made using the Blessed script. Aspect ratio is kept but a little data on the left and right is lost. It is an effective and efficient approach that can be applied to image understanding in numerous scenarios, especially when examples are scarce. like 233. 1 1 comment Evnl2020 • 1 yr. I am late on this post. Copy it to your favorite word processor, and apply it the same way as before, by pasting it into the Prompt field and clicking the blue arrow button under Generate. 0 was released in November 2022 and has been entirely funded and developed by Stability AI. The text to image sampling script within Stable Diffusion, known as "txt2img", consumes a text prompt in addition to assorted option parameters covering. Open up your browser, enter "127. ¿Quieres instalar stable diffusion en tu computador y disfrutar de todas sus ventajas? En este tutorial te enseñamos cómo hacerlo paso a paso y sin complicac. Text-to-Image with Stable Diffusion. creates original designs within seconds. Stable diffusion has been making huge waves recently in the AI and art communities (if you don’t know what that is feel free to check out this earlier post). fixは高解像度の画像が生成できるオプションです。. ai and more. This model uses a frozen CLIP ViT-L/14 text. Steps. But it is not the easiest software to use. Trial users get 200 free credits to create prompts, which are entered in the Prompt box. 5 or XL. Diffusion Model就是图像生成领域近年出现的"颠覆性"方法,将图像生成效果和稳定性拔高到了一个新的高度。. With stable diffusion, it really creates some nice stuff for what is already available, like a pizza with specific toppings [0]. Overview Stable Diffusion V3 APIs Text2Image API generates an image from a text prompt. Items you don't want in the image. In this post, I will show how to edit the prompt to image function to add. Stable Diffusion 1. 使用MediaPipe的面部网格注释器的修改输出,在LAION-Face数据集的一个子集上训练了ControlNet,以便在生成面部图像时提供新级别的控. photo of perfect green apple with stem, water droplets, dramatic lighting. It is common to use negative embeddings for anime. Text-to-image models like Stable Diffusion generate an image from a text prompt. 6 API acts as a replacement for Stable Diffusion 1. Local Installation. The CLIP interrogator has two parts: one is the BLIP model, which takes on the function of decoding and reasoning about the text description. . safetensors (5. 手順1:教師データ等を準備する. Get the result. The program needs 16gb of regular RAM to run smoothly. Apply the filter: Apply the stable diffusion filter to your image and observe the results. But it’s not sufficient because the GPU requirements to run these models are still prohibitively expensive for most consumers. You can also upload and replicate non-AI generated images. Documentation is lacking. It was pre-trained being conditioned on the ImageNet-1k classes. fix” to generate images at images larger would be possible using Stable Diffusion alone. Let’s start generating variations to show you how low and high denoising strengths alter your results: Prompt: realistic photo of a road in the middle of an autumn forest with trees in. Are there online Stable diffusion sites that do img2img? 10 upvotes · 7 comments r/StableDiffusion Comfyui + AnimateDiff Text2Vid youtu. You can receive up to four options per prompt. zip. Mage Space has very limited free features, so it may as well be a paid app. So 4 seeds per prompt, 8 total. AUTOMATIC1111のモデルデータは「"stable-diffusion-webuimodelsStable-diffusion"」の中にあります。 正則化画像の用意. Also there is post tagged here where all the links to all resources are. Stable Diffusion - Image to Prompts Run 934. ,AI绘画stable diffusion,AI辅助室内设计controlnet-语义分割控制测试-3. Predictions typically complete within 1 seconds. Sort of new here. 1. . The following resources can be helpful if you're looking for more. Second day with Animatediff, SD1. Additional training is achieved by training a base model with an additional dataset you are. Example outputs . 因為是透過 Stable Diffusion Model 算圖,除了放大解析度外,還能增加細部細節!. In this quick episode we do a simple workflow where we upload an image into our SDXL graph inside of ComfyUI and add additional noise to produce an altered i. Take careful note of the syntax of the example that’s already there. img2txt. ckpt). (Optimized for stable-diffusion (clip ViT-L/14)) 2. But the […]Stable Diffusion是2022年發布的深度學習 文本到图像生成模型。 它主要用於根據文本的描述產生詳細圖像,儘管它也可以應用於其他任務,如內補繪製、外補繪製,以及在提示詞指導下產生圖生圖的转变。. StableDiffusion. . Using stable diffusion and these prompts hand-in-hand, you can easily create stunning and high-quality logos in seconds without needing any design experience. stable-diffusion-LOGO-fine-tuned model trained by nicky007. Unlike other subject-driven generation models, BLIP-Diffusion introduces a new multimodal encoder which is pre-trained to provide subject representation. portrait of a beautiful death queen in a beautiful mansion painting by craig mullins and leyendecker, studio ghibli fantasy close - up shot. Compress the prompt and fixes. xformers: 7 it/s (I recommend this) AITemplate: 10. Download and install the latest Git here. txt2img2img is an. 4 min read. Also, because the Payload source code is fully written in. plugin already! NOTE: Once installed, you will be able to generate images without a subscrip. I) Main use cases of stable diffusion There are a lot of options of how to use stable diffusion, but here are the four main use cases:. This step downloads the Stable Diffusion software (AUTOMATIC1111). 1. In general, the best stable diffusion prompts will have this form: “A [type of picture] of a [main subject], [style cues]* ”.