One click install and run script (but you still must install python and git)
Outpainting
Inpainting
Color Sketch
Prompt Matrix
Stable Diffusion Upscale
Attention, specify parts of text that the model should pay more attention to
a man in a ((tuxedo)) - will pay more attention to tuxedo
a man in a (tuxedo:1.21) - alternative syntax
select text and press Ctrl+Up or Ctrl+Down (or Command+Up or Command+Down if you’re on a MacOS) to automatically adjust attention to selected text (code contributed by anonymous user)
Loopback, run img2img processing multiple times
X/Y/Z plot, a way to draw a 3 dimensional plot of images with different parameters
Textual Inversion
have as many embeddings as you want and use any names you like for them
use multiple embeddings with different numbers of vectors per token
works with half precision floating point numbers
train embeddings on 8GB (also reports of 6GB working)
Extras tab with:
GFPGAN, neural network that fixes faces
CodeFormer, face restoration tool as an alternative to GFPGAN
RealESRGAN, neural network upscaler
ESRGAN, neural network upscaler with a lot of third party models
SwinIR and Swin2SR (see here), neural network upscalers
LDSR, Latent diffusion super resolution upscaling
Resizing aspect ratio options
Sampling method selection
Adjust sampler eta values (noise multiplier)
More advanced noise setting options
Interrupt processing at any time
4GB video card support (also reports of 2GB working)
Correct seeds for batches
Live prompt token length validation
Generation parameters
parameters you used to generate images are saved with that image
in PNG chunks for PNG, in EXIF for JPEG
can drag the image to PNG info tab to restore generation parameters and automatically copy them into UI
can be disabled in settings
drag and drop an image/text-parameters to promptbox
Read Generation Parameters Button, loads parameters in promptbox to UI
Settings page
Running arbitrary python code from UI (must run with --allow-code to enable)
Mouseover hints for most UI elements
Possible to change defaults/mix/max/step values for UI elements via text config
Tiling support, a checkbox to create images that can be tiled like textures
Progress bar and live image generation preview
Can use a separate neural network to produce previews with almost none VRAM or compute requirement
Negative prompt, an extra text field that allows you to list what you don’t want to see in generated image
Styles, a way to save part of prompt and easily apply them via dropdown later
Variations, a way to generate same image but with tiny differences
Seed resizing, a way to generate same image but at slightly different resolution
CLIP interrogator, a button that tries to guess prompt from an image
Prompt Editing, a way to change prompt mid-generation, say to start making a watermelon and switch to anime girl midway
Batch Processing, process a group of files using img2img
Img2img Alternative, reverse Euler method of cross attention control
Highres Fix, a convenience option to produce high resolution pictures in one click without usual distortions
Reloading checkpoints on the fly
Checkpoint Merger, a tab that allows you to merge up to 3 checkpoints into one
sudo pacman -S yay
yay -S python311 # do not confuse with python3.11 package
Only for 3.11
Then set up env variable in launch script
export python_cmd=”python3.11”
or in webui-user.sh
python_cmd=”python3.11”
2. Navigate to the directory you would like the webui to be installed and execute the following command:
```bash
wget -q https://raw.githubusercontent.com/AUTOMATIC1111/stable-diffusion-webui/master/webui.sh
Stable Diffusion web UI
A web interface for Stable Diffusion, implemented using Gradio library.
Features
Detailed feature showcase with images:
((tuxedo))
- will pay more attention to tuxedo(tuxedo:1.21)
- alternative syntaxCtrl+Up
orCtrl+Down
(orCommand+Up
orCommand+Down
if you’re on a MacOS) to automatically adjust attention to selected text (code contributed by anonymous user)--allow-code
to enable)AND
a cat :1.2 AND a dog AND a penguin :2.2
--xformers
to commandline args)Installation and Running
Make sure the required dependencies are met and follow the instructions available for:
Alternatively, use online services (like Google Colab):
Installation on Windows 10/11 with NVidia-GPUs using release package
sd.webui.zip
from v1.0.0-pre and extract its contents.update.bat
.run.bat
.Automatic Installation on Windows
git clone https://github.com/AUTOMATIC1111/stable-diffusion-webui.git
.webui-user.bat
from Windows Explorer as normal, non-administrator, user.Automatic Installation on Linux
Ubuntu 24.04
sudo add-apt-repository ppa:deadsnakes/ppa sudo apt update sudo apt install python3.11Manjaro/Arch
sudo pacman -S yay yay -S python311 # do not confuse with python3.11 package
Only for 3.11
Then set up env variable in launch script
export python_cmd=”python3.11”
or in webui-user.sh
python_cmd=”python3.11”
Or just clone the repo wherever you want:
webui.sh
.webui-user.sh
for options.Installation on Apple Silicon
Find the instructions here.
Contributing
Here’s how to add code to this repo: Contributing
Documentation
The documentation was moved from this README over to the project’s wiki.
For the purposes of getting Google and other search engines to crawl the wiki, here’s a link to the (not for humans) crawlable wiki.
Credits
Licenses for borrowed code can be found in
Settings -> Licenses
screen, and also inhtml/licenses.html
file.