Skip to content

Install and Run on AMD GPUs

ClashSAN edited this page Feb 24, 2023 · 31 revisions

Windows

(In Testing) For Windows users, try this fork using Direct-ml. If you get this working, it would be nice to share here: Discussion

make sure you have the modified repositories in stable-diffusion-webui-directml/repositories/:

(rename them to k-diffusion and stable-diffusion-stability-ai) Place any stable diffusion checkpoint (ckpt or safetensor) in the models/Stable-diffusion directory, and double-click webui-user.bat. If you have 4-6gb vram, try adding these flags to webui-user.bat like so:

COMMANDLINE_ARGS=--opt-sub-quad-attention --lowvram --disable-nan-check

You can add --autolaunch to auto open the url for you.

If it looks like it is stuck when installing gfpgan or gfgan, press enter and it should continue.

(The rest below are installation guides for linux with rocm.)

Automatic Installation

(As of 1/15/23 you can just run webui-user.sh and pytorch+rocm should be automatically installed for you.)

  1. Install Python 3.10.6

  2. git clone https://github.com/AUTOMATIC1111/stable-diffusion-webui

  3. Place stable diffusion checkpoint (model.ckpt) in the models/Stable-diffusion directory

  4. For many AMD gpus you MUST Add --precision full --no-half to COMMANDLINE_ARGS= in webui-user.sh to avoid black squares or crashing.*

  5. Run webui.sh

*Certain cards like the Radeon RX 6000 Series and the RX 500 Series will function normally without the option --precision full --no-half, saving plenty of vram. (noted here.)

Running natively

Execute the following:

git clone https://github.com/AUTOMATIC1111/stable-diffusion-webui
cd stable-diffusion-webui
python -m venv venv
source venv/bin/activate
python -m pip install --upgrade pip wheel

# It's possible that you don't need "--precision full", dropping "--no-half" however crashes my drivers
TORCH_COMMAND='pip install torch torchvision --extra-index-url https://download.pytorch.org/whl/rocm5.1.1' python launch.py --precision full --no-half

In following runs you will only need to execute:

cd stable-diffusion-webui
# Optional: "git pull" to update the repository
source venv/bin/activate

# It's possible that you don't need "--precision full", dropping "--no-half" however crashes my drivers
TORCH_COMMAND='pip install torch torchvision --extra-index-url https://download.pytorch.org/whl/rocm5.1.1' python launch.py --precision full --no-half

The first generation after starting the WebUI might take very long, and you might see a message similar to this:

MIOpen(HIP): Warning [SQLiteBase] Missing system database file: gfx1030_40.kdb Performance may degrade. Please follow instructions to install: https://github.com/ROCmSoftwarePlatform/MIOpen#installing-miopen-kernels-package

The next generations should work with regular performance. You can follow the link in the message, and if you happen to use the same operating system, follow the steps there to fix this issue. If there is no clear way to compile or install the MIOpen kernels for your operating system, consider following the "Running inside Docker"-guide below.

Running inside Docker

Pull the latest rocm/pytorch Docker image, start the image and attach to the container (taken from the rocm/pytorch documentation): docker run -it --network=host --device=/dev/kfd --device=/dev/dri --group-add=video --ipc=host --cap-add=SYS_PTRACE --security-opt seccomp=unconfined -v $HOME/dockerx:/dockerx rocm/pytorch

Execute the following inside the container:

cd /dockerx
git clone https://github.com/AUTOMATIC1111/stable-diffusion-webui
cd stable-diffusion-webui
python -m venv venv
source venv/bin/activate
python -m pip install --upgrade pip wheel

# It's possible that you don't need "--precision full", dropping "--no-half" however crashes my drivers
TORCH_COMMAND='pip install torch torchvision --extra-index-url https://download.pytorch.org/whl/rocm5.1.1' REQS_FILE='requirements.txt' python launch.py --precision full --no-half

Following runs will only require you to restart the container, attach to it again and execute the following inside the container: Find the container name from this listing: docker container ls --all, select the one matching the rocm/pytorch image, restart it: docker container restart <container-id> then attach to it: docker exec -it <container-id> bash.

cd /dockerx/stable-diffusion-webui
# Optional: "git pull" to update the repository
source venv/bin/activate

# It's possible that you don't need "--precision full", dropping "--no-half" however crashes my drivers
TORCH_COMMAND='pip install torch torchvision --extra-index-url https://download.pytorch.org/whl/rocm5.1.1' REQS_FILE='requirements.txt' python launch.py --precision full --no-half

The /dockerx folder inside the container should be accessible in your home directory under the same name.

Updating Python version inside Docker

If the web UI becomes incompatible with the pre-installed Python 3.7 version inside the Docker image, here are instructions on how to update it (assuming you have successfully followed "Running inside Docker"):

Execute the following inside the container:

apt install python3.9-full # Confirm every prompt
update-alternatives --install /usr/local/bin/python python /usr/bin/python3.9 1
echo 'PATH=/usr/local/bin:$PATH' >> ~/.bashrc

Then restart the container and attach again. If you check python --version it should now say Python 3.9.5 or newer.

Run rm -rf /dockerx/stable-diffusion-webui/venv inside the container and then follow the steps in "Running inside Docker" again, skipping the git clone https://github.com/AUTOMATIC1111/stable-diffusion-webui and using the modified launch-command below instead:

TORCH_COMMAND='pip install torch torchvision --extra-index-url https://download.pytorch.org/whl/rocm5.1.1' python launch.py --precision full --no-half

It's possible that you don't need "--precision full", dropping "--no-half" however it may not work for everyone. Certain cards like the Radeon RX 6000 Series and the RX 500 Series will function normally without the option --precision full --no-half, saving plenty of vram. (noted here.)

Always use this new launch-command from now on, also when restarting the web UI in following runs.

Clone this wiki locally