blockchainchad / progrockdiffusion

A command line version of Disco Diffusion

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

progrockdiffusion

As of October 2022 I am no longer actively updating Prog Rock Diffusion. Occasional bug fixes or minor improvements may happen.

A command line version of Disco Diffusion. A GUI is available with a focused set of settings here: GUI Instructions

Quick note on updating:

A new version was released today, Sept 16 2022, which includes two new models for Open Clip. As a result you will need up update Open Clip. To do this, delete the open-clip directory inside your PRD directory, then run the command to fetch the newer version: (run this in your progrockdiffusion directory)

git clone https://github.com/mlfoundations/open_clip.git

Hardware prerequisites

You will need at least 16gb of RAM if not using a GPU. An Nvidia GPU is highly recommended! The speed improvement is massive. 8gb is probably the minimum amount of GPU memory.

This author has an RTX 3080 with 10gb and it runs fairly well, but some advanced features are not possible with "only" 10gb.

You'll also need between 20 and 40gb of free disk space, depending on which models you enable.

Software prerequisties

Linux

Ubuntu 20.04 or similar (A docker environment, VM, or Windows Subsystem for Linux should work provided it can access your GPU).

If using a GPU: CUDA 11.4+ (installation instructions can be found here: https://developer.nvidia.com/cuda-11-4-1-download-archive).

Windows

Windows 10 or 11 (If using a GPU, NVIDIA drivers installed) Other versions may work but are untested

MacOS

Minimal testing has been done with the latest MacOS on an M1 Macbook Air. PLEASE NOTE: GPU acceleration is not yet supported. It will work, but it will be slow.

If using an NVIDIA GPU, test NVIDIA drivers

You can test that your environment is working properly by running:

nvidia-smi

The output should indicate a driver version, CUDA version, and so on. If you get an error, stop here and troubleshoot how to get Nvidia drivers, CUDA, and/or a connection to your GPU with the environment you're using.

First time setup

Download and install Anaconda

[Linux]

sudo apt update
sudo apt upgrade -y
wget https://repo.anaconda.com/archive/Anaconda3-2021.11-Linux-x86_64.sh
bash Anaconda3-2021.11-Linux-x86_64.sh
respond 'yes' to accept license terms and provide install dir when prompted
respond 'yes' to run conda initialization

Logout and back in for the changes to take effect

[Windows]

Download from here and install: https://www.anaconda.com/products/individual

From the start menu, open a "Anaconda Powershell Prompt" (Powershell is important)

[MacOS]

Install Homebrew:

/bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/install.sh)"

Restart your terminal, then:

brew install miniforge
conda init zsh

Restart your terminal again.

Create prog rock diffusion env

[Linux and Windows]

conda create --name progrockdiffusion python=3.7

[MacOS]

conda create --name progrockdiffusion python=3.8

[All Platforms]

conda activate progrockdiffusion
conda install -c anaconda git

Now change to whatever base directory you want ProgRockDiffusion to go into.

Clone the prog rock diffusion repo

git clone https://github.com/lowfuel/progrockdiffusion.git
cd progrockdiffusion

Note: the "cd" command above is important, as the next steps will add additional libraries and data to ProgRockDiffusion

From here on out, this is the directory you'll want to be in when you use the program.

Install the required libraries and tools

git clone https://github.com/crowsonkb/guided-diffusion
git clone https://github.com/openai/CLIP.git
git clone https://github.com/mlfoundations/open_clip.git
git clone https://github.com/assafshocher/ResizeRight.git
git clone https://github.com/CompVis/taming-transformers
pip install -e ./CLIP
pip install -e ./guided-diffusion
pip install -e ./taming-transformers
pip install lpips datetime timm json5 numexpr

Basic or GPU Accelerated PyTorch

You defnitely should install the GPU version if you have an NVIDIA card. It's almost 30x faster. Otherwise, you can install the CPU version instead (required for MacOS)

EITHER Install GPU accelerated PyTorch

pip install torch==1.10.2+cu113 -f https://download.pytorch.org/whl/cu113/torch_stable.html
pip install torchvision==0.11.3+cu113 -f https://download.pytorch.org/whl/cu113/torch_stable.html
pip install torchaudio==0.10.2+cu113 -f https://download.pytorch.org/whl/cu113/torch_stable.html

OR install the basic CPU version of PyTorch (warning - very slow!)

pip install torch==1.11.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html
pip install torchvision==0.12.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html
pip install torchaudio==0.11.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html

Install remaining libraries and tools

[MacOS]

export GRPC_PYTHON_BUILD_SYSTEM_OPENSSL=1
export GRPC_PYTHON_BUILD_SYSTEM_ZLIB=1
pip install grpcio

[All Platforms]

pip install ipywidgets omegaconf pytorch_lightning einops
pip install matplotlib pandas
pip install 'Pillow>=9.1.0'
conda install opencv

[Linux] Depending on your Linux platform, you may get an error about libGL.so.1 If you do, try installing these dependencies:

sudo apt-get install ffmpeg libsm6 libxext6 -y

[Linux] Finally:

sudo apt install imagemagick

Use

NOTE: On your first run it might appear to hang. Let it go for a good while, though, as it might just be downloading models. Somtimes there is no feedback during the download process (why? Who knows)

If you've opened a new terminal or powershell prompt, you may need to activate your ProgRockDiffusion session again:

conda activate progrockdiffusion

CD to the directory where you installed ProgRockDiffusion. Now you're ready!

The simplest way to run it is:

[Linux]

python3 prd.py

[Windows and MacOS]

python prd.py

This will generate an image using the settings from "settings.json", which you could edit to adjust the defaults (or, better yet, make a copy of it and tell prd to use an alternative settings file using the command line arguments below).

Note: On windows you'll type "python" instead of "python3" in the commands below.

usage: python3 prd.py [-h] [-s SETTINGS] [-o OUTPUT] [-p PROMPT]

Generate images from text prompts.
By default, the supplied settings.json file will be used.
You can edit that, and/or use the options below to fine tune:

Optional arguments:
  -h, --help            show this help message and exit
  -s SETTINGS, --settings SETTINGS
                        A settings JSON file to use, best to put in quotes
                        Can be specified more than once to layer settings on top of one another
  -o OUTPUT, --output OUTPUT
                        What output directory to use within images_out
  -p PROMPT, --prompt PROMPT
                        Override the prompt
  -i, --ignoreseed
                        Use a random seed instead of what is in your settings file

  -c, --cpu CORES
                        Force CPU mode, and (optionally) specify how many threads to run.

  --cuda DEVICE-ID
                        Specify which CUDA device ID to use for rendering (default: 0).

  -g PERCENT, --geninit PERCENT:
                        Will save an image called geninit.png at PERCENT of overall steps, for use with --useinit

  -u, --useinit:
                        Forces use of geninit.png as an init_image starting at 20% of defined steps.

Usage examples:

To use the Default output directory and settings from settings.json:
 python3 prd.py

To use your own settings.json file (note that putting it in quotes can help parse errors):
 python3 prd.py -s "some_directory/mysettings.json"

Note that multiple settings files are allowed. They're parsed in order. The values present are applied over any previous value:
 python3 prd.py -s "some_directory/mysettings.json" -s "highres.json"

To quickly just override the output directory name and the prompt:
 python3 prd.py -p "A cool image of the author of this program" -o Coolguy

Multiple prompts with weight values are supported:
 python3 prd.py -p "A cool image of the author of this program" -p "Pale Blue Sky:.5"

You can ignore the seed coming from a settings file by adding -i, resulting in a new random seed

To force use of the CPU for image generation, add a -c or --cpu (warning: VERY slow):
 python3 prd.py -c

To specify which CUDA device to use (advanced) by device ID (default is 0):
 python3 prd.py --cuda 1

Simply edit the settings.json file provided, or BETTER YET copy it and make several that include your favorite settings. Note that multiple settings files can be specified in your command, and they'll be loaded in order. Settings.json is always loaded, and any specified after that are layered on top (they only need to contain the settings you want to tweak). For example you could have a settings file that just contains a higher width, height, and more steps, for when you want to make a high-quality image. Layer that on top of your regular settings and it will apply those values without changing anything else.

GUI Instructions

Visual Diffusion

Visual Diffusion is a GUI for progrockdiffusion. Please direct all GUI related issue/features/requests to Visual Diffusion

image

Install

cd progrockdiffusion
git clone https://github.com/KnoBuddy/visualdiffusion/

[Windows]

conda activate progrockdiffusion
cd progrockdiffusion
python visualdiffusion\prdgui.py

[Linux/MacOS]

conda activate progrockdiffusion
cd progrockdiffusion
python3 visualdiffusion\prdgui.py

Tips and Troubleshooting

Get a random artist

In your prompt, if you use _artist_ instead of an artists name, an artist will be picked at random from artists.txt

If you get an error about pandas needing a different verison of numpy, you can try:

pip install --force-reinstall numpy

If you are getting "OMP: Error #15: Initializing libiomp5md.dll, but found libiomp5md.dll already initialized"

This seems to be because of a Pytorch compiling bug for Intel CPUs. You can set an environment variable that will fix this, either on your machine (if you know how to do that), or by editing prd.py. To do it by editing prd.py, find the line that says "import os" and add the following right below it:

os.environ['KMP_DUPLICATE_LIB_OK']='True'

Switch between GPU and CPU modes

Let's assume you installed the GPU version. You can adjust these instructions if you did CPU first, of course. Clone your existing conda environment:

conda create --name prdcpu --clone progrockdiffusion
conda activate prdcpu

Now install the CPU version of pytorch:

pip install torch==1.11.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html
pip install torchvision==0.12.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html
pip install torchaudio==0.11.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html

All set! You can now switch between the two by simply doing:

conda activate progrockdiffusion
conda activate prdcpu

Store clip models in PRD models directory

By default clip will store its models in your home directory, under '.cache/clip' This can be useful if you use multiple PRD directories, or PRD as well as Disco Diffusion, as you will only need to store the models once. However, some people prefer to keep the models with PRD, so that they can be stored on a drive different to the one their home directory is on. To store them in the models folder, set an environment variable of LOCAL_CLIP_MODELS (to any value, it doesn't matter it just needs to be set). Note: This does not effect OpenClip models, which has no option to change where the models are stored

Notes

  • Animation is not supported

About

A command line version of Disco Diffusion

License:MIT License


Languages

Language:Python 98.4%Language:Dockerfile 1.6%