Generative AI Art with Forge – Quickstart Guide

Last UpdatedChanges
3/13/2023First Version
6/10/2023Shock update! Forge End-of-Life?
8/20/2024Surprise Major Update!

What is Forge?

?? Signs of Life! – 8/20/2024 ??

The repository received a surprise Major Update this week, bringing Flux model support, and a number of interface enhancements!

?? Announcement from Forge Developer – 6/8/2024 ??

In a surprise announcement from Developer Illyasviel, posted June 8th 2024, it was confirmed what many had feared – Forge is now an experimental interface/repository, and all users are encouraged to switch back to Automatic1111 WebUI.

Going forward, do not install Forge looking for a useable Stable Diffusion Interface.

Forge is an interface for creating images (inference) with Stable Diffusion, aimed at advanced users who may have previously used Automatic 1111. Building on Automatic 1111, Forge offers many improvements, additions and adjustments to both the back and front-end, to enhance the image creation experience.

Forge comes from the same developer who brought us the excellent beginner/intermediate level interface Fooocus (see our Quickstart guide!) and the original ControlNet implementation (see our ControlNet guide!).

Why should I be excited by Forge?

While Forge is built on the bones of Automatic 1111, through some coding magic it promises tremendous image generation speed increases and stability over the original interface;

Compared to original Automatic 1111 (for SDXL inference at 1024px), you can expect the following speed increases and VRAM utilization improvements:

  • If you use common GPU like 8GB vram, you can expect to get about 30~45% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 700MB to 1.3GB, the maximum diffusion resolution (that will not out of memory (OOM)) will increase about 2x to 3x, and the maximum diffusion batch size (that will not OOM) will increase about 4x to 6x.
  • If you use less powerful GPU like 6GB vram, you can expect to get about 60~75% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 800MB to 1.5GB, the maximum diffusion resolution (that will not OOM) will increase about 3x, and the maximum diffusion batch size (that will not OOM) will increase about 4x.
  • If you use powerful GPU like 4090 with 24GB vram, you can expect to get about 3~6% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 1GB to 1.4GB, the maximum diffusion resolution (that will not OOM) will increase about 1.6x, and the maximum diffusion batch size (that will not OOM) will increase about 2x.

And the claims are true!

Forge also integrates many advanced extensions and features, including Self-Attention Guidance, Kohya High Res Fix, FreeU, StyleAlign, Hypertile, Stable Video Diffusion, and Z123.

Civitai - Forge Extensions
Integrated features

How can I use Forge?

The good news is that installing Forge is super simple! There are two methods – one for users familiar with Github (anyone who has previously used Auto1111 will most likely be), and one extremely straightforward method for less experienced users. We’ll look at the extremely straightforward method!

On Windows, click here to download the zipped package! It’s 1.7 GB zipped, which, when uncompressed, expands to around 5.3 GB, and when installed, takes up about 10 GB of disk space (without any models).

Once you have downloaded and extracted the zipped package, you’ll see the following files;

Civitai - Forge Installation

It’s important to first run update.bat – this ensures the latest files are pulled down for install.

After running update.bat, run run.bat to install and launch!

On subsequent launches, you’ll need to find webui-user.bat to launch the Forge Interface.

Civitai - Forge Launch

Architecture

Per the Forge documentation;

Forge backend removes all WebUI's codes related to resource management and reworked everything. All previous CMD flags like medvram, lowvram, medvram-sdxl, precision full, no half, no half vae, attention_xxx, upcast unet, ... are all REMOVED. Adding these flags will not cause error but they will not do anything now. We highly encourage Forge users to remove all cmd flags and let Forge to decide how to load models.

Without any cmd flag, Forge can run SDXL with 4GB vram and SD1.5 with 2GB vram.

Note: There was some drama in the Forge Github about the backend being “stolen” from ComfyUI, to which the developer responded. The claim was shown to be unsubstantiated – forge is 100% Automatic 1111, with coding and infrastructure changes to boost performance.

Advanced Features

Thanks to a Unet Patcher, many advanced features are implemented in Forge which are not natively available, supported, or much slower and VRAM hungry in Automatic 1111. Some are explored below;

IP Adapter with Masking

Via ControlNet and the IP Adapter Preprocessor, we can supply two images, and mask them together to provide a single unified image! We’ll expand upon this section in future to show exactly how to do this!

Civitai - Forge IP Adapter Example

New Samplers

  • DDPM
  • DDPM Karras
  • DPM++ 2M Turbo
  • DPM++ 2M SDE Turbo
  • LCM Karras
  • Euler A Turbo

Stable Video Diffusion (SVD)

Forge supports Stable Video Diffusion natively, with reduced requirements to generate short SVD videos. You can also easily pass your generations from the txt2img tab into the SVD tab for generation!

Requirements, Notes, & Limitations

  • Forge has very low VRAM requirements in comparison to Automatic 1111, and other interfaces, but you’ll still need a minimum of 4GB of VRAM for SDXL image generation, and 2GB of VRAM for SD 1.5 image generation!
  • Not all of the thousands of Automatic 1111 extensions work with Forge! You may come across extensions which install, but don’t load and throw errors in the console.

The Future

We’ll continue to expand this quickstart guide with more information as it becomes available!