Go to file
2022-12-13 16:06:06 -08:00
.dev_scripts Replace --full_precision with --precision that works even if not specified 2022-09-20 17:08:00 -04:00
.github Merge remote-tracking branch 'origin/main' into dev/diffusers 2022-12-13 16:06:06 -08:00
assets merge release-candidate-1-3-2 into main. 2022-11-12 17:17:07 +00:00
backend Merge remote-tracking branch 'origin/main' into dev/diffusers 2022-12-11 11:23:52 -08:00
binary_installer Merge remote-tracking branch 'origin/main' into dev/diffusers 2022-12-11 11:23:52 -08:00
configs Merge remote-tracking branch 'origin/main' into dev/diffusers 2022-12-10 21:19:57 -08:00
docker-build Make Dockerfile cloud ready (tested on runpod) (#1950) 2022-12-12 17:54:42 -05:00
docs Typo fix in INSTALL_AUTOMATED.md (#1968) 2022-12-13 19:13:28 +01:00
environments-and-requirements Merge remote-tracking branch 'origin/main' into dev/diffusers 2022-12-11 11:23:52 -08:00
frontend Fresh Frontend Build 2022-12-13 20:45:37 +13:00
installer (install) fix segfault on macos when using homebrew 2022-12-13 11:39:08 -05:00
ldm Merge branch 'main' into dev/diffusers 2022-12-12 23:09:04 -08:00
notebooks Merge dev into main for 2.2.0 (#1642) 2022-11-30 16:12:23 -05:00
scripts Merge remote-tracking branch 'origin/main' into dev/diffusers 2022-12-12 11:17:25 -08:00
server add option to show intermediate latent space 2022-11-02 17:53:11 -04:00
static Merge dev into main for 2.2.0 (#1642) 2022-11-30 16:12:23 -05:00
tests Merge remote-tracking branch 'origin/main' into dev/diffusers 2022-12-13 16:06:06 -08:00
.dockerignore Optimized Docker build with support for external working directory (#1544) 2022-12-06 13:28:07 +01:00
.editorconfig Merge dev into main for 2.2.0 (#1642) 2022-11-30 16:12:23 -05:00
.gitattributes Update .gitattributes 2022-08-29 16:58:41 -05:00
.gitignore Add invokeai.init to gitignore 2022-12-11 11:19:22 -05:00
.gitmodules remove src directory, which is gumming up conda installs; addresses issue #77 2022-08-25 10:43:05 -04:00
.prettierrc.yaml change printWidth for markdown files to 80 2022-09-17 02:23:00 +02:00
CODE_OF_CONDUCT.md Merge dev into main for 2.2.0 (#1642) 2022-11-30 16:12:23 -05:00
InvokeAI_Statement_of_Values.md Merge dev into main for 2.2.0 (#1642) 2022-11-30 16:12:23 -05:00
LICENSE adding license using GitHub template 2022-10-17 12:09:24 -04:00
LICENSE-ModelWeights.txt added assertion checks for out-of-bound arguments; added various copyright and license agreement files 2022-08-24 09:22:27 -04:00
main.py Merge dev into main for 2.2.0 (#1642) 2022-11-30 16:12:23 -05:00
mkdocs.yml Merge dev into main for 2.2.0 (#1642) 2022-11-30 16:12:23 -05:00
README.md Clean up readme 2022-12-06 06:58:26 -05:00
setup.py Simple Installer for Unified Directory Structure, Initial Implementation (#1819) 2022-12-11 00:37:08 -05:00
shell.nix nix: add shell.nix file 2022-10-25 07:08:31 -04:00
Stable_Diffusion_v1_Model_Card.md stable diffusion 2022-08-10 16:30:49 +02:00

project logo

InvokeAI: A Stable Diffusion Toolkit

discord badge

latest release badge github stars badge github forks badge

CI checks on main badge CI checks on dev badge latest commit to dev badge

github open issues badge github open prs badge

This is a fork of CompVis/stable-diffusion, the open source text-to-image generator. It provides a streamlined process with various new features and options to aid the image generation process. It runs on Windows, macOS and Linux machines, with GPU cards with as little as 4 GB of RAM. It provides both a polished Web interface (see below), and an easy-to-use command-line interface.

Quick links: [How to Install] [Discord Server] [Documentation and Tutorials] [Code and Downloads] [Bug Reports] [Discussion, Ideas & Q&A]

Note: InvokeAI is rapidly evolving. Please use the Issues tab to report bugs and make feature requests. Be sure to use the provided templates. They will help us diagnose issues faster.

Getting Started with InvokeAI

For full installation and upgrade instructions, please see: InvokeAI Installation Overview

  1. Go to the bottom of the Latest Release Page
  2. Download the .zip file for your OS (Windows/macOS/Linux).
  3. Unzip the file.
  4. If you are on Windows, double-click on the install.bat script. On macOS, open a Terminal window, drag the file install.sh from Finder into the Terminal, and press return. On Linux, run install.sh.
  5. Wait a while, until it is done.
  6. The folder where you ran the installer from will now be filled with lots of files. If you are on Windows, double-click on the invoke.bat file. On macOS, open a Terminal window, drag invoke.sh from the folder into the Terminal, and press return. On Linux, run invoke.sh
  7. Press 2 to open the "browser-based UI", press enter/return, wait a minute or two for Stable Diffusion to start up, then open your browser and go to http://localhost:9090.
  8. Type banana sushi in the box on the top left and click Invoke:

Table of Contents

  1. Installation
  2. Hardware Requirements
  3. Features
  4. Latest Changes
  5. Troubleshooting
  6. Contributing
  7. Contributors
  8. Support
  9. Further Reading

Installation

This fork is supported across Linux, Windows and Macintosh. Linux users can use either an Nvidia-based card (with CUDA support) or an AMD card (using the ROCm driver). For full installation and upgrade instructions, please see: InvokeAI Installation Overview

Hardware Requirements

InvokeAI is supported across Linux, Windows and macOS. Linux users can use either an Nvidia-based card (with CUDA support) or an AMD card (using the ROCm driver).

System

You wil need one of the following:

  • An NVIDIA-based graphics card with 4 GB or more VRAM memory.
  • An Apple computer with an M1 chip.

We do not recommend the GTX 1650 or 1660 series video cards. They are unable to run in half-precision mode and do not have sufficient VRAM to render 512x512 images.

Memory

  • At least 12 GB Main Memory RAM.

Disk

  • At least 12 GB of free disk space for the machine learning model, Python, and all its dependencies.

Note

If you have a Nvidia 10xx series card (e.g. the 1080ti), please run the dream script in full-precision mode as shown below.

Similarly, specify full-precision mode on Apple M1 hardware.

Precision is auto configured based on the device. If however you encounter errors like 'expected type Float but found Half' or 'not implemented for Half' you can try starting invoke.py with the --precision=float32 flag to your initialization command

(invokeai) ~/InvokeAI$ python scripts/invoke.py --precision=float32

Or by updating your InvokeAI configuration file with this argument.

Features

Major Features

Other Features

Latest Changes

For our latest changes, view our Release Notes

Troubleshooting

Please check out our Q&A to get solutions for common installation problems and other issues.

Contributing

Anyone who wishes to contribute to this project, whether documentation, features, bug fixes, code cleanup, testing, or code reviews, is very much encouraged to do so.

To join, just raise your hand on the InvokeAI Discord server (#dev-chat) or the GitHub discussion board.

If you are unfamiliar with how to contribute to GitHub projects, here is a Getting Started Guide. A full set of contribution guidelines, along with templates, are in progress, but for now the most important thing is to make your pull request against the "development" branch, and not against "main". This will help keep public breakage to a minimum and will allow you to propose more radical changes.

We hope you enjoy using our software as much as we enjoy creating it, and we hope that some of those of you who are reading this will elect to become part of our community.

Welcome to InvokeAI!

Contributors

This fork is a combined effort of various people from across the world. Check out the list of all these amazing people. We thank them for their time, hard work and effort.

Support

For support, please use this repository's GitHub Issues tracking service. Feel free to send me an email if you use and like the script.

Original portions of the software are Copyright (c) 2020 Lincoln D. Stein

Further Reading

Please see the original README for more information on this software and underlying algorithm, located in the file README-CompViz.md.