Go to file
mauwii e2e86d2d11
fix INSTALL_DOCKER.md:
- fix title (Docker instead of "Before you begin")
- add headline with Emoji
- fix headlines to render toc correct
2022-10-11 00:32:04 +02:00
.dev_scripts Replace --full_precision with --precision that works even if not specified 2022-09-20 17:08:00 -04:00
.github rebrand CLI from "dream" to "invoke" 2022-10-08 09:32:06 -04:00
assets fix CompViz assets 2022-09-18 00:13:45 +02:00
backend Fixes CORS handling 2022-10-08 11:57:18 -04:00
configs Get for external TI .bin files to work 2022-10-08 13:18:19 -04:00
data stable diffusion 2022-08-10 16:30:49 +02:00
docker-build Upgrade GFPGAN to Version 1.4 2022-09-24 05:09:45 -04:00
docs fix INSTALL_DOCKER.md: 2022-10-11 00:32:04 +02:00
frontend rebuild front end 2022-10-10 09:19:52 -04:00
ldm run make_schedule() if it hasn't already been called 2022-10-09 18:21:13 -04:00
models add configs for training unconditional/class-conditional ldms 2021-12-22 15:57:23 +01:00
notebooks fix(venv): rename 'ldm' -> 'invokeai' 2022-10-10 13:04:03 -04:00
scripts rename all modules from ldm.dream to ldm.invoke 2022-10-08 11:37:23 -04:00
server rename all modules from ldm.dream to ldm.invoke 2022-10-08 11:37:23 -04:00
static * Changes for PR 2022-10-04 14:37:32 -04:00
tests Added linux to the workflows (#463) 2022-09-16 17:46:57 -04:00
.gitattributes Update .gitattributes 2022-08-29 16:58:41 -05:00
.gitignore multiple fixes in response to preflight testing bug reports 2022-10-05 12:44:16 -04:00
.gitmodules remove src directory, which is gumming up conda installs; addresses issue #77 2022-08-25 10:43:05 -04:00
.prettierrc.yaml change printWidth for markdown files to 80 2022-09-17 02:23:00 +02:00
environment-mac.yml fix(venv): rename 'ldm' -> 'invokeai' 2022-10-10 13:04:03 -04:00
environment.yml fix(venv): rename 'ldm' -> 'invokeai' 2022-10-10 13:04:03 -04:00
LICENSE added assertion checks for out-of-bound arguments; added various copyright and license agreement files 2022-08-24 09:22:27 -04:00
LICENSE-ModelWeights.txt added assertion checks for out-of-bound arguments; added various copyright and license agreement files 2022-08-24 09:22:27 -04:00
main.py Textual Inversion for M1 2022-09-27 01:39:17 +02:00
mkdocs.yml chng site_name to Stable Diffusion Toolkit Docs 2022-10-10 23:23:10 +02:00
pyproject.toml.hide Fix Mac Issue #723 2022-09-21 13:42:47 -04:00
README.md Update README.md 2022-10-10 12:36:50 -04:00
requirements-lin-AMD.txt TOIL(pip): Refactor pip requirements across the board (#531) 2022-09-14 07:01:58 -04:00
requirements-lin-win-colab-CUDA.txt TOIL(pip): Refactor pip requirements across the board (#531) 2022-09-14 07:01:58 -04:00
requirements-linux-arm64.txt multiple fixes in response to preflight testing bug reports 2022-10-05 12:44:16 -04:00
requirements-mac-MPS-CPU.txt add realesrgan to requirements.txt, remove nightie for torch and torchvision due to performance issues 2022-10-08 12:01:45 -04:00
requirements-mkdocs.txt update requirements-mkdocs.txt 2022-09-19 08:38:46 +02:00
requirements.txt add realesrgan to requirements.txt, remove nightie for torch and torchvision due to performance issues 2022-10-08 12:01:45 -04:00
setup.py release candidate for InvokeAI 2.0 2022-10-04 23:05:40 -04:00
Stable_Diffusion_v1_Model_Card.md stable diffusion 2022-08-10 16:30:49 +02:00

InvokeAI: A Stable Diffusion Toolkit

Formally known as lstein/stable-diffusion

project logo

discord badge

latest release badge github stars badge github forks badge

CI checks on main badge CI checks on dev badge latest commit to dev badge

github open issues badge github open prs badge

This is a fork of CompVis/stable-diffusion, the open source text-to-image generator. It provides a streamlined process with various new features and options to aid the image generation process. It runs on Windows, Mac and Linux machines, with GPU cards with as little as 4 GB of RAM. It provides both a polished Web interface (see below), and an easy-to-use command-line interface.

Quick links: [Discord Server] [Documentation and Tutorials] [Code and Downloads] [Bug Reports] [Discussion, Ideas & Q&A]

Note: This fork is rapidly evolving. Please use the Issues tab to report bugs and make feature requests. Be sure to use the provided templates. They will help aid diagnose issues faster.

Table of Contents

  1. Installation
  2. Hardware Requirements
  3. Features
  4. Latest Changes
  5. Troubleshooting
  6. Contributing
  7. Contributors
  8. Support
  9. Further Reading

Installation

This fork is supported across multiple platforms. You can find individual installation instructions below.

Hardware Requirements

System

You wil need one of the following:

  • An NVIDIA-based graphics card with 4 GB or more VRAM memory.
  • An Apple computer with an M1 chip.

Memory

  • At least 12 GB Main Memory RAM.

Disk

  • At least 6 GB of free disk space for the machine learning model, Python, and all its dependencies.

Note

If you have a Nvidia 10xx series card (e.g. the 1080ti), please run the dream script in full-precision mode as shown below.

Similarly, specify full-precision mode on Apple M1 hardware.

Precision is auto configured based on the device. If however you encounter errors like 'expected type Float but found Half' or 'not implemented for Half' you can try starting invoke.py with the --precision=float32 flag:

(ldm) ~/stable-diffusion$ python scripts/invoke.py --precision=float32

Features

Major Features

Other Features

Latest Changes

  • v2.0.0 (9 October 2022)

    • dream.py script renamed invoke.py. A dream.py script wrapper remains for backward compatibility.
    • Completely new WebGUI - launch with python3 scripts/invoke.py --web
    • Support for inpainting and outpainting
    • img2img runs on all k* samplers
    • Support for negative prompts
    • Support for CodeFormer face reconstruction
    • Support for Textual Inversion on Macintoshes
    • Support in both WebGUI and CLI for post-processing of previously-generated images using facial reconstruction, ESRGAN upscaling, outcropping (similar to DALL-E infinite canvas), and "embiggen" upscaling. See the !fix command.
    • New --hires option on invoke> line allows larger images to be created without duplicating elements, at the cost of some performance.
    • New --perlin and --threshold options allow you to add and control variation during image generation (see Thresholding and Perlin Noise Initialization
    • Extensive metadata now written into PNG files, allowing reliable regeneration of images and tweaking of previous settings.
    • Command-line completion in invoke.py now works on Windows, Linux and Mac platforms.
    • Improved command-line completion behavior. New commands added:
      • List command-line history with !history
      • Search command-line history with !search
      • Clear history with !clear
    • Deprecated --full_precision / -F. Simply omit it and invoke.py will auto configure. To switch away from auto use the new flag like --precision=float32.

For older changelogs, please visit the CHANGELOG.

Troubleshooting

Please check out our Q&A to get solutions for common installation problems and other issues.

Contributing

Anyone who wishes to contribute to this project, whether documentation, features, bug fixes, code cleanup, testing, or code reviews, is very much encouraged to do so. If you are unfamiliar with how to contribute to GitHub projects, here is a Getting Started Guide.

A full set of contribution guidelines, along with templates, are in progress, but for now the most important thing is to make your pull request against the "development" branch, and not against "main". This will help keep public breakage to a minimum and will allow you to propose more radical changes.

Contributors

This fork is a combined effort of various people from across the world. Check out the list of all these amazing people. We thank them for their time, hard work and effort.

Support

For support, please use this repository's GitHub Issues tracking service. Feel free to send me an email if you use and like the script.

Original portions of the software are Copyright (c) 2020 Lincoln D. Stein

Further Reading

Please see the original README for more information on this software and underlying algorithm, located in the file README-CompViz.md.