InvokeAI/docs/installation/INSTALL_DOCKER.md
2022-11-10 06:33:49 +01:00

9.5 KiB

title
Docker

:fontawesome-brands-docker: Docker

Before you begin

  • For end users: Install InvokeAI locally using the instructions for your OS.
  • For developers: For container-related development tasks or for enabling easy deployment to other environments (on-premises or cloud), follow these instructions. For general use, install locally to leverage your machine's GPU.

Why containers?

They provide a flexible, reliable way to build and deploy InvokeAI. You'll also use a Docker volume to store the largest model files and image outputs as a first step in decoupling storage and compute. Future enhancements can do this for other assets. See Processes under the Twelve-Factor App methodology for details on why running applications in such a stateless fashion is important.

You can specify the target platform when building the image and running the container. You'll also need to specify the InvokeAI requirements file that matches the container's OS and the architecture it will run on.

Developers on Apple silicon (M1/M2): You can't access your GPU cores from Docker containers and performance is reduced compared with running it directly on macOS but for development purposes it's fine. Once you're done with development tasks on your laptop you can build for the target platform and architecture and deploy to another environment with NVIDIA GPUs on-premises or in the cloud.

Installation on a Linux container

Prerequisites

Install Docker

On the Docker Desktop app, go to Preferences, Resources, Advanced. Increase the CPUs and Memory to avoid this Issue. You may need to increase Swap and Disk image size too.

Get a Huggingface-Token

Go to Hugging Face, create a token and temporary place it somewhere like a open texteditor window (but dont save it!, only keep it open, we need it in the next step)

Setup

Set the fork you want to use and other variables.

!!! tip

I preffer to save my env vars
in the repository root in a `.env` (or `.envrc`) file to automatically re-apply
them when I come back.

The build- and run- scripts contain default values for almost everything, besides the Hugging Face Token you created in the last step.

Some Suggestions of variables you may want to change besides the Token:

Environment-Variable Default value Description
HUGGINGFACE_TOKEN No default, but required! This is the only required variable, without you can't get the checkpoint
ARCH x86_64 if you are using a ARM based CPU
INVOKEAI_TAG invokeai-x86_64 the Container Repository / Tag which will be used
INVOKEAI_CONDA_ENV_FILE environment-lin-cuda.yml since environment.yml wouldn't work with aarch
INVOKEAI_GIT invoke-ai/InvokeAI the repository to use
INVOKEAI_BRANCH main the branch to checkout

Build the Image

I provided a build script, which is located in docker-build/build.sh but still needs to be executed from the Repository root.

./docker-build/build.sh

The build Script not only builds the container, but also creates the docker volume if not existing yet, or if empty it will just download the models.

Run the Container

After the build process is done, you can run the container via the provided docker-build/run.sh script

./docker-build/run.sh

When used without arguments, the container will start the website and provide you the link to open it. But if you want to use some other parameters you can also do so.

!!! example

```bash
docker-build/run.sh --from_file tests/validate_pr_prompt.txt
```

The output folder is located on the volume which is also used to store the model.

Find out more about available CLI-Parameter at [features/CLI.md](../features/CLI.md)

!!! warning "Deprecated"

From here on you will find the rest of the previous Docker-Docs, which will still
provide some usefull informations.

Usage (time to have fun)

Startup

If you're on a Linux container the invoke script is automatically started and the output dir set to the Docker volume you created earlier.

If you're directly on macOS follow these startup instructions.
With the Conda environment activated (conda activate ldm), run the interactive interface that combines the functionality of the original scripts txt2img and img2img:
Use the more accurate but VRAM-intensive full precision math because half-precision requires autocast and won't work.
By default the images are saved in outputs/img-samples/.

python3 scripts/invoke.py --full_precision

You'll get the script's prompt. You can see available options or quit.

invoke> -h
invoke> q

Text to Image

For quick (but bad) image results test with 5 steps (default 50) and 1 sample image. This will let you know that everything is set up correctly.
Then increase steps to 100 or more for good (but slower) results.
The prompt can be in quotes or not.

invoke> The hulk fighting with sheldon cooper -s5 -n1
invoke> "woman closeup highly detailed"  -s 150
# Reuse previous seed and apply face restoration
invoke> "woman closeup highly detailed"  --steps 150 --seed -1 -G 0.75

You'll need to experiment to see if face restoration is making it better or worse for your specific prompt.

If you're on a container the output is set to the Docker volume. You can copy it wherever you want.
You can download it from the Docker Desktop app, Volumes, my-vol, data.
Or you can copy it from your Mac terminal. Keep in mind docker cp can't expand *.png so you'll need to specify the image file name.

On your host Mac (you can use the name of any container that mounted the volume):

docker cp dummy:/data/000001.928403745.png /Users/<your-user>/Pictures

Image to Image

You can also do text-guided image-to-image translation. For example, turning a sketch into a detailed drawing.

strength is a value between 0.0 and 1.0 that controls the amount of noise that is added to the input image. Values that approach 1.0 allow for lots of variations but will also produce images that are not semantically consistent with the input. 0.0 preserves image exactly, 1.0 replaces it completely.

Make sure your input image size dimensions are multiples of 64 e.g. 512x512. Otherwise you'll get Error: product of dimension sizes > 2**31'. If you still get the error try a different size like 512x256.

If you're on a Docker container, copy your input image into the Docker volume

docker cp /Users/<your-user>/Pictures/sketch-mountains-input.jpg dummy:/data/

Try it out generating an image (or more). The invoke script needs absolute paths to find the image so don't use ~.

If you're on your Mac

invoke> "A fantasy landscape, trending on artstation" -I /Users/<your-user>/Pictures/sketch-mountains-input.jpg --strength 0.75  --steps 100 -n4

If you're on a Linux container on your Mac

invoke> "A fantasy landscape, trending on artstation" -I /data/sketch-mountains-input.jpg --strength 0.75  --steps 50 -n1

Web Interface

You can use the invoke script with a graphical web interface. Start the web server with:

python3 scripts/invoke.py --full_precision --web

If it's running on your Mac point your Mac web browser to http://127.0.0.1:9090

Press Control-C at the command line to stop the web server.

Notes

Some text you can add at the end of the prompt to make it very pretty:

cinematic photo, highly detailed, cinematic lighting, ultra-detailed, ultrarealistic, photorealism, Octane Rendering, cyberpunk lights, Hyper Detail, 8K, HD, Unreal Engine, V-Ray, full hd, cyberpunk, abstract, 3d octane render + 4k UHD + immense detail + dramatic lighting + well lit + black, purple, blue, pink, cerulean, teal, metallic colours, + fine details, ultra photoreal, photographic, concept art, cinematic composition, rule of thirds, mysterious, eerie, photorealism, breathtaking detailed, painting art deco pattern, by hsiao, ron cheng, john james audubon, bizarre compositions, exquisite detail, extremely moody lighting, painted by greg rutkowski makoto shinkai takashi takeuchi studio ghibli, akihiko yoshida

The original scripts should work as well.

python3 scripts/orig_scripts/txt2img.py --help
python3 scripts/orig_scripts/txt2img.py --ddim_steps 100 --n_iter 1 --n_samples 1  --plms --prompt "new born baby kitten. Hyper Detail, Octane Rendering, Unreal Engine, V-Ray"
python3 scripts/orig_scripts/txt2img.py --ddim_steps 5   --n_iter 1 --n_samples 1  --plms --prompt "ocean" # or --klms