InvokeAI/docker
2024-07-26 13:57:33 -04:00
..
.env.sample fix(docker): change 'nvidia' profile name to 'cuda' 2024-07-09 09:47:29 -04:00
docker-compose.yml fix(docker): change 'nvidia' profile name to 'cuda' 2024-07-09 09:47:29 -04:00
docker-entrypoint.sh fix(docker): ensure 'chown' does not break on read-only fs; fixes #6264 2024-07-09 09:47:29 -04:00
Dockerfile Update docker image to use pnpm version 8 2024-07-26 13:57:33 -04:00
README.md docs: overhaul Docker documentation, add to main README 2024-07-09 09:47:29 -04:00
run.sh fix(docker): change 'nvidia' profile name to 'cuda' 2024-07-09 09:47:29 -04:00
runpod-readme.md (docker) add README used by the Runpod template 2023-07-12 16:51:15 -04:00

Invoke in Docker

  • Ensure that Docker can use the GPU on your system
  • This documentation assumes Linux, but should work similarly under Windows with WSL2
  • We don't recommend running Invoke in Docker on macOS at this time. It works, but very slowly.

Quickstart :lightning:

No docker compose, no persistence, just a simple one-liner using the official images:

CUDA:

docker run --runtime=nvidia --gpus=all --publish 9090:9090 ghcr.io/invoke-ai/invokeai

ROCm:

docker run --device /dev/kfd --device /dev/dri --publish 9090:9090 ghcr.io/invoke-ai/invokeai:main-rocm

Open http://localhost:9090 in your browser once the container finishes booting, install some models, and generate away!

Tip

To persist your data (including downloaded models) outside of the container, add a --volume/-v flag to the above command, e.g.: docker run --volume /some/local/path:/invokeai <...the rest of the command>

Customize the container

We ship the run.sh script, which is a convenient wrapper around docker compose for cases where custom image build args are needed. Alternatively, the familiar docker compose commands work just as well.

cd docker
cp .env.sample .env
# edit .env to your liking if you need to; it is well commented.
./run.sh

It will take a few minutes to build the image the first time. Once the application starts up, open http://localhost:9090 in your browser to invoke!

Docker setup in detail

Linux

  1. Ensure builkit is enabled in the Docker daemon settings (/etc/docker/daemon.json)
  2. Install the docker compose plugin using your package manager, or follow a tutorial.
    • The deprecated docker-compose (hyphenated) CLI probably won't work. Update to a recent version.
  3. Ensure docker daemon is able to access the GPU.

macOS

Tip

You'll be better off installing Invoke directly on your system, because Docker can not use the GPU on macOS.

If you are still reading:

  1. Ensure Docker has at least 16GB RAM
  2. Enable VirtioFS for file sharing
  3. Enable docker compose V2 support

This is done via Docker Desktop preferences.

Configure the Invoke Environment

  1. Make a copy of .env.sample and name it .env (cp .env.sample .env (Mac/Linux) or copy example.env .env (Windows)). Make changes as necessary. Set INVOKEAI_ROOT to an absolute path to the desired location of the InvokeAI runtime directory. It may be an existing directory from a previous installation (post 4.0.0).
  2. Execute run.sh

The image will be built automatically if needed.

The runtime directory (holding models and outputs) will be created in the location specified by INVOKEAI_ROOT. The default location is ~/invokeai. Navigate to the Model Manager tab and install some models before generating.

Use a GPU

  • Linux is recommended for GPU support in Docker.
  • WSL2 is required for Windows.
  • only x86_64 architecture is supported.

The Docker daemon on the system must be already set up to use the GPU. In case of Linux, this involves installing nvidia-docker-runtime and configuring the nvidia runtime as default. Steps will be different for AMD. Please see Docker/NVIDIA/AMD documentation for the most up-to-date instructions for using your GPU with Docker.

To use an AMD GPU, set GPU_DRIVER=rocm in your .env file before running ./run.sh.

Customize

Check the .env.sample file. It contains some environment variables for running in Docker. Copy it, name it .env, and fill it in with your own values. Next time you run run.sh, your custom values will be used.

You can also set these values in docker-compose.yml directly, but .env will help avoid conflicts when code is updated.

Values are optional, but setting INVOKEAI_ROOT is highly recommended. The default is ~/invokeai. Example:

INVOKEAI_ROOT=/Volumes/WorkDrive/invokeai
HUGGINGFACE_TOKEN=the_actual_token
CONTAINER_UID=1000
GPU_DRIVER=cuda

Any environment variables supported by InvokeAI can be set here. See the Configuration docs for further detail.

Even More Customizing!

See the docker-compose.yml file. The command instruction can be uncommented and used to run arbitrary startup commands. Some examples below.

Reconfigure the runtime directory

Can be used to download additional models from the supported model list

In conjunction with INVOKEAI_ROOT can be also used to initialize a runtime directory

command:
  - invokeai-configure
  - --yes

Or install models:

command:
  - invokeai-model-install