InvokeAI/docs/installation/040_INSTALL_DOCKER.md
2023-10-19 11:26:36 -04:00

5.3 KiB

title
Installing with Docker

:fontawesome-brands-docker: Docker

!!! warning "macOS and AMD GPU Users"

We highly recommend to Install InvokeAI locally using [these instructions](INSTALLATION.md),
because Docker containers can not access the GPU on macOS.

!!! warning "AMD GPU Users"

Container support for AMD GPUs has been reported to work by the community, but has not received
extensive testing. Please make sure to set the `GPU_DRIVER=rocm` environment variable (see below), and
use the `build.sh` script to build the image for this to take effect at build time.

!!! tip "Linux and Windows Users"

For optimal performance, configure your Docker daemon to access your machine's GPU.
Docker Desktop on Windows [includes GPU support](https://www.docker.com/blog/wsl-2-gpu-support-for-docker-desktop-on-nvidia-gpus/).
Linux users should install and configure the [NVIDIA Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html)

Why containers?

They provide a flexible, reliable way to build and deploy InvokeAI. See Processes under the Twelve-Factor App methodology for details on why running applications in such a stateless fashion is important.

The container is configured for CUDA by default, but can be built to support AMD GPUs by setting the GPU_DRIVER=rocm environment variable at Docker image build time.

Developers on Apple silicon (M1/M2): You can't access your GPU cores from Docker containers and performance is reduced compared with running it directly on macOS but for development purposes it's fine. Once you're done with development tasks on your laptop you can build for the target platform and architecture and deploy to another environment with NVIDIA GPUs on-premises or in the cloud.

TL;DR

This assumes properly configured Docker on Linux or Windows/WSL2. Read on for detailed customization options.

```bash
# docker compose commands should be run from the `docker` directory
cd docker
docker compose up
```

Installation in a Linux container (desktop)

Prerequisites

Install Docker

On the Docker Desktop app, go to Preferences, Resources, Advanced. Increase the CPUs and Memory to avoid this Issue. You may need to increase Swap and Disk image size too.

Get a Huggingface-Token

Besides the Docker Agent you will need an Account on huggingface.co.

After you succesfully registered your account, go to huggingface.co/settings/tokens, create a token and copy it, since you will need in for the next step.

Setup

Set up your environmnent variables. In the docker directory, make a copy of env.sample and name it .env. Make changes as necessary.

Any environment variables supported by InvokeAI can be set here - please see the CONFIGURATION for further detail.

At a minimum, you might want to set the INVOKEAI_ROOT environment variable to point to the location where you wish to store your InvokeAI models, configuration, and outputs.

Environment-Variable Default value Description
INVOKEAI_ROOT ~/invokeai Required - the location of your InvokeAI root directory. It will be created if it does not exist.
HUGGING_FACE_HUB_TOKEN InvokeAI will work without it, but some of the integrations with HuggingFace (like downloading from models from private repositories) may not work
GPU_DRIVER cuda Optionally change this to rocm to build the image for AMD GPUs. NOTE: Use the build.sh script to build the image for this to take effect.

Build the Image

Use the standard docker compose build command from within the docker directory.

If using an AMD GPU: a: set the GPU_DRIVER=rocm environment variable in docker-compose.yml and continue using docker compose build as usual, or b: set GPU_DRIVER=rocm in the .env file and use the build.sh script, provided for convenience

Run the Container

Use the standard docker compose up command, and generally the docker compose CLI as usual.

Once the container starts up (and configures the InvokeAI root directory if this is a new installation), you can access InvokeAI at http://localhost:9090