diff --git a/docker-compose.yml b/docker-compose.yml index 550b507..0cacffd 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -28,7 +28,7 @@ services: <<: *base_service profiles: ["auto"] build: ./services/AUTOMATIC1111 - image: sd-auto:44 + image: sd-auto:45 environment: - CLI_ARGS=--allow-code --medvram --xformers --enable-insecure-extension-access --api @@ -43,7 +43,7 @@ services: <<: *base_service profiles: ["invoke"] build: ./services/invoke/ - image: sd-invoke:22 + image: sd-invoke:23 environment: - PRELOAD=true - CLI_ARGS= diff --git a/services/AUTOMATIC1111/Dockerfile b/services/AUTOMATIC1111/Dockerfile index 632ebb2..317e407 100644 --- a/services/AUTOMATIC1111/Dockerfile +++ b/services/AUTOMATIC1111/Dockerfile @@ -26,7 +26,7 @@ RUN . /clone.sh clip-interrogator https://github.com/pharmapsychotic/clip-interr FROM alpine:3.17 as xformers RUN apk add --no-cache aria2 -RUN aria2c -x 5 --dir / --out wheel.whl 'https://github.com/AbdBarho/stable-diffusion-webui-docker/releases/download/4.1.0/xformers-0.0.16.dev421-cp310-cp310-manylinux2014_x86_64.whl' +RUN aria2c -x 5 --dir / --out wheel.whl 'https://github.com/AbdBarho/stable-diffusion-webui-docker/releases/download/5.0.0/xformers-0.0.17.dev449-cp310-cp310-manylinux2014_x86_64.whl' FROM python:3.10.9-slim @@ -67,7 +67,7 @@ RUN --mount=type=cache,target=/root/.cache/pip \ # Note: don't update the sha of previous versions because the install will take forever # instead, update the repo state in a later step -ARG SHA=3993aa43e95bb60b9df977946af988f119525a71 +ARG SHA=3715ece0adce7bf7c5e9c5ab3710b2fdc3848f39 RUN --mount=type=cache,target=/root/.cache/pip < req.txt -pip install -r req.txt -rm req.txt +pip install . +# egrep -v '^-e .' environments-and-requirements/requirements-lin-cuda.txt > req.txt +# pip install -r req.txt +# rm req.txt EOF @@ -62,9 +63,10 @@ RUN --mount=type=cache,target=/root/.cache/pip \ RUN touch invokeai.init COPY . /docker/ +# mkdir configs && cp invokeai/configs/INITIAL_MODELS.yaml configs/models.yaml -ENV ROOT=/stable-diffusion PYTHONPATH="${PYTHONPATH}:${ROOT}" PRELOAD=false CLI_ARGS="" HF_HOME=/root/.cache/huggingface +ENV PYTHONUNBUFFERED=1 ROOT=/stable-diffusion PYTHONPATH="${PYTHONPATH}:${ROOT}" PRELOAD=false CLI_ARGS="" HF_HOME=/root/.cache/huggingface EXPOSE 7860 ENTRYPOINT ["/docker/entrypoint.sh"] -CMD python3 -u scripts/invoke.py --web --host 0.0.0.0 --port 7860 --config /docker/models.yaml --root_dir ${ROOT} --outdir /output/invoke ${CLI_ARGS} +CMD invokeai --web --host 0.0.0.0 --port 7860 --config /docker/models.yaml --root_dir ${ROOT} --outdir /output/invoke ${CLI_ARGS} diff --git a/services/invoke/entrypoint.sh b/services/invoke/entrypoint.sh index a8209bb..83d41a1 100755 --- a/services/invoke/entrypoint.sh +++ b/services/invoke/entrypoint.sh @@ -38,9 +38,9 @@ for to_path in "${!MOUNTS[@]}"; do echo Mounted $(basename "${from_path}") done -if "${PRELOAD}" == "true"; then - set -Eeuo pipefail - python3 -u scripts/preload_models.py --skip-sd-weights --root ${ROOT} --config_file /docker/models.yaml -fi +# if "${PRELOAD}" == "true"; then +# set -Eeuo pipefail +# python3 -u scripts/preload_models.py --skip-sd-weights --root ${ROOT} --config_file /docker/models.yaml +# fi exec "$@" diff --git a/services/invoke/models.yaml b/services/invoke/models.yaml index 727f8e6..c631164 100644 --- a/services/invoke/models.yaml +++ b/services/invoke/models.yaml @@ -9,7 +9,7 @@ stable-diffusion-1.5: description: Stable Diffusion version 1.5 weights: /data/StableDiffusion/v1-5-pruned-emaonly.ckpt vae: /data/VAE/vae-ft-mse-840000-ema-pruned.ckpt - config: ./configs/stable-diffusion/v1-inference.yaml + config: /stable-diffusion/invokeai/configs/stable-diffusion/v1-inference.yaml width: 512 height: 512 default: true @@ -17,7 +17,7 @@ inpainting-1.5: description: RunwayML SD 1.5 model optimized for inpainting weights: /data/StableDiffusion/sd-v1-5-inpainting.ckpt vae: /data/VAE/vae-ft-mse-840000-ema-pruned.ckpt - config: ./configs/stable-diffusion/v1-inpainting-inference.yaml + config: /stable-diffusion/invokeai/configs/stable-diffusion/v1-inpainting-inference.yaml width: 512 height: 512 default: false