InvokeAI/invokeai/backend
psychedelicious 2526ef52c5 fix(nodes): workaround seamless multi gpu error #6010
The seamless logic errors when a second GPU is selected. I don't understand why, but a workaround is to skip the model patching when there there are no seamless axes specified.

This is also just a good practice regardless - don't patch the model unless we need to. Probably a negligible perf impact.

Closes #6010
2024-03-29 08:56:38 +11:00
..
image_util tidy: "fit_image_to_resolution" -> "resize_image_to_resolution" 2024-03-21 07:02:57 -07:00
ip_adapter final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
model_hash feat(mm): rename "blake3" to "blake3_multi" 2024-03-22 08:26:36 +11:00
model_manager docs(mm): update ModelSearch 2024-03-28 12:35:41 +11:00
onnx final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
stable_diffusion fix(nodes): workaround seamless multi gpu error #6010 2024-03-29 08:56:38 +11:00
tiles feat(nodes): extract LATENT_SCALE_FACTOR to constants.py 2024-03-01 10:42:33 +11:00
util chore: ruff 2024-03-27 08:16:27 -04:00
__init__.py consolidate model manager parts into a single class 2024-03-01 10:42:33 +11:00
lora.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
model_patcher.py chore: ruff 2024-03-01 10:42:33 +11:00
raw_model.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
textual_inversion.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00