InvokeAI/invokeai/backend
Lincoln Stein 532f82cb97
Optimize RAM to VRAM transfer (#6312)
* avoid copying model back from cuda to cpu

* handle models that don't have state dicts

* add assertions that models need a `device()` method

* do not rely on torch.nn.Module having the device() method

* apply all patches after model is on the execution device

* fix model patching in latents too

* log patched tokenizer

* closes #6375

---------

Co-authored-by: Lincoln Stein <lstein@gmail.com>
2024-05-24 17:06:09 +00:00
..
image_util feat(nodes): use new blur_if_nsfw method 2024-05-14 07:23:38 +10:00
ip_adapter Create a UNetAttentionPatcher for patching UNet models with CustomAttnProcessor2_0 modules. 2024-04-09 08:12:12 -04:00
model_hash feat(mm): rename "blake3" to "blake3_multi" 2024-03-22 08:26:36 +11:00
model_manager Optimize RAM to VRAM transfer (#6312) 2024-05-24 17:06:09 +00:00
onnx final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
stable_diffusion cleanup: seamless unused older code cleanup 2024-05-13 08:11:08 +10:00
tiles feat(nodes): extract LATENT_SCALE_FACTOR to constants.py 2024-03-01 10:42:33 +11:00
util Re-enable app shutdown actions (#6244) 2024-04-19 06:45:42 -04:00
__init__.py consolidate model manager parts into a single class 2024-03-01 10:42:33 +11:00
lora.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
model_patcher.py chore: ruff 2024-03-01 10:42:33 +11:00
raw_model.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
textual_inversion.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00