InvokeAI/invokeai/backend/model_manager/load
Lincoln Stein 532f82cb97
Optimize RAM to VRAM transfer (#6312)
* avoid copying model back from cuda to cpu

* handle models that don't have state dicts

* add assertions that models need a `device()` method

* do not rely on torch.nn.Module having the device() method

* apply all patches after model is on the execution device

* fix model patching in latents too

* log patched tokenizer

* closes #6375

---------

Co-authored-by: Lincoln Stein <lstein@gmail.com>
2024-05-24 17:06:09 +00:00
..
convert_cache [mm] Do not write diffuser model to disk when convert_cache set to zero (#6072) 2024-03-29 16:11:08 -04:00
model_cache Optimize RAM to VRAM transfer (#6312) 2024-05-24 17:06:09 +00:00
model_loaders chore: clean up merge conflicts 2024-04-03 20:28:00 +05:30
__init__.py chore: ruff 2024-03-01 10:42:33 +11:00
load_base.py [mm] Do not write diffuser model to disk when convert_cache set to zero (#6072) 2024-03-29 16:11:08 -04:00
load_default.py [util] Add generic torch device class (#6174) 2024-04-15 13:12:49 +00:00
memory_snapshot.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
model_loader_registry.py Experiment with using absolute paths within model management 2024-03-08 15:36:14 -05:00
model_util.py make model manager v2 ready for PR review 2024-03-01 10:42:33 +11:00
optimizations.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00