InvokeAI/invokeai/backend/model_manager/load
Lincoln Stein a3cb5da130
Improve RAM<->VRAM memory copy performance in LoRA patching and elsewhere (#6490)
* allow model patcher to optimize away the unpatching step when feasible

* remove lazy_offloading functionality

* allow model patcher to optimize away the unpatching step when feasible

* remove lazy_offloading functionality

* do not save original weights if there is a CPU copy of state dict

* Update invokeai/backend/model_manager/load/load_base.py

Co-authored-by: Ryan Dick <ryanjdick3@gmail.com>

* documentation fixes requested during penultimate review

* add non-blocking=True parameters to several torch.nn.Module.to() calls, for slight performance increases

* fix ruff errors

* prevent crash on non-cuda-enabled systems

---------

Co-authored-by: Lincoln Stein <lstein@gmail.com>
Co-authored-by: Kent Keirsey <31807370+hipsterusername@users.noreply.github.com>
Co-authored-by: Ryan Dick <ryanjdick3@gmail.com>
2024-06-13 17:10:03 +00:00
..
convert_cache make download and convert cache keys safe for filename length 2024-04-28 12:24:36 -04:00
model_cache Improve RAM<->VRAM memory copy performance in LoRA patching and elsewhere (#6490) 2024-06-13 17:10:03 +00:00
model_loaders add support for generic loading of diffusers directories 2024-06-07 13:54:30 +10:00
__init__.py add support for generic loading of diffusers directories 2024-06-07 13:54:30 +10:00
load_base.py Merge remote-tracking branch 'origin/main' into lstein/feat/simple-mm2-api 2024-06-07 14:23:41 +10:00
load_default.py make download and convert cache keys safe for filename length 2024-04-28 12:24:36 -04:00
memory_snapshot.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
model_loader_registry.py Experiment with using absolute paths within model management 2024-03-08 15:36:14 -05:00
model_util.py make model manager v2 ready for PR review 2024-03-01 10:42:33 +11:00
optimizations.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00