InvokeAI/invokeai/backend
Lincoln Stein 21a60af881
when unlocking models, offload_unlocked_models should prune to vram limit only (#6450)
Co-authored-by: Lincoln Stein <lstein@gmail.com>
2024-05-29 03:01:21 +00:00
..
image_util feat(nodes): use new blur_if_nsfw method 2024-05-14 07:23:38 +10:00
ip_adapter Create a UNetAttentionPatcher for patching UNet models with CustomAttnProcessor2_0 modules. 2024-04-09 08:12:12 -04:00
model_hash feat(mm): rename "blake3" to "blake3_multi" 2024-03-22 08:26:36 +11:00
model_manager when unlocking models, offload_unlocked_models should prune to vram limit only (#6450) 2024-05-29 03:01:21 +00:00
onnx final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
stable_diffusion cleanup: seamless unused older code cleanup 2024-05-13 08:11:08 +10:00
tiles feat(nodes): extract LATENT_SCALE_FACTOR to constants.py 2024-03-01 10:42:33 +11:00
util Re-enable app shutdown actions (#6244) 2024-04-19 06:45:42 -04:00
__init__.py consolidate model manager parts into a single class 2024-03-01 10:42:33 +11:00
lora.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
model_patcher.py chore: ruff 2024-03-01 10:42:33 +11:00
raw_model.py final tidying before marking PR as ready for review 2024-03-01 10:42:33 +11:00
textual_inversion.py Add a callout about the hackiness of dropping tokens in the TextualInversionManager. 2024-05-28 05:11:54 -07:00