InvokeAI/invokeai/backend/model_manager
Lincoln Stein 3d6d89feb4
[mm] Do not write diffuser model to disk when convert_cache set to zero (#6072)
* pass model config to _load_model

* make conversion work again

* do not write diffusers to disk when convert_cache set to 0

* adding same model to cache twice is a no-op, not an assertion error

* fix issues identified by psychedelicious during pr review

* following conversion, avoid redundant read of cached submodels

* fix error introduced while merging

---------

Co-authored-by: Lincoln Stein <lstein@gmail.com>
2024-03-29 16:11:08 -04:00
..
load [mm] Do not write diffuser model to disk when convert_cache set to zero (#6072) 2024-03-29 16:11:08 -04:00
metadata fix(mm): config.json to indicates diffusers model 2024-03-13 21:02:29 +11:00
util install model if diffusers or single file, cleaned up backend logic to not mess with existing model install 2024-03-13 21:02:29 +11:00
__init__.py [mm] Do not write diffuser model to disk when convert_cache set to zero (#6072) 2024-03-29 16:11:08 -04:00
config.py fix(mm): default settings pydantic error 2024-03-25 07:40:52 -04:00
convert_ckpt_to_diffusers.py [mm] Do not write diffuser model to disk when convert_cache set to zero (#6072) 2024-03-29 16:11:08 -04:00
libc_util.py Tidy names and locations of modules 2024-03-01 10:42:33 +11:00
merge.py fix(config): use new get_config across the app, use correct settings 2024-03-19 09:24:28 +11:00
probe.py fix(mm): move depth variant config to sd2 2024-03-27 07:48:54 -04:00
search.py docs(mm): update ModelSearch 2024-03-28 12:35:41 +11:00
starter_models.py fix(mm): remove proteus model 2024-03-22 02:22:03 -07:00