InvokeAI/scripts
Lincoln Stein 0c8f0e3386 add max_load_models parameter for model cache control
- ldm.generate.Generator() now takes an argument named `max_load_models`.
  This is an integer that limits the model cache size. When the cache
  reaches the limit, it will start purging older models from cache.

- CLI takes an argument --max_load_models, default to 2. This will keep
  one model in GPU and the other in CPU and switch back and forth
  quickly.

- To not cache models at all, pass --max_load_models=1
2022-10-31 08:53:16 -04:00
..
orig_scripts rename all modules from ldm.dream to ldm.invoke 2022-10-08 11:37:23 -04:00
dream.py Reword deprecation warning for dream.py 2022-10-14 07:33:10 -04:00
images2prompt.py rename all modules from ldm.dream to ldm.invoke 2022-10-08 11:37:23 -04:00
invoke.py add max_load_models parameter for model cache control 2022-10-31 08:53:16 -04:00
legacy_api.py Add back old dream.py as legacy_api.py 2022-10-12 20:35:56 -04:00
merge_embeddings.py Fix merging embeddings (#226) 2022-08-31 11:24:11 -04:00
preload_models.py add --no-interactive mode 2022-10-30 00:33:48 -04:00
sd-metadata.py rename all modules from ldm.dream to ldm.invoke 2022-10-08 11:37:23 -04:00