Sergey Borisov
2e7fc055c4
Support both pre and post 4.31.0 transformers
2023-07-19 06:15:17 +03:00
Brandon Rising
f4e52fafac
Fix as part of merging main in
2023-07-18 23:05:33 -04:00
Lincoln Stein
0f7e329e76
restore access token-saving code
2023-07-18 22:58:56 -04:00
Brandon Rising
ee7b36cea5
Merge branch 'main' into onnx-testing
2023-07-18 22:56:41 -04:00
Lincoln Stein
a690cca5b5
make convert work with both 4.30.2 and 4.31.0
2023-07-18 22:18:13 -04:00
Brandon Rising
e201ad2f51
Switch to io_binding for run, testing different session options
2023-07-18 21:54:54 -04:00
Sergey Borisov
0aa7193d3b
Load text_model.embeddings.position_ids outsude state_dict
2023-07-19 04:18:43 +03:00
Lincoln Stein
2fbf245c3d
Merge branch 'main' into release/invokeai-3-0-beta
...
-- this adds the upscaling support
2023-07-18 21:17:15 -04:00
Lincoln Stein
39c14eb2ac
fix pretrained model download to work with xl
2023-07-18 21:10:33 -04:00
Lincoln Stein
893e199677
Merge branch 'main' into feat/ui/upscale
2023-07-18 19:18:55 -04:00
blessedcoolant
186e98da5e
Merge branch 'main' into fix/mem_cleanup
2023-07-19 10:10:32 +12:00
Eugene Brodsky
dea9a5da7a
Avoid crash if unable to modify the model config file ( #3824 )
...
* fix whitespace; remove invisible characters
* log error and proceed if unable to modify the model config
2023-07-18 16:33:19 -04:00
Sergey Borisov
bda0000acd
Cleanup vram after models offloading, tweak to cleanup local variable references on ram offload
2023-07-18 23:21:18 +03:00
Brandon Rising
35d5ef9118
Emit step completions
2023-07-18 12:35:07 -04:00
StAlKeR7779
889b77d3d6
Merge branch 'main' into save_vram
2023-07-18 16:55:48 +03:00
Sergey Borisov
bc11296a5e
Disable lazy offloading on disabled vram cache, move resulted tensors to cpu(to not stack vram tensors in cache), fix - text encoder not freed(detach)
2023-07-18 16:20:25 +03:00
psychedelicious
42c440c73f
Merge branch 'main' into feat/ui/upscale
2023-07-18 22:08:02 +10:00
Lincoln Stein
9c3c393b84
merge with main
2023-07-18 07:00:55 -04:00
psychedelicious
56098f370c
feat(nodes): add RealESRGAN_x2plus.pth
, update upscale nodes
...
- add `RealESRGAN_x2plus.pth` model to installer
- add `RealESRGAN_x2plus.pth` to `realesrgan` node
- rename `RealESRGAN` to `ESRGAN` in nodes
- make `scale_factor` optional in `img_scale` node
2023-07-18 14:55:18 +10:00
Lincoln Stein
1353bf98b3
add specific exception for model probe failures
2023-07-17 23:08:39 -04:00
Lincoln Stein
025cda3815
fix 424 error on model import
2023-07-17 22:21:11 -04:00
blessedcoolant
13da881953
Merge branch 'main' into sdxl-support
2023-07-18 13:34:07 +12:00
blessedcoolant
ec3c15ead0
Merge branch 'main' into mm-ui
2023-07-18 12:58:57 +12:00
Sergey Borisov
0fce35c54c
Cleanup, fix variable name, fix controlnet for sequential and cross attention guidance
2023-07-17 23:53:50 +03:00
Brandon Rising
bcce70fca6
Testing different session opts, added timings for testing
2023-07-17 16:27:33 -04:00
Sergey Borisov
1c680a7147
Fix - encoder_attention_mask not passed before to unet, even if passed it will broke sequential guidance run, so rewrite logic
2023-07-17 23:13:37 +03:00
Lincoln Stein
0ea8d3c30c
prevent crash on rename operation on models in models directory
2023-07-17 07:50:06 -04:00
Lincoln Stein
84a13ff8e1
Merge branch 'mm-ui' of github.com:blessedcoolant/InvokeAI into mm-ui
2023-07-17 07:29:35 -04:00
Lincoln Stein
3fba262c94
expose paths as absolute to web api
2023-07-17 07:29:26 -04:00
Lincoln Stein
107ca6bf47
expose model paths as absolute to web models API
2023-07-17 07:26:05 -04:00
blessedcoolant
cbfd1d1b27
Merge branch 'main' into feat/standalone_diffusers_ti
2023-07-17 22:01:52 +12:00
blessedcoolant
aebd595607
Merge branch 'main' into mm-ui
2023-07-17 13:49:25 +12:00
Lincoln Stein
ed88e72412
correct cannot assign to field 'unconditioned_embeddings' error
2023-07-16 21:06:40 -04:00
Sergey Borisov
6aefd8600a
Fix error with long prompts when controlnet used
2023-07-16 21:06:40 -04:00
Lincoln Stein
ccb43d5a91
make check for 2.3 root directory more stringent
2023-07-16 20:43:15 -04:00
Kent Keirsey
675a92401c
Merge branch 'main' into lstein/default-model-install
2023-07-16 19:32:59 -04:00
Sergey Borisov
b61c83e836
Allow bin extension to detect diffusers-ti provided as file
2023-07-17 00:32:17 +03:00
Lincoln Stein
2bc3e36bc0
add missing exception name
2023-07-16 16:14:28 -04:00
Lincoln Stein
6fbb5ce780
add renaming capabilities to model update API route
2023-07-16 14:17:05 -04:00
Brandon Rising
932112b640
testing being super wasteful with data
2023-07-16 00:17:33 -04:00
Lincoln Stein
b767b5d44c
user must adjust terminal size on Windows
2023-07-15 23:19:50 -04:00
Lincoln Stein
e95cb3aa71
Merge branch 'lstein/default-model-install' into release/invokeai-3-0-beta
2023-07-15 20:16:51 -04:00
Lincoln Stein
5b5d5ec978
Merge branch 'main' into sdxl-support
2023-07-15 19:49:57 -04:00
Lincoln Stein
ccbfa5d862
resolve conflicts
2023-07-15 19:47:50 -04:00
Lincoln Stein
7fa394912d
Merge branch 'main' into lstein/default-model-install
2023-07-15 18:26:35 -04:00
Lincoln Stein
6b0a158ffa
Merge branch 'main' into lstein/default-model-install
2023-07-15 18:23:34 -04:00
Lincoln Stein
c90345d6a3
deprecate the face restoration option
2023-07-15 18:23:32 -04:00
Lincoln Stein
f66ead0819
Merge branch 'main' into update-textual-inversion-training
2023-07-15 17:44:45 -04:00
Kent Keirsey
77b0129b4c
Merge branch 'main' into lstein/migrate-fix
2023-07-15 10:37:56 -04:00
Lincoln Stein
e01706f5f5
add fp16 support to controlnet models
2023-07-15 10:37:11 -04:00
Lincoln Stein
a111539059
migrate script now initializes destination root if needed
2023-07-15 09:59:34 -04:00
Lincoln Stein
32e7e52d69
Merge branch 'main' into lstein/default-model-install
2023-07-15 08:30:22 -04:00
psychedelicious
d234bf1cb9
feat(install): display full ESRGAN model filenames during installation
2023-07-15 21:27:10 +10:00
blessedcoolant
808b2de709
Merge branch 'main' into lstein/model-manager-route-enhancements
2023-07-15 16:56:54 +12:00
Lincoln Stein
2faa7cee37
add rename_model route
2023-07-14 23:03:18 -04:00
Brandon
467414f214
Merge branch 'main' into update-textual-inversion-training
2023-07-14 22:32:09 -04:00
Brandon Rising
f88a338be0
Setup textual inversion training with new model manager
2023-07-14 21:58:51 -04:00
Lincoln Stein
b306247eb5
remove clipseg model install
2023-07-14 20:39:42 -04:00
Lincoln Stein
a45f7ce355
add --list-models command
2023-07-14 19:52:47 -04:00
Lincoln Stein
eb9d74653d
set default models for realesrgan, controlnet and text inversion
2023-07-14 19:03:41 -04:00
Sergey Borisov
7093e5d033
Pad conditionings using zeros and encoder_attention_mask
2023-07-15 00:52:54 +03:00
Brandon Rising
bd7b59910d
Testing onnx in new ui updates
2023-07-14 14:24:15 -04:00
Lincoln Stein
e71ce83e9c
Merge branch 'main' into lstein/model-manager-route-enhancements
2023-07-14 13:52:55 -04:00
Lincoln Stein
8600aad12b
multiple enhancements to model manager REACT API
...
1. add a /sync route for synchronizing the in-memory model lists to
models.yaml, the models directory, and the autoimport directories.
2. add optional destination_directories to convert_model and merge_model
operations.
3. add /ckpt_confs route for retrieving known legacy checkpoint configuration
files.
4. add /search route for finding all models in a directory located in the server
filesystem
2023-07-14 13:45:16 -04:00
Lincoln Stein
ad076b1174
add model directory search route
2023-07-14 11:14:33 -04:00
blessedcoolant
16e93c6455
Merge branch 'main' into mm-ui
2023-07-14 15:46:53 +12:00
Brandon Rising
524888bf3b
Merge branch 'main' into feat/onnx
2023-07-13 14:23:57 -04:00
blessedcoolant
9348dc8e0d
Merge branch 'main' into fix/controlnet_cfg_inj_cond
2023-07-14 01:11:04 +12:00
blessedcoolant
5a546e66f1
Merge branch 'main' into fix-inpainting
2023-07-13 20:42:13 +12:00
psychedelicious
eb0d55263b
fix(mm): make model config attribute names consistent
...
Our model fields use `model_name`, but the API response uses `name`. Some places use `model_type` but the API response used `type`.
Changed the API response to provide `model_name` and `model_type`, which simplifies how we manage models on the client substantially.
2023-07-13 15:40:05 +10:00
blessedcoolant
7e3b9f1320
fix: Inpaint not working with some schedulers
...
Co-Authored-By: StAlKeR7779 <7768370+StAlKeR7779@users.noreply.github.com>
2023-07-13 15:06:03 +12:00
blessedcoolant
71e34ac256
Merge branch 'main' into mm-ui
2023-07-13 12:48:43 +12:00
Sergey Borisov
67c8cf4bc2
Controlnet model detection
2023-07-12 08:50:19 -04:00
Sergey Borisov
a328986b43
Less naive model detection
2023-07-12 08:50:19 -04:00
blessedcoolant
5a6ad99d4e
feat: Restore Delete Model Functionality
2023-07-12 16:39:07 +12:00
StAlKeR7779
b8a9b499df
Merge branch 'main' into fix/controlnet_cfg_inj_cond
2023-07-11 23:43:47 +03:00
Lincoln Stein
25591788c1
fix conflicts
2023-07-11 15:55:10 -04:00
Lincoln Stein
dab03fb646
rename gpu_mem_reserved to max_vram_cache_size
...
To be consistent with max_cache_size, the amount of memory to hold in
VRAM for model caching is now controlled by the max_vram_cache_size
configuration parameter.
2023-07-11 15:25:39 -04:00
Lincoln Stein
d32f9f7cb0
reverse logic of gpu_mem_reserved
...
- gpu_mem_reserved now indicates the amount of VRAM that will be reserved
for model caching (similar to max_cache_size).
2023-07-11 15:16:40 -04:00
Lincoln Stein
6b93c1451f
do not crash when probing an unknown model type
2023-07-11 10:56:47 -04:00
Lincoln Stein
bf2b5b5cd4
improvements to sdxl support in model manager
...
- Move SDXL-related models to models/sdxl.py
- Create separate base type BaseModelType.StableDiffusionXLRefiner for the refiner
models.
2023-07-09 20:42:03 -04:00
Lincoln Stein
5759a390f9
introduce gpu_mem_reserved configuration parameter
2023-07-09 18:35:04 -04:00
Lincoln Stein
130249a2dd
add model loading support for SDXL
2023-07-09 15:47:06 -04:00
Lincoln Stein
8d7dba937d
fix undefined variable
2023-07-09 14:37:45 -04:00
Lincoln Stein
d6cb0e54b3
don't unload models from GPU until the space is needed
2023-07-09 14:26:30 -04:00
Lincoln Stein
5f7435955e
if models.yaml doesn't exist, rebuild it
2023-07-08 15:13:51 -04:00
Lincoln Stein
69ef1e1e56
speculative change to upgrade script
2023-07-08 11:45:26 -04:00
Lincoln Stein
10d3bccf32
Mac MPS FP16 fixes ( #3641 )
...
This PR is to allow FP16 precision to work on Macs with MPS. In
addition, it centralizes the torch fixes/workarounds required for MPS
into a new backend utility `mps_fixes.py`. This is conditionally
imported in `api_app.py`/`cli_app.py`.
Many MANY thanks to @StAlKeR7779 for patiently working to debug and fix
these issues.
2023-07-07 17:43:23 -04:00
Eugene Brodsky
97b2ec58e2
Merge branch 'main' into release/invokeai-3-0-alpha
2023-07-07 14:18:12 -04:00
Lincoln Stein
56f4712814
fix checkpoint VAE handling in migrate script
2023-07-07 09:34:42 -04:00
Lincoln Stein
9f58ed35cf
improve user migration experience
...
- No longer fail root directory probing if invokeai.yaml is missing
(test is now whether a `models/core` directory exists).
- Migrate script does not overwrite previously-installed models.
- Can run migrate script on an existing 2.3 version directory
with --from and --to pointing to same 2.3 root.
2023-07-07 08:18:46 -04:00
blessedcoolant
7aa918677e
Merge branch 'main' into feat/clip_skip
2023-07-07 16:21:53 +12:00
Lincoln Stein
54f3686e3b
merge with main, fix conflicts
2023-07-06 15:21:45 -04:00
Lincoln Stein
e9352227f3
add merge api
2023-07-06 15:12:34 -04:00
blessedcoolant
bc5371eeee
Merge branch 'main' into feat/clip_skip
2023-07-07 06:03:39 +12:00
Lincoln Stein
e573a533ae
remove redundant import
2023-07-06 13:24:58 -04:00
Lincoln Stein
581be42c75
Merge branch 'main' into lstein/model-manager-router-api
2023-07-06 13:20:36 -04:00
Lincoln Stein
90c66aab3d
merge with upstream
2023-07-06 13:17:02 -04:00
Lincoln Stein
3e925fbf34
model merging API ready for testing
2023-07-06 13:15:15 -04:00
Lincoln Stein
ec7c2f07c6
model merge backend, CLI and TUI working
2023-07-06 12:21:42 -04:00
blessedcoolant
b229fe19aa
Merge branch 'main' into lstein/configure-max-cache-size
2023-07-07 01:52:12 +12:00
Sergey Borisov
04b57c408f
Add clip skip option to prompt node
2023-07-06 16:09:40 +03:00
blessedcoolant
6f1268e2b1
Merge branch 'main' into lstein/more-model-loading-fixes
2023-07-07 00:32:22 +12:00
Lincoln Stein
8f5fcb188c
Merge branch 'main' into lstein/model-manager-router-api
2023-07-05 23:16:43 -04:00
Lincoln Stein
f7daa6e71d
all methods now return OPENAPI_MODEL_CONFIGS; convert uses PUT
2023-07-05 23:13:01 -04:00
Lincoln Stein
3691b55565
fix autoimport crash
2023-07-05 21:53:08 -04:00
Lincoln Stein
f610045a14
Merge branch 'main' into mps-fp16-fixes
2023-07-05 21:01:48 -04:00
Lincoln Stein
a7cbcae176
expose max_cache_size to invokeai-configure interface
2023-07-05 20:59:57 -04:00
Lincoln Stein
0a6dccd607
expose max_cache_size to invokeai-configure interface
2023-07-05 20:59:14 -04:00
Lincoln Stein
43c51ff157
Merge branch 'main' into lstein/more-model-loading-fixes
2023-07-05 20:48:15 -04:00
Lincoln Stein
cfa3b2419c
partial implementation of merge
2023-07-05 20:25:47 -04:00
Lincoln Stein
d4550b3059
clean up lint errors in lora.py
2023-07-05 19:18:25 -04:00
Lincoln Stein
83d3a043da
merge latest changes from main
2023-07-05 19:15:53 -04:00
gogurtenjoyer
169ff6368b
Update mps_fixes.py - additional torch op for nodes
...
This fixes scaling in the nodes UI.
2023-07-05 17:47:23 -04:00
Lincoln Stein
71dad6d404
Merge branch 'main' into ti-ui
2023-07-05 16:57:31 -04:00
Lincoln Stein
685a47cc7d
fix crash during lora application
2023-07-05 16:40:47 -04:00
Lincoln Stein
cb947bcbf0
Merge branch 'main' into lstein/fix-migrate3-textencoder
2023-07-05 16:23:00 -04:00
Lincoln Stein
f8bbec8572
Recognize and load diffusers-style LoRAs (.bin)
...
Prevent double-reporting of autoimported models
- closes #3636
Allow autoimport of diffusers-style LoRA models
- closes #3637
2023-07-05 16:21:23 -04:00
Lincoln Stein
863336acbb
Recognize and load diffusers-style LoRAs (.bin)
...
Prevent double-reporting of autoimported models
- closes #3636
Allow autoimport of diffusers-style LoRA models
- closes #3637
2023-07-05 16:19:16 -04:00
Lincoln Stein
90ae8ce26a
prevent model install crash "torch needs to be restarted with spawn"
2023-07-05 16:18:20 -04:00
Lincoln Stein
ad5d90aca8
prevent model install crash "torch needs to be restarted with spawn"
2023-07-05 15:38:07 -04:00
Lincoln Stein
5b6dd47b9f
add API for model convert
2023-07-05 15:13:21 -04:00
Lincoln Stein
5027d0a603
accept @psychedelicious suggestions above
2023-07-05 14:50:57 -04:00
Lincoln Stein
9f9ce08e44
Merge branch 'main' into lstein/remove-hardcoded-cuda-device
2023-07-05 13:38:33 -04:00
Lincoln Stein
021e1eca8e
Merge branch 'main' into mps-fp16-fixes
2023-07-05 13:19:52 -04:00
Lincoln Stein
5fe722900d
allow clip-vit-large-patch14 text encoder to coexist with tokenizer in same directory
2023-07-05 13:15:08 -04:00
Lincoln Stein
cf173b522b
allow clip-vit-large-patch14 text encoder to coexist with tokenizer in same directory
2023-07-05 13:14:41 -04:00
blessedcoolant
9e2d63ef97
Merge branch 'main' into fix/ckpt_convert_scan
2023-07-06 05:01:34 +12:00
Sergey Borisov
0ac9dca926
Fix loading diffusers ti
2023-07-05 19:46:00 +03:00
Lincoln Stein
bd82c4ace0
model installer confirms deletion of models
2023-07-05 09:57:23 -04:00
Lincoln Stein
9edf78dd2e
merge with main
2023-07-05 09:12:54 -04:00
Lincoln Stein
6112197edf
convert implemented; need router
2023-07-05 09:05:05 -04:00
gogurtenjoyer
ba7345deb4
Merge branch 'main' into mps-fp16-fixes
2023-07-05 07:38:41 -04:00
Sergey Borisov
ee042ab76d
Fix ckpt scanning on conversion
2023-07-05 14:18:30 +03:00
blessedcoolant
780e77d2ae
Merge branch 'main' into fix/clip_path
2023-07-05 22:45:52 +12:00
Sergey Borisov
e3fc1b3816
Fix clip path in migrate script
2023-07-05 13:43:09 +03:00
Lincoln Stein
307a01d604
when migrating models, changes / to _ in model names to avoid breaking model name keys
2023-07-05 20:27:03 +10:00
Sergey Borisov
2beb8f049e
Fix model detection
2023-07-05 09:43:46 +03:00
blessedcoolant
639d88afd6
revert: inference_mode to no_grad
2023-07-05 16:39:15 +12:00
blessedcoolant
c0501ed5c2
fix: Slow loading of Loras
...
Co-Authored-By: StAlKeR7779 <7768370+StAlKeR7779@users.noreply.github.com>
2023-07-05 12:47:34 +10:00
gogurtenjoyer
233869b56a
Mac MPS FP16 fixes
...
This PR is to allow FP16 precision to work on Macs with MPS. In addition, it centralizes the torch fixes/workarounds
required for MPS into a new backend utility file `mps_fixes.py`. This is conditionally imported in `api_app.py`/`cli_app.py`.
Many MANY thanks to StAlKeR7779 for patiently working to debug and fix these issues.
2023-07-04 18:10:53 -04:00
Lincoln Stein
5d099f4a49
update_model working
2023-07-04 17:26:57 -04:00
Lincoln Stein
752b4d50cf
model_delete method now working
2023-07-04 10:40:32 -04:00
Lincoln Stein
c1c49d9a76
import model returns 404 for invalid path, 409 for duplicate model
2023-07-04 10:08:10 -04:00
Lincoln Stein
96bf92ead4
add the import model router
2023-07-04 14:35:47 +10:00
Lincoln Stein
fc419546bc
Merge branch 'main' into lstein/remove-hardcoded-cuda-device
2023-07-03 14:10:47 -04:00
Lincoln Stein
cfd09214d3
Merge branch 'main' into lstein/fix-vae-conversion-crash
2023-07-03 14:03:13 -04:00
Lincoln Stein
b128ba81db
Merge branch 'main' into lstein/remove-hardcoded-cuda-device
2023-07-03 13:58:14 -04:00
Lincoln Stein
d6de11bd56
resolve merge conflict
2023-07-03 12:19:11 -04:00
Lincoln Stein
ed86d0b708
Union[foo, None]=>Optional[foo]
2023-07-03 12:17:45 -04:00