InvokeAI/invokeai/frontend
psychedelicious b76d2cd716 fix(ui): handle base model compat when recalling parameters
We had a one-behind issue with recalling metadata items that had a model.

For example, when recalling LoRAs, we check against the current main model to decide whether or not the requested LoRA is compatible and may be recalled.

When recalling all params, we are often also recalling the main model, but the compat logic didn't compare against this new main model.

The logic is updated to check against the new main model, if one is being set.

Closes #5512
2024-01-23 14:13:18 -06:00
..
CLI Resolving merge conflicts for flake8 2023-08-18 15:52:04 +10:00
install Update recommends db backup when installing RC (#5381) 2024-01-02 22:44:45 +00:00
merge chore: ruff check - fix flake8-comprensions 2023-11-11 10:55:23 +11:00
training chore: ruff check - fix flake8-comprensions 2023-11-11 10:55:23 +11:00
web fix(ui): handle base model compat when recalling parameters 2024-01-23 14:13:18 -06:00
__init__.py Run python black 2023-07-28 09:46:44 -04:00
legacy_launch_invokeai.py isort wip 2023-09-12 13:01:58 -04:00