InvokeAI/invokeai/backend
2024-07-04 09:08:25 -04:00
..
image_util Merge branch 'main' into lstein/feat/simple-mm2-api 2024-05-17 22:54:03 -04:00
ip_adapter Update calc_model_size_by_data(...) to handle all expected model types, and to log an error if an unexpected model type is received. 2024-07-04 09:08:25 -04:00
model_hash Run ruff 2024-06-26 21:46:59 +10:00
model_manager Update calc_model_size_by_data(...) to handle all expected model types, and to log an error if an unexpected model type is received. 2024-07-04 09:08:25 -04:00
onnx Improve RAM<->VRAM memory copy performance in LoRA patching and elsewhere (#6490) 2024-06-13 17:10:03 +00:00
stable_diffusion Expose the VAE tile_size on the VAE encode and decode invocations. 2024-07-02 09:07:03 -04:00
tiles feat(nodes): extract LATENT_SCALE_FACTOR to constants.py 2024-03-01 10:42:33 +11:00
util fix(backend): mps should not use non_blocking 2024-06-27 19:15:23 +10:00
__init__.py consolidate model manager parts into a single class 2024-03-01 10:42:33 +11:00
lora.py fix(backend): mps should not use non_blocking 2024-06-27 19:15:23 +10:00
model_patcher.py ruff format 2024-06-27 09:45:13 -04:00
raw_model.py Improve RAM<->VRAM memory copy performance in LoRA patching and elsewhere (#6490) 2024-06-13 17:10:03 +00:00
textual_inversion.py Update calc_model_size_by_data(...) to handle all expected model types, and to log an error if an unexpected model type is received. 2024-07-04 09:08:25 -04:00