From 6a77634b34be4c996c3fafbe9bf594e51f1651bb Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Sat, 11 Mar 2023 17:14:03 -0500 Subject: [PATCH] remove unneeded generate initializer routines --- invokeai/app/services/generate_initializer.py | 35 ------------------- .../backend/model_management/model_manager.py | 2 +- 2 files changed, 1 insertion(+), 36 deletions(-) delete mode 100644 invokeai/app/services/generate_initializer.py diff --git a/invokeai/app/services/generate_initializer.py b/invokeai/app/services/generate_initializer.py deleted file mode 100644 index 6dd65e69b1..0000000000 --- a/invokeai/app/services/generate_initializer.py +++ /dev/null @@ -1,35 +0,0 @@ -import os -import sys -import torch -import traceback -from argparse import Namespace -from omegaconf import OmegaConf - -import invokeai.version -from ...backend import Globals - -def load_face_restoration(opt): - try: - gfpgan, codeformer, esrgan = None, None, None - if opt.restore or opt.esrgan: - from invokeai.backend.restoration import Restoration - - restoration = Restoration() - if opt.restore: - gfpgan, codeformer = restoration.load_face_restore_models( - opt.gfpgan_model_path - ) - else: - print(">> Face restoration disabled") - if opt.esrgan: - esrgan = restoration.load_esrgan(opt.esrgan_bg_tile) - else: - print(">> Upscaling disabled") - else: - print(">> Face restoration and upscaling disabled") - except (ModuleNotFoundError, ImportError): - print(traceback.format_exc(), file=sys.stderr) - print(">> You may need to install the ESRGAN and/or GFPGAN modules") - return gfpgan, codeformer, esrgan - - diff --git a/invokeai/backend/model_management/model_manager.py b/invokeai/backend/model_management/model_manager.py index 50ea13a2d0..06b1490c93 100644 --- a/invokeai/backend/model_management/model_manager.py +++ b/invokeai/backend/model_management/model_manager.py @@ -90,7 +90,7 @@ class ModelManager(object): """ return model_name in self.config - def get_model(self, model_name: str=None): + def get_model(self, model_name: str=None)->dict: """ Given a model named identified in models.yaml, return the model object. If in RAM will load into GPU VRAM.