From 77c5c18542569ed261af80efd2d89b1ea9e62ffd Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Wed, 2 Aug 2023 09:11:24 -0400 Subject: [PATCH 1/6] add slider for VRAM cache --- .../backend/install/invokeai_configure.py | 38 ++++++++++++++++++- invokeai/frontend/install/widgets.py | 2 +- 2 files changed, 38 insertions(+), 2 deletions(-) diff --git a/invokeai/backend/install/invokeai_configure.py b/invokeai/backend/install/invokeai_configure.py index 4bf2a484a1..9136604b14 100755 --- a/invokeai/backend/install/invokeai_configure.py +++ b/invokeai/backend/install/invokeai_configure.py @@ -12,10 +12,12 @@ import io import os import shutil import textwrap +import torch import traceback import yaml import warnings from argparse import Namespace +from enum import Enum from pathlib import Path from shutil import get_terminal_size from typing import get_type_hints @@ -49,6 +51,7 @@ from invokeai.frontend.install.widgets import ( CenteredButtonPress, FileBox, IntTitleSlider, + FloatTitleSlider, set_min_terminal_size, CyclingForm, MIN_COLS, @@ -76,6 +79,9 @@ Default_config_file = config.model_conf_path SD_Configs = config.legacy_conf_path PRECISION_CHOICES = ["auto", "float16", "float32"] +HAS_CUDA = torch.cuda.is_available() +_, MAX_VRAM = torch.cuda.mem_get_info() if HAS_CUDA else (0, 0) +MAX_VRAM /= 1073741824 # GB in bytes INIT_FILE_PREAMBLE = """# InvokeAI initialization file # This is the InvokeAI initialization file, which contains command-line default values. @@ -86,6 +92,12 @@ INIT_FILE_PREAMBLE = """# InvokeAI initialization file logger = InvokeAILogger.getLogger() +class DummyWidgetValue(Enum): + zero = 0 + true = True + false = False + + # -------------------------------------------- def postscript(errors: None): if not any(errors): @@ -378,13 +390,36 @@ Use cursor arrows to make a checkbox selection, and space to toggle. ) self.max_cache_size = self.add_widget_intelligent( IntTitleSlider, - name="Size of the RAM cache used for fast model switching (GB)", + name="RAM cache size. The larger this is, the more models can be kept in memory rather than loading from disk each time (GB)", value=old_opts.max_cache_size, out_of=20, lowest=3, begin_entry_at=6, scroll_exit=True, ) + if HAS_CUDA: + self.nextrely += 1 + self.add_widget_intelligent( + npyscreen.TitleFixedText, + name="VRAM cache size. Make this large enough to hold an entire model, but not more than half your available VRAM (GB)", + begin_entry_at=0, + editable=False, + color="CONTROL", + scroll_exit=True, + ) + self.nextrely -= 1 + self.max_vram_cache_size = self.add_widget_intelligent( + npyscreen.Slider, + value=old_opts.max_vram_cache_size, + out_of=round(MAX_VRAM * 2) / 2, + lowest=0.0, + relx=8, + step=0.25, + begin_entry_at=MAX_VRAM * 0.55, + scroll_exit=True, + ) + else: + self.max_vram_cache_size = DummyWidgetValue.zero self.nextrely += 1 self.outdir = self.add_widget_intelligent( FileBox, @@ -476,6 +511,7 @@ https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENS "outdir", "free_gpu_mem", "max_cache_size", + "max_vram_cache_size", "xformers_enabled", "always_use_cpu", ]: diff --git a/invokeai/frontend/install/widgets.py b/invokeai/frontend/install/widgets.py index 10da15bf13..c1481035d5 100644 --- a/invokeai/frontend/install/widgets.py +++ b/invokeai/frontend/install/widgets.py @@ -164,7 +164,7 @@ class FloatSlider(npyscreen.Slider): class FloatTitleSlider(npyscreen.TitleText): - _entry_type = FloatSlider + _entry_type = npyscreen.Slider class SelectColumnBase: From 880727436c9e3b5c531bf7034c30bbd30b231579 Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Wed, 2 Aug 2023 09:43:52 -0400 Subject: [PATCH 2/6] fix default vram cache size calculation --- invokeai/app/services/config.py | 4 ++-- invokeai/backend/install/invokeai_configure.py | 8 +++++--- 2 files changed, 7 insertions(+), 5 deletions(-) diff --git a/invokeai/app/services/config.py b/invokeai/app/services/config.py index a9a4a64f75..264a93955c 100644 --- a/invokeai/app/services/config.py +++ b/invokeai/app/services/config.py @@ -173,7 +173,7 @@ from typing import ClassVar, Dict, List, Set, Literal, Union, get_origin, get_ty INIT_FILE = Path("invokeai.yaml") DB_FILE = Path("invokeai.db") LEGACY_INIT_FILE = Path("invokeai.init") - +DEFAULT_MAX_VRAM = 2.75 class InvokeAISettings(BaseSettings): """ @@ -395,7 +395,7 @@ class InvokeAIAppConfig(InvokeAISettings): free_gpu_mem : bool = Field(default=False, description="If true, purge model from GPU after each generation.", category='Memory/Performance') max_loaded_models : int = Field(default=3, gt=0, description="(DEPRECATED: use max_cache_size) Maximum number of models to keep in memory for rapid switching", category='DEPRECATED') max_cache_size : float = Field(default=6.0, gt=0, description="Maximum memory amount used by model cache for rapid switching", category='Memory/Performance') - max_vram_cache_size : float = Field(default=2.75, ge=0, description="Amount of VRAM reserved for model storage", category='Memory/Performance') + max_vram_cache_size : float = Field(default=DEFAULT_MAX_VRAM, ge=0, description="Amount of VRAM reserved for model storage", category='Memory/Performance') gpu_mem_reserved : float = Field(default=2.75, ge=0, description="DEPRECATED: use max_vram_cache_size. Amount of VRAM reserved for model storage", category='DEPRECATED') nsfw_checker : bool = Field(default=True, description="DEPRECATED: use Web settings to enable/disable", category='DEPRECATED') precision : Literal[tuple(['auto','float16','float32','autocast'])] = Field(default='auto',description='Floating point precision', category='Memory/Performance') diff --git a/invokeai/backend/install/invokeai_configure.py b/invokeai/backend/install/invokeai_configure.py index 9136604b14..2e95993d75 100755 --- a/invokeai/backend/install/invokeai_configure.py +++ b/invokeai/backend/install/invokeai_configure.py @@ -82,6 +82,7 @@ PRECISION_CHOICES = ["auto", "float16", "float32"] HAS_CUDA = torch.cuda.is_available() _, MAX_VRAM = torch.cuda.mem_get_info() if HAS_CUDA else (0, 0) MAX_VRAM /= 1073741824 # GB in bytes +MAX_VRAM_CACHE_RATIO = 0.55 # first guess of optimal vram cache based on total available INIT_FILE_PREAMBLE = """# InvokeAI initialization file # This is the InvokeAI initialization file, which contains command-line default values. @@ -568,9 +569,11 @@ def edit_opts(program_opts: Namespace, invokeai_opts: Namespace) -> argparse.Nam def default_startup_options(init_file: Path) -> Namespace: opts = InvokeAIAppConfig.get_config() + # dynamically adust vram for memory size + if not init_file.exists(): + opts.max_vram_cache_size = round((MAX_VRAM * MAX_VRAM_CACHE_RATIO)*4) / 4 return opts - def default_user_selections(program_opts: Namespace) -> InstallSelections: try: installer = ModelInstall(config) @@ -628,7 +631,6 @@ def maybe_create_models_yaml(root: Path): # ------------------------------------- def run_console_ui(program_opts: Namespace, initfile: Path = None) -> (Namespace, Namespace): - # parse_args() will read from init file if present invokeai_opts = default_startup_options(initfile) invokeai_opts.root = program_opts.root @@ -711,7 +713,6 @@ def migrate_init_file(legacy_format: Path): # ------------------------------------- def migrate_models(root: Path): from invokeai.backend.install.migrate_to_3 import do_migrate - do_migrate(root, root) @@ -813,6 +814,7 @@ def main(): models_to_download = default_user_selections(opt) new_init_file = config.root_path / "invokeai.yaml" + if opt.yes_to_all: write_default_options(opt, new_init_file) init_options = Namespace(precision="float32" if opt.full_precision else "float16") From 29ac252501133dde3d2092d8b1e66e79f6de9799 Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Wed, 2 Aug 2023 09:44:06 -0400 Subject: [PATCH 3/6] blackify --- invokeai/app/services/config.py | 1 + invokeai/backend/install/invokeai_configure.py | 6 ++++-- 2 files changed, 5 insertions(+), 2 deletions(-) diff --git a/invokeai/app/services/config.py b/invokeai/app/services/config.py index 264a93955c..1509176478 100644 --- a/invokeai/app/services/config.py +++ b/invokeai/app/services/config.py @@ -175,6 +175,7 @@ DB_FILE = Path("invokeai.db") LEGACY_INIT_FILE = Path("invokeai.init") DEFAULT_MAX_VRAM = 2.75 + class InvokeAISettings(BaseSettings): """ Runtime configuration settings in which default values are diff --git a/invokeai/backend/install/invokeai_configure.py b/invokeai/backend/install/invokeai_configure.py index 2e95993d75..6cf1101b4c 100755 --- a/invokeai/backend/install/invokeai_configure.py +++ b/invokeai/backend/install/invokeai_configure.py @@ -82,7 +82,7 @@ PRECISION_CHOICES = ["auto", "float16", "float32"] HAS_CUDA = torch.cuda.is_available() _, MAX_VRAM = torch.cuda.mem_get_info() if HAS_CUDA else (0, 0) MAX_VRAM /= 1073741824 # GB in bytes -MAX_VRAM_CACHE_RATIO = 0.55 # first guess of optimal vram cache based on total available +MAX_VRAM_CACHE_RATIO = 0.55 # first guess of optimal vram cache based on total available INIT_FILE_PREAMBLE = """# InvokeAI initialization file # This is the InvokeAI initialization file, which contains command-line default values. @@ -571,9 +571,10 @@ def default_startup_options(init_file: Path) -> Namespace: opts = InvokeAIAppConfig.get_config() # dynamically adust vram for memory size if not init_file.exists(): - opts.max_vram_cache_size = round((MAX_VRAM * MAX_VRAM_CACHE_RATIO)*4) / 4 + opts.max_vram_cache_size = round((MAX_VRAM * MAX_VRAM_CACHE_RATIO) * 4) / 4 return opts + def default_user_selections(program_opts: Namespace) -> InstallSelections: try: installer = ModelInstall(config) @@ -713,6 +714,7 @@ def migrate_init_file(legacy_format: Path): # ------------------------------------- def migrate_models(root: Path): from invokeai.backend.install.migrate_to_3 import do_migrate + do_migrate(root, root) From 5de42be4a6bf9a5b2435693e588c529856327257 Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Wed, 2 Aug 2023 14:27:13 -0400 Subject: [PATCH 4/6] reduce VRAM cache default; take max RAM from system --- invokeai/app/services/config.py | 5 ++--- invokeai/backend/install/invokeai_configure.py | 18 +++++++++--------- 2 files changed, 11 insertions(+), 12 deletions(-) diff --git a/invokeai/app/services/config.py b/invokeai/app/services/config.py index 1509176478..e88523eb8f 100644 --- a/invokeai/app/services/config.py +++ b/invokeai/app/services/config.py @@ -24,7 +24,7 @@ InvokeAI: sequential_guidance: false precision: float16 max_cache_size: 6 - max_vram_cache_size: 2.7 + max_vram_cache_size: 0.5 always_use_cpu: false free_gpu_mem: false Features: @@ -173,8 +173,7 @@ from typing import ClassVar, Dict, List, Set, Literal, Union, get_origin, get_ty INIT_FILE = Path("invokeai.yaml") DB_FILE = Path("invokeai.db") LEGACY_INIT_FILE = Path("invokeai.init") -DEFAULT_MAX_VRAM = 2.75 - +DEFAULT_MAX_VRAM = 0.5 class InvokeAISettings(BaseSettings): """ diff --git a/invokeai/backend/install/invokeai_configure.py b/invokeai/backend/install/invokeai_configure.py index 6cf1101b4c..c329d7b3fb 100755 --- a/invokeai/backend/install/invokeai_configure.py +++ b/invokeai/backend/install/invokeai_configure.py @@ -10,6 +10,7 @@ import sys import argparse import io import os +import psutil import shutil import textwrap import torch @@ -79,10 +80,13 @@ Default_config_file = config.model_conf_path SD_Configs = config.legacy_conf_path PRECISION_CHOICES = ["auto", "float16", "float32"] +GB = 1073741824 # GB in bytes HAS_CUDA = torch.cuda.is_available() _, MAX_VRAM = torch.cuda.mem_get_info() if HAS_CUDA else (0, 0) -MAX_VRAM /= 1073741824 # GB in bytes -MAX_VRAM_CACHE_RATIO = 0.55 # first guess of optimal vram cache based on total available + + +MAX_VRAM /= GB +MAX_RAM = psutil.virtual_memory().total / GB INIT_FILE_PREAMBLE = """# InvokeAI initialization file # This is the InvokeAI initialization file, which contains command-line default values. @@ -391,9 +395,9 @@ Use cursor arrows to make a checkbox selection, and space to toggle. ) self.max_cache_size = self.add_widget_intelligent( IntTitleSlider, - name="RAM cache size. The larger this is, the more models can be kept in memory rather than loading from disk each time (GB)", + name="RAM cache size (GB). Make this at least large enough to hold a single model. Larger sizes will allow you to switch between models quickly without reading from disk.", value=old_opts.max_cache_size, - out_of=20, + out_of=MAX_RAM, lowest=3, begin_entry_at=6, scroll_exit=True, @@ -402,7 +406,7 @@ Use cursor arrows to make a checkbox selection, and space to toggle. self.nextrely += 1 self.add_widget_intelligent( npyscreen.TitleFixedText, - name="VRAM cache size. Make this large enough to hold an entire model, but not more than half your available VRAM (GB)", + name="VRAM cache size (GB). Reserving a small amount of VRAM will modestly speed up the start of image generation.", begin_entry_at=0, editable=False, color="CONTROL", @@ -416,7 +420,6 @@ Use cursor arrows to make a checkbox selection, and space to toggle. lowest=0.0, relx=8, step=0.25, - begin_entry_at=MAX_VRAM * 0.55, scroll_exit=True, ) else: @@ -569,9 +572,6 @@ def edit_opts(program_opts: Namespace, invokeai_opts: Namespace) -> argparse.Nam def default_startup_options(init_file: Path) -> Namespace: opts = InvokeAIAppConfig.get_config() - # dynamically adust vram for memory size - if not init_file.exists(): - opts.max_vram_cache_size = round((MAX_VRAM * MAX_VRAM_CACHE_RATIO) * 4) / 4 return opts From ec48779080a160b324d521ac0379a25da4150e89 Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Wed, 2 Aug 2023 14:28:19 -0400 Subject: [PATCH 5/6] blackify --- invokeai/app/services/config.py | 1 + invokeai/backend/install/invokeai_configure.py | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/invokeai/app/services/config.py b/invokeai/app/services/config.py index 25a7b8ffb3..271d617ff8 100644 --- a/invokeai/app/services/config.py +++ b/invokeai/app/services/config.py @@ -175,6 +175,7 @@ DB_FILE = Path("invokeai.db") LEGACY_INIT_FILE = Path("invokeai.init") DEFAULT_MAX_VRAM = 0.5 + class InvokeAISettings(BaseSettings): """ Runtime configuration settings in which default values are diff --git a/invokeai/backend/install/invokeai_configure.py b/invokeai/backend/install/invokeai_configure.py index c329d7b3fb..692eff8867 100755 --- a/invokeai/backend/install/invokeai_configure.py +++ b/invokeai/backend/install/invokeai_configure.py @@ -86,7 +86,7 @@ _, MAX_VRAM = torch.cuda.mem_get_info() if HAS_CUDA else (0, 0) MAX_VRAM /= GB -MAX_RAM = psutil.virtual_memory().total / GB +MAX_RAM = psutil.virtual_memory().total / GB INIT_FILE_PREAMBLE = """# InvokeAI initialization file # This is the InvokeAI initialization file, which contains command-line default values. From f56f19710dc7771c9b59ea119016caaaa27d8e5c Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Tue, 8 Aug 2023 12:27:25 -0400 Subject: [PATCH 6/6] allow user to interactively resize screen before UI runs --- .../backend/install/invokeai_configure.py | 15 +++-- invokeai/frontend/install/model_install.py | 13 ++-- invokeai/frontend/install/widgets.py | 59 +++++++++++-------- 3 files changed, 52 insertions(+), 35 deletions(-) diff --git a/invokeai/backend/install/invokeai_configure.py b/invokeai/backend/install/invokeai_configure.py index 692eff8867..18fb2352f3 100755 --- a/invokeai/backend/install/invokeai_configure.py +++ b/invokeai/backend/install/invokeai_configure.py @@ -52,11 +52,11 @@ from invokeai.frontend.install.widgets import ( CenteredButtonPress, FileBox, IntTitleSlider, - FloatTitleSlider, set_min_terminal_size, CyclingForm, MIN_COLS, MIN_LINES, + WindowTooSmallException, ) from invokeai.backend.install.legacy_arg_parsing import legacy_parser from invokeai.backend.install.model_install_backend import ( @@ -395,7 +395,7 @@ Use cursor arrows to make a checkbox selection, and space to toggle. ) self.max_cache_size = self.add_widget_intelligent( IntTitleSlider, - name="RAM cache size (GB). Make this at least large enough to hold a single model. Larger sizes will allow you to switch between models quickly without reading from disk.", + name="RAM cache size (GB). Make this at least large enough to hold a single full model.", value=old_opts.max_cache_size, out_of=MAX_RAM, lowest=3, @@ -440,7 +440,7 @@ Use cursor arrows to make a checkbox selection, and space to toggle. self.autoimport_dirs = {} self.autoimport_dirs["autoimport_dir"] = self.add_widget_intelligent( FileBox, - name=f"Folder to recursively scan for new checkpoints, ControlNets, LoRAs and TI models", + name="Folder to recursively scan for new checkpoints, ControlNets, LoRAs and TI models", value=str(config.root_path / config.autoimport_dir), select_dir=True, must_exist=False, @@ -635,9 +635,10 @@ def run_console_ui(program_opts: Namespace, initfile: Path = None) -> (Namespace invokeai_opts = default_startup_options(initfile) invokeai_opts.root = program_opts.root - # The third argument is needed in the Windows 11 environment to - # launch a console window running this program. - set_min_terminal_size(MIN_COLS, MIN_LINES) + if not set_min_terminal_size(MIN_COLS, MIN_LINES): + raise WindowTooSmallException( + "Could not increase terminal size. Try running again with a larger window or smaller font size." + ) # the install-models application spawns a subprocess to install # models, and will crash unless this is set before running. @@ -842,6 +843,8 @@ def main(): postscript(errors=errors) if not opt.yes_to_all: input("Press any key to continue...") + except WindowTooSmallException as e: + logger.error(str(e)) except KeyboardInterrupt: print("\nGoodbye! Come back soon.") diff --git a/invokeai/frontend/install/model_install.py b/invokeai/frontend/install/model_install.py index 0633553a3d..55e45931e3 100644 --- a/invokeai/frontend/install/model_install.py +++ b/invokeai/frontend/install/model_install.py @@ -28,7 +28,6 @@ from npyscreen import widget from invokeai.backend.util.logging import InvokeAILogger from invokeai.backend.install.model_install_backend import ( - ModelInstallList, InstallSelections, ModelInstall, SchedulerPredictionType, @@ -41,12 +40,12 @@ from invokeai.frontend.install.widgets import ( SingleSelectColumns, TextBox, BufferBox, - FileBox, set_min_terminal_size, select_stable_diffusion_config_file, CyclingForm, MIN_COLS, MIN_LINES, + WindowTooSmallException, ) from invokeai.app.services.config import InvokeAIAppConfig @@ -156,7 +155,7 @@ class addModelsForm(CyclingForm, npyscreen.FormMultiPage): BufferBox, name="Log Messages", editable=False, - max_height=15, + max_height=6, ) self.nextrely += 1 @@ -693,7 +692,11 @@ def select_and_download_models(opt: Namespace): # needed to support the probe() method running under a subprocess torch.multiprocessing.set_start_method("spawn") - set_min_terminal_size(MIN_COLS, MIN_LINES) + if not set_min_terminal_size(MIN_COLS, MIN_LINES): + raise WindowTooSmallException( + "Could not increase terminal size. Try running again with a larger window or smaller font size." + ) + installApp = AddModelApplication(opt) try: installApp.run() @@ -787,6 +790,8 @@ def main(): curses.echo() curses.endwin() logger.info("Goodbye! Come back soon.") + except WindowTooSmallException as e: + logger.error(str(e)) except widget.NotEnoughSpaceForWidget as e: if str(e).startswith("Height of 1 allocated"): logger.error("Insufficient vertical space for the interface. Please make your window taller and try again") diff --git a/invokeai/frontend/install/widgets.py b/invokeai/frontend/install/widgets.py index c1481035d5..156503f0b8 100644 --- a/invokeai/frontend/install/widgets.py +++ b/invokeai/frontend/install/widgets.py @@ -21,31 +21,40 @@ MIN_COLS = 130 MIN_LINES = 38 +class WindowTooSmallException(Exception): + pass + + # ------------------------------------- -def set_terminal_size(columns: int, lines: int): - ts = get_terminal_size() - width = max(columns, ts.columns) - height = max(lines, ts.lines) - +def set_terminal_size(columns: int, lines: int) -> bool: OS = platform.uname().system - if OS == "Windows": - pass - # not working reliably - ask user to adjust the window - # _set_terminal_size_powershell(width,height) - elif OS in ["Darwin", "Linux"]: - _set_terminal_size_unix(width, height) + screen_ok = False + while not screen_ok: + ts = get_terminal_size() + width = max(columns, ts.columns) + height = max(lines, ts.lines) - # check whether it worked.... - ts = get_terminal_size() - pause = False - if ts.columns < columns: - print("\033[1mThis window is too narrow for the user interface.\033[0m") - pause = True - if ts.lines < lines: - print("\033[1mThis window is too short for the user interface.\033[0m") - pause = True - if pause: - input("Maximize the window then press any key to continue..") + if OS == "Windows": + pass + # not working reliably - ask user to adjust the window + # _set_terminal_size_powershell(width,height) + elif OS in ["Darwin", "Linux"]: + _set_terminal_size_unix(width, height) + + # check whether it worked.... + ts = get_terminal_size() + if ts.columns < columns or ts.lines < lines: + print( + f"\033[1mThis window is too small for the interface. InvokeAI requires {columns}x{lines} (w x h) characters, but window is {ts.columns}x{ts.lines}\033[0m" + ) + resp = input( + "Maximize the window and/or decrease the font size then press any key to continue. Type [Q] to give up.." + ) + if resp.upper().startswith("Q"): + break + else: + screen_ok = True + return screen_ok def _set_terminal_size_powershell(width: int, height: int): @@ -80,14 +89,14 @@ def _set_terminal_size_unix(width: int, height: int): sys.stdout.flush() -def set_min_terminal_size(min_cols: int, min_lines: int): +def set_min_terminal_size(min_cols: int, min_lines: int) -> bool: # make sure there's enough room for the ui term_cols, term_lines = get_terminal_size() if term_cols >= min_cols and term_lines >= min_lines: - return + return True cols = max(term_cols, min_cols) lines = max(term_lines, min_lines) - set_terminal_size(cols, lines) + return set_terminal_size(cols, lines) class IntSlider(npyscreen.Slider):