From 2c6e89ece981125062fa62e005b2999ff831090f Mon Sep 17 00:00:00 2001 From: Brandon Rising Date: Mon, 22 Jan 2024 15:04:09 -0500 Subject: [PATCH] [DO NOT MERGE] Label Testing PR --- invokeai/app/api/routers/download_queue.py | 1 + invokeai/app/invocations/latent.py | 2 ++ .../invocation_processor/invocation_processor_default.py | 1 + invokeai/backend/model_management/lora.py | 2 +- invokeai/frontend/web/package.json | 1 + .../src/features/controlAdapters/hooks/useAddControlAdapter.ts | 2 +- pyproject.toml | 2 +- 7 files changed, 8 insertions(+), 3 deletions(-) diff --git a/invokeai/app/api/routers/download_queue.py b/invokeai/app/api/routers/download_queue.py index 92b658c370..af53771d4d 100644 --- a/invokeai/app/api/routers/download_queue.py +++ b/invokeai/app/api/routers/download_queue.py @@ -24,6 +24,7 @@ download_queue_router = APIRouter(prefix="/v1/download_queue", tags=["download_q ) async def list_downloads() -> List[DownloadJob]: """Get a list of active and inactive jobs.""" + print("test") queue = ApiDependencies.invoker.services.download_queue return queue.list_jobs() diff --git a/invokeai/app/invocations/latent.py b/invokeai/app/invocations/latent.py index b77363ceb8..1915bd8a08 100644 --- a/invokeai/app/invocations/latent.py +++ b/invokeai/app/invocations/latent.py @@ -75,6 +75,8 @@ from .model import ModelInfo, UNetField, VaeField if choose_torch_device() == torch.device("mps"): from torch import mps +print("test") + DEFAULT_PRECISION = choose_precision(choose_torch_device()) SAMPLER_NAME_VALUES = Literal[tuple(SCHEDULER_MAP.keys())] diff --git a/invokeai/app/services/invocation_processor/invocation_processor_default.py b/invokeai/app/services/invocation_processor/invocation_processor_default.py index 09608dca2b..bf59b295c3 100644 --- a/invokeai/app/services/invocation_processor/invocation_processor_default.py +++ b/invokeai/app/services/invocation_processor/invocation_processor_default.py @@ -23,6 +23,7 @@ class DefaultInvocationProcessor(InvocationProcessorABC): self.__threadLimit = BoundedSemaphore(1) self.__invoker = invoker self.__stop_event = Event() + print("test") self.__invoker_thread = Thread( name="invoker_processor", target=self.__process, diff --git a/invokeai/backend/model_management/lora.py b/invokeai/backend/model_management/lora.py index d72f55794d..d5b6113f2b 100644 --- a/invokeai/backend/model_management/lora.py +++ b/invokeai/backend/model_management/lora.py @@ -113,7 +113,7 @@ class ModelPatcher: for layer_key, layer in lora.layers.items(): if not layer_key.startswith(prefix): continue - + print("test") # TODO(ryand): A non-negligible amount of time is currently spent resolving LoRA keys. This # should be improved in the following ways: # 1. The key mapping could be more-efficiently pre-computed. This would save time every time a diff --git a/invokeai/frontend/web/package.json b/invokeai/frontend/web/package.json index 28dfd84fcf..a47ac5c8d9 100644 --- a/invokeai/frontend/web/package.json +++ b/invokeai/frontend/web/package.json @@ -20,6 +20,7 @@ ], "scripts": { "dev": "concurrently \"vite dev\" \"pnpm run theme:watch\"", + "hi": "echo test", "dev:host": "concurrently \"vite dev --host\" \"pnpm run theme:watch\"", "build": "pnpm run lint && vite build", "typegen": "node scripts/typegen.js", diff --git a/invokeai/frontend/web/src/features/controlAdapters/hooks/useAddControlAdapter.ts b/invokeai/frontend/web/src/features/controlAdapters/hooks/useAddControlAdapter.ts index c08adc7f6f..83216dba70 100644 --- a/invokeai/frontend/web/src/features/controlAdapters/hooks/useAddControlAdapter.ts +++ b/invokeai/frontend/web/src/features/controlAdapters/hooks/useAddControlAdapter.ts @@ -16,7 +16,7 @@ export const useAddControlAdapter = (type: ControlAdapterType) => { const firstCompatibleModel = models.filter((m) => baseModel ? m.base_model === baseModel : true )[0]; - + console.log("test") if (firstCompatibleModel) { return firstCompatibleModel; } diff --git a/pyproject.toml b/pyproject.toml index 5ebdd83e22..09e7816f65 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta" [project] name = "InvokeAI" -description = "An implementation of Stable Diffusion which provides various new features and options to aid the image generation process" +description = "An test implementation of Stable Diffusion which provides various new features and options to aid the image generation process" requires-python = ">=3.10, <3.12" readme = { content-type = "text/markdown", file = "README.md" } keywords = ["stable-diffusion", "AI"]