From 8cb2fa86008d16232b22bb4e653c9869522dbc8a Mon Sep 17 00:00:00 2001 From: StAlKeR7779 Date: Tue, 25 Apr 2023 04:29:17 +0300 Subject: [PATCH] Restore log_tokenization check --- invokeai/app/invocations/compel.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/invokeai/app/invocations/compel.py b/invokeai/app/invocations/compel.py index 87f6f0fcca..653c60b686 100644 --- a/invokeai/app/invocations/compel.py +++ b/invokeai/app/invocations/compel.py @@ -103,7 +103,7 @@ class CompelInvocation(BaseInvocation): positive_prompt: Union[FlattenedPrompt, Blend] = Compel.parse_prompt_string(self.positive_prompt) negative_prompt: Union[FlattenedPrompt, Blend] = Compel.parse_prompt_string(self.negative_prompt) - if True: #getattr(Globals, "log_tokenization", False): + if getattr(Globals, "log_tokenization", False): log_tokenization(positive_prompt, negative_prompt, tokenizer=tokenizer) # TODO: add lora(with model and clip field types)