mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
Load model in inpaint when using free_gpu_mem option (#1938)
* Load model in inpaint when using free_gpu_mem option * Passing free_gpu_mem option to inpaint generator
This commit is contained in:
parent
96a12099ed
commit
f4e7383490
@ -806,6 +806,7 @@ class Generate:
|
||||
if not self.generators.get('inpaint'):
|
||||
from ldm.invoke.generator.inpaint import Inpaint
|
||||
self.generators['inpaint'] = Inpaint(self.model, self.precision)
|
||||
self.generators['inpaint'].free_gpu_mem = self.free_gpu_mem
|
||||
return self.generators['inpaint']
|
||||
|
||||
# "omnibus" supports the runwayML custom inpainting model, which does
|
||||
|
@ -291,6 +291,9 @@ class Inpaint(Img2Img):
|
||||
masked_region = (1.0-inpaint_replace) * inverted_mask * z_enc + inpaint_replace * inverted_mask * l_noise
|
||||
z_enc = z_enc * mask_image + masked_region
|
||||
|
||||
if self.free_gpu_mem and self.model.model.device != self.model.device:
|
||||
self.model.model.to(self.model.device)
|
||||
|
||||
# decode it
|
||||
samples = sampler.decode(
|
||||
z_enc,
|
||||
|
Loading…
x
Reference in New Issue
Block a user