2022-08-28 21:27:43 +00:00
|
|
|
import json
|
|
|
|
import base64
|
|
|
|
import mimetypes
|
2022-08-28 20:37:27 +00:00
|
|
|
import os
|
2022-08-28 21:27:43 +00:00
|
|
|
from http.server import BaseHTTPRequestHandler, ThreadingHTTPServer
|
2022-08-27 01:10:13 +00:00
|
|
|
from ldm.dream.pngwriter import PngWriter
|
2022-08-30 15:55:40 +00:00
|
|
|
from threading import Event
|
|
|
|
|
|
|
|
class CanceledException(Exception):
|
|
|
|
pass
|
2022-08-28 21:27:43 +00:00
|
|
|
|
|
|
|
class DreamServer(BaseHTTPRequestHandler):
|
|
|
|
model = None
|
2022-08-30 15:55:40 +00:00
|
|
|
canceled = Event()
|
2022-08-28 21:27:43 +00:00
|
|
|
|
|
|
|
def do_GET(self):
|
|
|
|
if self.path == "/":
|
|
|
|
self.send_response(200)
|
|
|
|
self.send_header("Content-type", "text/html")
|
|
|
|
self.end_headers()
|
|
|
|
with open("./static/dream_web/index.html", "rb") as content:
|
|
|
|
self.wfile.write(content.read())
|
2022-08-29 15:53:27 +00:00
|
|
|
elif self.path == "/config.js":
|
|
|
|
# unfortunately this import can't be at the top level, since that would cause a circular import
|
|
|
|
from ldm.gfpgan.gfpgan_tools import gfpgan_model_exists
|
|
|
|
self.send_response(200)
|
|
|
|
self.send_header("Content-type", "application/javascript")
|
|
|
|
self.end_headers()
|
|
|
|
config = {
|
|
|
|
'gfpgan_model_exists': gfpgan_model_exists
|
|
|
|
}
|
|
|
|
self.wfile.write(bytes("let config = " + json.dumps(config) + ";\n", "utf-8"))
|
2022-08-30 15:55:40 +00:00
|
|
|
elif self.path == "/cancel":
|
|
|
|
self.canceled.set()
|
|
|
|
self.send_response(200)
|
|
|
|
self.send_header("Content-type", "application/json")
|
|
|
|
self.end_headers()
|
|
|
|
self.wfile.write(bytes('{}', 'utf8'))
|
2022-08-28 21:33:30 +00:00
|
|
|
else:
|
|
|
|
path = "." + self.path
|
2022-08-29 21:43:34 +00:00
|
|
|
cwd = os.path.realpath(os.getcwd())
|
|
|
|
is_in_cwd = os.path.commonprefix((os.path.realpath(path), cwd)) == cwd
|
2022-08-28 21:33:30 +00:00
|
|
|
if not (is_in_cwd and os.path.exists(path)):
|
|
|
|
self.send_response(404)
|
|
|
|
return
|
|
|
|
mime_type = mimetypes.guess_type(path)[0]
|
2022-08-28 21:27:43 +00:00
|
|
|
if mime_type is not None:
|
|
|
|
self.send_response(200)
|
|
|
|
self.send_header("Content-type", mime_type)
|
|
|
|
self.end_headers()
|
|
|
|
with open("." + self.path, "rb") as content:
|
|
|
|
self.wfile.write(content.read())
|
2022-08-28 20:37:27 +00:00
|
|
|
else:
|
2022-08-28 21:27:43 +00:00
|
|
|
self.send_response(404)
|
|
|
|
|
|
|
|
def do_POST(self):
|
|
|
|
self.send_response(200)
|
|
|
|
self.send_header("Content-type", "application/json")
|
|
|
|
self.end_headers()
|
|
|
|
|
2022-08-29 15:53:27 +00:00
|
|
|
# unfortunately this import can't be at the top level, since that would cause a circular import
|
|
|
|
from ldm.gfpgan.gfpgan_tools import gfpgan_model_exists
|
|
|
|
|
2022-08-28 21:27:43 +00:00
|
|
|
content_length = int(self.headers['Content-Length'])
|
|
|
|
post_data = json.loads(self.rfile.read(content_length))
|
|
|
|
prompt = post_data['prompt']
|
|
|
|
initimg = post_data['initimg']
|
|
|
|
iterations = int(post_data['iterations'])
|
|
|
|
steps = int(post_data['steps'])
|
|
|
|
width = int(post_data['width'])
|
|
|
|
height = int(post_data['height'])
|
|
|
|
cfgscale = float(post_data['cfgscale'])
|
2022-08-29 18:06:18 +00:00
|
|
|
sampler_name = post_data['sampler']
|
2022-08-29 15:53:27 +00:00
|
|
|
gfpgan_strength = float(post_data['gfpgan_strength']) if gfpgan_model_exists else 0
|
2022-08-29 02:42:31 +00:00
|
|
|
upscale_level = post_data['upscale_level']
|
|
|
|
upscale_strength = post_data['upscale_strength']
|
|
|
|
upscale = [int(upscale_level),float(upscale_strength)] if upscale_level != '' else None
|
2022-08-29 16:36:48 +00:00
|
|
|
progress_images = 'progress_images' in post_data
|
|
|
|
seed = self.model.seed if int(post_data['seed']) == -1 else int(post_data['seed'])
|
2022-08-28 21:27:43 +00:00
|
|
|
|
2022-08-30 15:55:40 +00:00
|
|
|
self.canceled.clear()
|
2022-08-28 21:27:43 +00:00
|
|
|
print(f"Request to generate with prompt: {prompt}")
|
2022-08-29 16:08:18 +00:00
|
|
|
# In order to handle upscaled images, the PngWriter needs to maintain state
|
|
|
|
# across images generated by each call to prompt2img(), so we define it in
|
|
|
|
# the outer scope of image_done()
|
|
|
|
config = post_data.copy() # Shallow copy
|
|
|
|
config['initimg'] = ''
|
|
|
|
|
|
|
|
images_generated = 0 # helps keep track of when upscaling is started
|
|
|
|
images_upscaled = 0 # helps keep track of when upscaling is completed
|
2022-08-31 04:21:04 +00:00
|
|
|
pngwriter = PngWriter("./outputs/img-samples/")
|
2022-08-29 16:08:18 +00:00
|
|
|
|
2022-08-31 04:21:04 +00:00
|
|
|
prefix = pngwriter.unique_prefix()
|
2022-08-29 16:08:18 +00:00
|
|
|
# if upscaling is requested, then this will be called twice, once when
|
|
|
|
# the images are first generated, and then again when after upscaling
|
|
|
|
# is complete. The upscaling replaces the original file, so the second
|
|
|
|
# entry should not be inserted into the image list.
|
2022-08-29 15:53:27 +00:00
|
|
|
def image_done(image, seed, upscaled=False):
|
2022-08-31 04:21:04 +00:00
|
|
|
name = f'{prefix}.{seed}.png'
|
|
|
|
path = pngwriter.save_image_and_prompt_to_png(image, f'{prompt} -S{seed}', name)
|
2022-08-29 16:08:18 +00:00
|
|
|
|
|
|
|
# Append post_data to log, but only once!
|
|
|
|
if not upscaled:
|
|
|
|
with open("./outputs/img-samples/dream_web_log.txt", "a") as log:
|
2022-08-31 04:21:04 +00:00
|
|
|
log.write(f"{path}: {json.dumps(config)}\n")
|
|
|
|
|
2022-08-29 16:08:18 +00:00
|
|
|
self.wfile.write(bytes(json.dumps(
|
2022-08-31 04:33:42 +00:00
|
|
|
{'event': 'result', 'url': path, 'seed': seed, 'config': config}
|
2022-08-29 16:08:18 +00:00
|
|
|
) + '\n',"utf-8"))
|
|
|
|
|
|
|
|
# control state of the "postprocessing..." message
|
|
|
|
upscaling_requested = upscale or gfpgan_strength>0
|
|
|
|
nonlocal images_generated # NB: Is this bad python style? It is typical usage in a perl closure.
|
|
|
|
nonlocal images_upscaled # NB: Is this bad python style? It is typical usage in a perl closure.
|
|
|
|
if upscaled:
|
|
|
|
images_upscaled += 1
|
|
|
|
else:
|
|
|
|
images_generated +=1
|
|
|
|
if upscaling_requested:
|
|
|
|
action = None
|
|
|
|
if images_generated >= iterations:
|
|
|
|
if images_upscaled < iterations:
|
|
|
|
action = 'upscaling-started'
|
|
|
|
else:
|
|
|
|
action = 'upscaling-done'
|
|
|
|
if action:
|
|
|
|
x = images_upscaled+1
|
|
|
|
self.wfile.write(bytes(json.dumps(
|
|
|
|
{'event':action,'processed_file_cnt':f'{x}/{iterations}'}
|
|
|
|
) + '\n',"utf-8"))
|
2022-08-27 01:10:13 +00:00
|
|
|
|
2022-08-31 04:21:04 +00:00
|
|
|
step_writer = PngWriter('./outputs/intermediates/')
|
|
|
|
step_index = 1
|
2022-08-29 16:36:48 +00:00
|
|
|
def image_progress(sample, step):
|
2022-08-30 15:55:40 +00:00
|
|
|
if self.canceled.is_set():
|
|
|
|
self.wfile.write(bytes(json.dumps({'event':'canceled'}) + '\n', 'utf-8'))
|
|
|
|
raise CanceledException
|
2022-08-31 04:21:04 +00:00
|
|
|
path = None
|
2022-08-29 16:36:48 +00:00
|
|
|
# since rendering images is moderately expensive, only render every 5th image
|
|
|
|
# and don't bother with the last one, since it'll render anyway
|
2022-08-31 04:21:04 +00:00
|
|
|
nonlocal step_index
|
2022-08-29 16:36:48 +00:00
|
|
|
if progress_images and step % 5 == 0 and step < steps - 1:
|
2022-08-31 02:30:12 +00:00
|
|
|
image = self.model._sample_to_image(sample)
|
2022-08-31 04:21:04 +00:00
|
|
|
name = f'{prefix}.{seed}.{step_index}.png'
|
|
|
|
metadata = f'{prompt} -S{seed} [intermediate]'
|
|
|
|
path = step_writer.save_image_and_prompt_to_png(image, metadata, name)
|
|
|
|
step_index += 1
|
2022-08-27 01:10:13 +00:00
|
|
|
self.wfile.write(bytes(json.dumps(
|
2022-08-31 04:21:04 +00:00
|
|
|
{'event': 'step', 'step': step + 1, 'url': path}
|
2022-08-27 01:10:13 +00:00
|
|
|
) + '\n',"utf-8"))
|
2022-08-28 21:27:43 +00:00
|
|
|
|
2022-08-30 15:55:40 +00:00
|
|
|
try:
|
|
|
|
if initimg is None:
|
|
|
|
# Run txt2img
|
|
|
|
self.model.prompt2image(prompt,
|
|
|
|
iterations=iterations,
|
|
|
|
cfg_scale = cfgscale,
|
|
|
|
width = width,
|
|
|
|
height = height,
|
|
|
|
seed = seed,
|
|
|
|
steps = steps,
|
|
|
|
gfpgan_strength = gfpgan_strength,
|
|
|
|
upscale = upscale,
|
|
|
|
sampler_name = sampler_name,
|
|
|
|
step_callback=image_progress,
|
|
|
|
image_callback=image_done)
|
|
|
|
else:
|
|
|
|
# Decode initimg as base64 to temp file
|
|
|
|
with open("./img2img-tmp.png", "wb") as f:
|
|
|
|
initimg = initimg.split(",")[1] # Ignore mime type
|
|
|
|
f.write(base64.b64decode(initimg))
|
|
|
|
|
|
|
|
try:
|
|
|
|
# Run img2img
|
|
|
|
self.model.prompt2image(prompt,
|
|
|
|
init_img = "./img2img-tmp.png",
|
|
|
|
iterations = iterations,
|
|
|
|
cfg_scale = cfgscale,
|
|
|
|
seed = seed,
|
|
|
|
steps = steps,
|
|
|
|
sampler_name = sampler_name,
|
|
|
|
gfpgan_strength=gfpgan_strength,
|
|
|
|
upscale = upscale,
|
|
|
|
step_callback=image_progress,
|
|
|
|
image_callback=image_done)
|
|
|
|
finally:
|
|
|
|
# Remove the temp file
|
|
|
|
os.remove("./img2img-tmp.png")
|
|
|
|
except CanceledException:
|
|
|
|
print(f"Canceled.")
|
|
|
|
return
|
2022-08-28 21:27:43 +00:00
|
|
|
|
2022-08-27 01:10:13 +00:00
|
|
|
print(f"Prompt generated!")
|
2022-08-28 21:27:43 +00:00
|
|
|
|
|
|
|
|
|
|
|
class ThreadingDreamServer(ThreadingHTTPServer):
|
|
|
|
def __init__(self, server_address):
|
|
|
|
super(ThreadingDreamServer, self).__init__(server_address, DreamServer)
|