mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
Run FLUX VAE decoding in the user's preferred dtype rather than float32. Tested, and seems to work well at float16.
This commit is contained in:
@ -65,6 +65,7 @@ class FluxVAELoader(ModelLoader):
|
||||
model = AutoEncoder(params)
|
||||
sd = load_file(model_path)
|
||||
model.load_state_dict(sd, assign=True)
|
||||
model.to(dtype=self._torch_dtype)
|
||||
|
||||
return model
|
||||
|
||||
|
Reference in New Issue
Block a user