mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
~7% speedup by switch to += in ldm.modules.attention. (#569)
Tested on 8GB eGPU nvidia setup so YMMV. Re-land with .clone() fix, context #508
This commit is contained in:
parent
94a7b3cc07
commit
dd3fff1d3e
@ -281,9 +281,9 @@ class BasicTransformerBlock(nn.Module):
|
||||
|
||||
def _forward(self, x, context=None):
|
||||
x = x.contiguous() if x.device.type == 'mps' else x
|
||||
x = self.attn1(self.norm1(x)) + x
|
||||
x = self.attn2(self.norm2(x), context=context) + x
|
||||
x = self.ff(self.norm3(x)) + x
|
||||
x += self.attn1(self.norm1(x.clone()))
|
||||
x += self.attn2(self.norm2(x.clone()), context=context)
|
||||
x += self.ff(self.norm3(x.clone()))
|
||||
return x
|
||||
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user