mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
move tokenizer into cpu cache as well
This commit is contained in:
@ -678,8 +678,6 @@ class Generate:
|
||||
self.embedding_path, self.precision == 'float32' or self.precision == 'autocast'
|
||||
)
|
||||
|
||||
# model.to doesn't change the cond_stage_model.device used to move the tokenizer output, so set it here
|
||||
self.model.cond_stage_model.device = self.device
|
||||
self._set_sampler()
|
||||
|
||||
for m in self.model.modules():
|
||||
|
Reference in New Issue
Block a user