mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
On hyperthreaded CPUs we get two threads operating on the queue by default on each core. This cases two threads to process queue items. This results in pytorch errors and sometimes generates garbage. Locking this to single thread makes sense because we are bound by the number of GPUs in the system, not by CPU cores. And to parallelize across GPUs we should just start multiple processors (and use async instead of threading) Fixes #3289 |
||
---|---|---|
.. | ||
api | ||
cli | ||
invocations | ||
models | ||
services | ||
util | ||
api_app.py | ||
cli_app.py |