mirror of
https://github.com/invoke-ai/InvokeAI
synced 2024-08-30 20:32:17 +00:00
eb6e6548ed
BLAKE3 has poor performance on spinning disks when parallelized. See https://github.com/BLAKE3-team/BLAKE3/issues/31 - Replace `skip_model_hash` setting with `hashing_algorithm`. Any algorithm we support is accepted. - Add `random` algorithm: hashes a UUID with BLAKE3 to create a random "hash". Equivalent to the previous skip functionality. - Add `blake3_single` algorithm: hashes on a single thread using BLAKE3, fixes the aforementioned performance issue - Update model probe to accept the algorithm to hash with as an optional arg, defaulting to `blake3` - Update all calls of the probe to use the app's configured hashing algorithm - Update an external script that probes models - Update tests - Move ModelHash into its own module to avoid circuclar import issues |
||
---|---|---|
.. | ||
embeddings | ||
image_util | ||
install | ||
ip_adapter | ||
model_hash | ||
model_manager | ||
onnx | ||
stable_diffusion | ||
tiles | ||
training | ||
util | ||
__init__.py | ||
lora.py | ||
model_patcher.py | ||
raw_model.py | ||
textual_inversion.py |