|
9ed53af520
|
Run Ruff
|
2024-08-26 20:17:50 -04:00 |
|
|
56fda669fd
|
Manage quantization of models within the loader
|
2024-08-26 20:17:50 -04:00 |
|
|
1fa6bddc89
|
WIP on moving from diffusers to FLUX
|
2024-08-26 20:17:50 -04:00 |
|
|
d3a5ca5247
|
More improvements for LLM.int8() - not fully tested.
|
2024-08-26 20:17:50 -04:00 |
|
|
f01f56a98e
|
LLM.int8() quantization is working, but still some rough edges to solve.
|
2024-08-26 20:17:50 -04:00 |
|
|
99b0f79784
|
Clean up NF4 implementation.
|
2024-08-26 20:17:50 -04:00 |
|
|
eeabb7ebe5
|
Make quantized loading fast for both T5XXL and FLUX transformer.
|
2024-08-26 20:17:50 -04:00 |
|