InvokeAI/invokeai/backend/quantization
2024-08-21 08:59:19 -04:00
..
bnb_llm_int8.py More improvements for LLM.int8() - not fully tested. 2024-08-21 08:59:19 -04:00
bnb_nf4.py LLM.int8() quantization is working, but still some rough edges to solve. 2024-08-21 08:59:19 -04:00
fast_quantized_diffusion_model.py Make quantized loading fast for both T5XXL and FLUX transformer. 2024-08-21 08:59:19 -04:00
fast_quantized_transformers_model.py Make quantized loading fast for both T5XXL and FLUX transformer. 2024-08-21 08:59:19 -04:00
load_flux_model_bnb_llm_int8.py LLM.int8() quantization is working, but still some rough edges to solve. 2024-08-21 08:59:19 -04:00
load_flux_model_bnb_nf4.py WIP on moving from diffusers to FLUX 2024-08-21 08:59:19 -04:00