Brandon Rising
|
a63f842a13
|
Select dev/schnell based on state dict, use correct max seq len based on dev/schnell, and shift in inference, separate vae flux params into separate config
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
4bd7fda694
|
Install sub directories with folders correctly, ensure consistent dtype of tensors in flux pipeline and vae
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
81f0886d6f
|
Working inference node with quantized bnb nf4 checkpoint
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
2eb87f3306
|
Remove unused param on _run_vae_decoding in flux text to image
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
1bd90e0fd4
|
Run ruff, setup initial text to image node
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
436f18ff55
|
Add backend functions and classes for Flux implementation, Update the way flux encoders/tokenizers are loaded for prompt encoding, Update way flux vae is loaded
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
2d9042fb93
|
Run Ruff
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
9ed53af520
|
Run Ruff
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
56fda669fd
|
Manage quantization of models within the loader
|
2024-08-26 20:17:50 -04:00 |
|
Brandon Rising
|
5f59a828f9
|
Setup flux model loading in the UI
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
1fa6bddc89
|
WIP on moving from diffusers to FLUX
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
f01f56a98e
|
LLM.int8() quantization is working, but still some rough edges to solve.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
99b0f79784
|
Clean up NF4 implementation.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
e1eb104345
|
NF4 inference working
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
a52c899c6d
|
Split a FluxTextEncoderInvocation out from the FluxTextToImageInvocation. This has the advantage that we benfit from automatic caching when the prompt isn't changed.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
eeabb7ebe5
|
Make quantized loading fast for both T5XXL and FLUX transformer.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
3cf0365a35
|
Make float16 inference work with FLUX on 24GB GPU.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
5870742bb9
|
Add support for 8-bit quantizatino of the FLUX T5XXL text encoder.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
01d8c62c57
|
Make 8-bit quantization save/reload work for the FLUX transformer. Reload is still very slow with the current optimum.quanto implementation.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
55a242b2d6
|
Minor improvements to FLUX workflow.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
45263b339f
|
Got FLUX schnell working with 8-bit quantization. Still lots of rough edges to clean up.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
3319491861
|
Use the FluxPipeline.encode_prompt() api rather than trying to run the two text encoders separately.
|
2024-08-26 20:17:50 -04:00 |
|
Ryan Dick
|
b39031ea53
|
First draft of FluxTextToImageInvocation.
|
2024-08-26 20:17:50 -04:00 |
|