Flux Blockwise (Mixed Precision Model)
I had to build several custom tools to allow for the mixed precision model, to my knowledge it is the first built like this.
Faster and more accurate then any other FP8 quantized model currently available
Works in Comfy and Forge but forge needs to be set to BF16 UNET
Comfy load as a diffuser model USE DEFAULT WEIGHT
FP16 Upcasting should not be used unless absolutely necessary such as running CPU or IPEX
FORGE - set COMMANDLINE_ARGS= --unet-in-bf16 --vae-in-fp32
Other then the need to force forge into BF16, (FP32 VAE optionally) it should work the same as the DEV model with the added benefit of being 5GB smaller then the full BF16
It turns out that every quantized model including my own up to this point to my knowledge has been built UN-optimally per blackforest.
Only the UNET blocks should be quantized in the diffuser model, also they should be upcast to BF16 and not FP16 (Comfy does this correctly)
I am currently trying to workout how to follow Blackforest recommendations but using GGUF
Description
FAQ
Comments (2)
I TRY THIS MODEL A LOT TIME .There is not any different with the fp8 dev .even use the different clip .WHY ? do something wrong?
The only difference is that the UNET blocks are in FP8 and the UNET Text blocks are in BF16 - you would need to run the model in BF16 to notice the enhanced accuracy with the 5GB or so reduction in size