CivArchive
    Electric Dreams - Electric Dreams v0.2
    NSFW
    Preview 2124155
    Preview 2124157
    Preview 2124159
    Preview 2124158
    Preview 2124156
    Preview 2124165
    Preview 2124168
    Preview 2124167
    Preview 2124166
    Preview 2124169

    Early version of my planned base model.

    Description

    FAQ

    Comments (8)

    YamerAug 21, 20233 reactions
    CivitAI

    This model is filled with junk data, SDXL models follow a fp16 structure, making it fp32 just adds a layer of unused data. Next time make the model fp16 only, there is no need for fp32. 馃憤

    GreyModeler
    Author
    Aug 21, 20236 reactions

    I just want to address this and your other comments as they are factually incorrect. While you can run FP16 fine, FP32 is not "filled with junk data" -- If you convert an FP16 to FP32 it will not increase anything but the size as the increased precision cannot be generated. With that said however if you are training FP32, and using FP32, it maintains actual used data. The reason this is FP32 is that it IS FP32. It can be converted to FP16 with loss, but it is not an FP16 model. You should avoid providing factually incorrect information as it will lead to propagation of misinformation.

    sikasolutionsworldwide709Aug 21, 20231 reaction

    @GreyModeler聽This is correct so far as I know to date

    Disty0Aug 23, 20232 reactions

    SDXL Base 1.0 model has 6 versions;
    FP16 Safetensors,
    FP16 Diffusers,
    FP32 Diffusers (Original),
    FP32 ONNX,
    FP32 OpenVINO.

    SDXL is not an FP16 model, It's converted to FP16.
    Only Safetensors version doesn't have the raw FP32.

    But i agree with FP16 version.
    FP16 version should be released alongside the FP32 since FP32 will get converted to FP16 anyway.

    GreyModeler
    Author
    Aug 23, 20232 reactions

    @Disty0聽Thanks for the clarification there -- I only had the original diffusers so didn't even know there was an FP16 release. With that said my training workflow is all switched to FP16 now, so v0.3 will be trained and released in FP16 exclusively. Also yes, I really should add an FP16 version for v0.2, just been preoccupied with training woes.

    incorrect statement

    MrNoirAug 30, 2023

    @GreyModeler聽i didnt understand, so we train b32 or b16?

    GreyModeler
    Author
    Sep 1, 2023

    @MrNoir聽That gets even more complicated -- While the v0.1 and 0.2 are full FP32, v0.3 and onward are different. For unet I'm maintaining FP16 weights, for the text encoders (which I am actively training unlike, to my knowledge most others training at this time [with good reason, they're a pain]) I'm using FP32 weights (I may still pull this down to FP16 in future). With that said however, there's model weights and training precision, as well as more data types for each, all with their own tradeoffs. There's no one right answer, but frankly from my testing, if you use FP16 or BF16 you're going to be fine, you don't generally need the precision of FP32, and I am not seeing the value of it vs the overhead in time and vram required to use it. With that said, BF16 while possibly better is problematic with some optimizers.

    Checkpoint
    SDXL 1.0

    Details

    Downloads
    297
    Platform
    CivitAI
    Platform Status
    Available
    Created
    8/20/2023
    Updated
    5/13/2026
    Deleted
    -

    Files

    electricDreams_electricDreamsV02.safetensors