CivArchive
    Preview 100066120
    Preview 93840209
    Preview 100093431
    Preview 93840481
    Preview 93839920
    Preview 93839904
    Preview 93840098
    Preview 93840447
    Preview 93840006
    Preview 93839851
    Preview 93839978
    Preview 93840152
    Preview 93840241
    Preview 93840742
    Preview 93840746
    Preview 93843021
    Preview 93840817
    Preview 93843572
    Preview 93844021
    Preview 93844083

    Nunchaku Fluxmania Legacy (by Spooknik)
    Select Pruned model fp16 (6.3 Gb)

    KREAMANIA

    I don't consider Kreamania to be a truly new model, it's just a slight improvement on Legacy, injecting a tiny bit of Krea (which, in my opinion, isn't a benchmark for realism) and adding some of my recent LoRA models. It deserved the name Legacy+ more than Kreamania.

    I find this version provides more detail and improved lighting. The composition remains almost identical to that of Legacy.

    Note : Don't use "freckles" in the prompt

    Oddity : Unexpected appearance of text

    https://civarchive.com/models/1808412/wan-refine-fluxmania

    https://civarchive.com/articles/17346/wan-refining-fluxmania-for-more-realism

    Link

    Quantized versions of the Fluxmania Legacy model:
    https://huggingface.co/belisarius/FLUX.1-dev-Fluxmania-Legacy-gguf/tree/main

    Showcase images are in PNG and include metadata, prompt & workflow.

    All licensing conditions for the Flux Dev model apply. You can find the details here.

    Model orientation : Photography - Realism - Portraits - Artistic nude

    Fluxmania III :

    File 22.07 GB : unet fp16 without VAE & clip

    File 11.16 GB : unet fp8 without VAE & clip
    File 15.91 GB : checkpoint fp8 (not bf16) with VAE & Clip L full fp32

    Settings : dpmpp_2m - sgm_uniform / cfg 3.5 / steps 25 - 30.

    I recommend using this version of Clip L with the Fluxmania model

    Fluxmania I & II : Unet fp8 (no VAE, no clip) / Settings : dpmpp_2m - sgm_uniform / cfg 3.5 / steps 25.

    Genuine : Unet fp16 (no VAE, no clip) / Settings : Euler Simple / cfg 3.5 / steps 30.

    Description

    Recommended settings : dpmpp_2m / sgm_uniform / 20 steps / flux guidance : 3.5

    FAQ

    Comments (127)

    markhassain3712Aug 12, 2025· 4 reactions
    CivitAI

    can u plz convert your lagacy model to nunchaku?

    Adel_AI
    Author
    Aug 12, 2025

    Sorry, I don't plan to do that.

    markhassain3712Aug 13, 2025· 2 reactions

    Adel_AI ohh ok but still i love that model so thnx for that ♥

    kennedysworksAug 13, 2025· 7 reactions
    CivitAI

    This is why we use flux even though there are models with many parameters like qwen and wan. This is because we respond flexibly according to the quality of the fine tuning creator. Thank you adel ai for always revealing amazing models

    SaoshyantAug 13, 2025· 2 reactions
    CivitAI

    I use Mac M1 and always have this problem with Samplers like SamplerCustomAdvanced which doesn't support dtype Float8_e4m3fn to MPS.

    in versions of Fluxmania I had to used "default" dtype weigth instead of Float8_e4m3fn.

    grateful if you could provide a version of fp16 that is suitable for M1.

    Adel_AI
    Author
    Aug 13, 2025· 1 reaction

    A few things to try:

    - Instead of "samplercustomadvanced," you can use either a standard "Ksampler" or "samplercustom."

    - Try using the E5M2 exponentiation.

    - You can use the fp8 version of the model with the "default" dtype.

    - Otherwise, I'm considering making an fp16 version, but not right away.

    SaoshyantAug 13, 2025· 2 reactions

    Dear Adel_AI 
    I tried all the ways but unfortunately it didn't work. :(

    same problem or similar with flux-krea and flux-kontext and very frustrated.

    I hope you will make the fp16 version soon as you came to the conclusion so quickly to release this version.

    Thank you very much.

    alhhentz134Aug 14, 2025· 1 reaction

    Saoshyant  : Try Krea or/and Kontext in GGUF format Works without problems on my M2 Ultra Studio & M4 air Max. For Kreamania, we're waiting for the author to find the necessary time to create either an fp16 version or even better, a quantized GGUF version.

    SaoshyantAug 16, 2025

    alhhentz134 I tried several versions of quantized GGUF. Unfortunately, the one that works without any problems is of very low quality.

    amazingbeautyAug 16, 2025
    CivitAI

    away from your model , what you thinkkrea added over flux dev ? because i don't believe it's just a try for more realistic images , sure it have some other improvements ?

    Adel_AI
    Author
    Aug 16, 2025· 1 reaction

    It is supposed to provide better guidance in addition to more refined realism compared to Flux1.Dev

    ledeni80Aug 16, 2025
    CivitAI

    Is there any way for Kreamania to work with Sdforge?

    Adel_AI
    Author
    Aug 16, 2025

    I am not a user of this interface, but several users confirm that all Fluxmania models work on Forge. Preferably, use models that include Clip and VAE

    ledeni80Aug 16, 2025

    Adel_AI Not my first rodeo. All of the past versions work fine. Kreamania is not the same as them, technically. The same settings give a blank image.

    Adel_AI
    Author
    Aug 16, 2025

    ledeni80 On previous versions, I always provided a "checkpoint" version with VAE and Clip, I think this is the best way to use it with Forge. For Kreamania, for now there is only the fp8 Unet version

    ledeni80Aug 16, 2025

    Adel_AI clip and vae are active from the day one of the installation. Never removed them. Not a single flux model had any problems up until now. This model is the only one I had a problem with. For example, I am using fluxmania_Legacy - Full Model fp8 (11.08 GB), without issues. So I don't know why this won't work.

    BIG_ASep 1, 2025

    Krea works perfectly on Forge. I'm not sure what issues you're encountering.

    lotnickSep 2, 2025

    With Clip and VAE, on Forge give just a black image

    Adel_AI
    Author
    Sep 2, 2025

    @lotnick Sorry, I'm no help with Forge, I've never used that interface.

    aneebartist683Aug 18, 2025
    CivitAI

    I tried to work on your png drag and drop workflow. Trust me my head is spinning. because you are genius I am not. So do you have a simplified version workflow where I can use these models? please

    Adel_AI
    Author
    Aug 18, 2025

    If you use my PNGs, delete or disable all unnecessary nodes to simplify the WF.

    Also, I store my prompts as images, not text, so I use I2T2I instead of T2I; for me, it's simpler and more intuitive. To simplify everything, remove the image loader and keep only the prompt input. If you still can't figure it out, I'll make you a simplified WF. But I think trying it yourself will help you better understand the logic of comfyui and progress.

    axlaimAug 19, 2025

    Adel_AI Hmm, I2T2I! That sounds interesting. Will have to try that. I'm loving t2i2i using sdxl to sd3.5 to flux. Do you use Florence/Joy Caption to set the initial text prompt?

    johndavAug 18, 2025· 6 reactions
    CivitAI

    Bbeen using your models for some time now (since version V) and first I want to say that it is the best i've ever tryied. Amazing work.

    I've been trying Flux Krea as well and comparing some results, Fluxmanias results on hands and feet are a bit worse. Do you have some tips for this? Maybe changing the sampler, steps, etc (I'm using the recommended settings)

    Adel_AI
    Author
    Aug 18, 2025

    I've always had pretty poor rendering for the feet, especially when the sole of the foot is facing up.

    I haven't particularly noticed any problematic rendering frequency for the hands.

    You could perhaps try res_2s/bong_tangent or heun2/sgm_uniform

    But the best thing would be to vary the seed until you get an acceptable result.

    I'll check that out myself.

    Thanks for your feedback.

    johndavAug 19, 2025

    Adel_AI I'll try out those combinations, thanks!

    axlaimAug 19, 2025
    CivitAI

    CUBLAS_STATUS_NOT_SUPPORTED

    I get this error when running Kreamania in comfy. No issues with any other versions of Fluxmania, or any other models: Flux, Chroma, SDXL, etc.

    Updated comfy, updated all but still the same.

    Adel_AI
    Author
    Aug 19, 2025

    Hi, I asked an AI assistant to help investigate the CUBLAS_STATUS_NOT_SUPPORTED error you're encountering with the Krea version of Fluxmania in ComfyUI. Here are a few suggestions you can try on your end:

    🛠️ Possible Causes & Fixes

    Model-specific operations: Krea might include a custom module or tensor operation that triggers a cuBLAS call not supported in your current setup — even if previous FP8 versions worked fine.

    CUDA/cuBLAS version mismatch: Ensure you're using CUDA 12.2 or newer. FP8 support is still evolving, and certain operations may fail silently on older versions.

    PyTorch build differences: Some builds of PyTorch (especially those without Triton or xformers) may not fully support FP8 operations. Try updating PyTorch or using a build with extended support.

    Driver version: Make sure your NVIDIA driver is up to date (≥ 535.xx). Older drivers can block FP8 execution paths.

    Fallback failure: If Krea uses FP8 without fallback to FP16/FP32, and your GPU or environment can't handle the operation natively, cuBLAS will throw that error.

    ✅ What You Can Try

    Force FP16 loading: If possible, modify the model loading script to cast tensors to float16 instead of float8.

    Test outside ComfyUI: Try loading the model in a minimal PyTorch script to see if the error persists. That can help isolate whether it's a model issue or a ComfyUI integration problem.

    Compare with working FP8 versions: If you have older Fluxmania FP8 models that worked, compare their structure or quantization method with Krea. There might be a subtle difference.

    Era1701Aug 20, 2025· 7 reactions
    CivitAI

    Absolutely wonderful work! It has demonstrated the enduring vitality of flux. However, could you please release the KREAMANIA fp16 model for those who have larger VRAM? FP8 will inevitably cause some loss in precision. Thank you very much!

    Adel_AI
    Author
    Aug 20, 2025· 4 reactions

    Thanks for your feedback, I plan to publish the fp16 in a few days. A lot of work to finish beforehand.

    BIG_ASep 1, 2025· 6 reactions
    CivitAI

    For creating realistic or photorealistic photography-style works, krea is recommended. For creating artistic illustration-style works, legacy is recommended.

    I'm not sure if the OP has similar reminders, but this is my actual experience after using it for a few days.

    Adel_AI
    Author
    Sep 1, 2025· 1 reaction

    Thanks for sharing your experience.
    Not sure if by “Krea” you meant “Kreamania,” the latest version of Fluxmania, compared to the previous “Legacy” one. Personally, I find both versions equally versatile, whether you're going for photorealism or a more artistic, illustration-style result. They both respond well depending on how the prompt is crafted and the creative direction you're aiming for.

    BIG_ASep 2, 2025· 1 reaction

    @Adel_AI This view is purely based on personal experience and carries significant subjective limitations. However, it is a fact that both versions are indeed excellent base models. Thank you for your explanation.

    Mr_FeiSep 3, 2025· 4 reactions
    CivitAI

    Does KREAMANIA not have an FP16 version?

    Adel_AI
    Author
    Sep 3, 2025· 1 reaction

    not yet

    amrpp79697Sep 9, 2025· 4 reactions
    CivitAI

    Are you going to release Kreamania in FP16 soon ? I have been using your Legacy FP16 model for quite sometime now on my 4090, its running great and quality is absolutely amazing. I would like very much to have your new Kreamania model in FP16 as well.

    craigmack34Sep 12, 2025· 7 reactions
    CivitAI

    Please. How do I remove that green dinosaur bs from my UI? Wtf?

    Adel_AI
    Author
    Sep 12, 2025· 5 reactions

    In the comfyUI settings,
    select the "KayTool" extension menu

    In the "GuLuLu" submenu, disable "GuLuLu Stay with Me"

    craigmack34Sep 14, 2025· 1 reaction

    @Adel_AI Thank you, Sir.

    ScratchyStarSep 13, 2025· 2 reactions
    CivitAI

    really great for arty nudes ! what loras would you suggest to enhance body shape, breast, cuteness, without changing to much model identity, especially grain and lighting ? thx for all

    Adel_AI
    Author
    Sep 13, 2025· 1 reaction

    I've finalized some nude-oriented LoRAs:

    https://civitai.com/models/1344651/nude-art

    https://civitai.com/models/1253652/red-district

    Thanks for your feedback

    ScratchyStarSep 13, 2025· 1 reaction

    @Adel_AI thx for advice, Ill check this out

    rlewisfr346Sep 14, 2025· 1 reaction
    CivitAI

    This model (Fluxmania-Krea) does not handle character LORAs well at all. Even LORAs trained on Krea make the model completely fall apart.
    EDIT: works OKish with Flux trained LORAs. I was trying to force a Krea trained LORA onto this, which is not a Krea checkpoint.

    Adel_AI
    Author
    Sep 17, 2025· 1 reaction

    This remark is not limited only to Kreamania but to all versions of Fluxmania, I have often been told about problems with character loras. Personally I do not use this type of Lora, I have not been able to see it for myself. I tried to finalize a model as versatile as possible, but that there are gaps, it seems normal to me

    GarbaggioTunaSep 22, 2025· 1 reaction
    CivitAI

    I've been using the 22G legacy for some time now, no issues, Kream runs out of memory (Forge.) Just FYI

    Adel_AI
    Author
    Sep 22, 2025· 1 reaction

    Sorry to hear Kreamania FP16 isn't working on your setup with Forge anymore. Just a tip—might be worth trying ComfyUI instead. It handles VRAM much more efficiently in my experience.

    GarbaggioTunaSep 23, 2025· 1 reaction

    @Adel_AI yeah i have a problem with the interface but will have to uh...FORGE through it eventually lol.
    Strange tho, seems as if the two checkpoints are the same physical size and i have a 3090 with 24G. Ah well, time to spend MOAR :)

    Adel_AI
    Author
    Sep 23, 2025· 1 reaction

    @GarbaggioTuna I have a 3070 8Gb, the 22Gb models work normally, with a little slowness, but without memory errors. This is one of the advantages of ComfyUI

    GarbaggioTunaSep 23, 2025· 1 reaction

    @Adel_AI I'm not concerned i am fine with Legacy, just thought it might be indicative of something for you. :) BTW no lora's involved...can give you the exact errors if ever interested.

    Adel_AI
    Author
    Sep 23, 2025· 1 reaction

    @GarbaggioTuna I don't use Forge and don't recommend it. I can't be of any help to anyone who encounters issues with this interface. Unfortunately, it's impossible to cover all the test areas.

    Thanks again for your feedback.

    GarbaggioTunaSep 23, 2025· 1 reaction

    @Adel_AI Gotcha, i understand totally. I just find that when something acts up between versions it might tell me something, but i have not done this kind of thing! ANYhoo, THANKS man i love your stuff.

    6vidit9Oct 3, 2025· 3 reactions
    CivitAI

    Mind-blowing checkpoint❤️

    Adel_AI
    Author
    Oct 3, 2025· 2 reactions

    Thanks for your feedback

    6vidit9Oct 3, 2025· 2 reactions

    @Adel_AI Legendary work, man. All your LoRAs, especially this checkpoint, are top-tier—some of the best Flux models on CivitAI.

    13_februaryOct 8, 2025· 1 reaction
    CivitAI

    When using ComfyUI with the fp8 version, Kreamania produces an entirely black image.

    Adel_AI
    Author
    Oct 9, 2025· 1 reaction

    There is probably a problem with your WF.

    wiferatorOct 9, 2025· 1 reaction
    CivitAI

    any chance of a Quantized version

    Adel_AI
    Author
    Oct 10, 2025· 1 reaction

    Sorry, not for now

    wiferatorOct 19, 2025· 2 reactions

    @Adel_AI early Christmas!

    TheHorrorStoryMakerOct 19, 2025· 2 reactions
    CivitAI

    this is fantastic thank you. if it isn't much trouble. may i ask for your workflow for the legacy checkpoint? when i drag the images posted for it. the workflow seems to be using different models. thanks again.

    Adel_AI
    Author
    Oct 19, 2025· 1 reaction

    You mean that in the summary, Legacy is mentioned as a checkpoint but in the workflow, it's a different model? Give me an example please

    TheHorrorStoryMakerOct 19, 2025· 1 reaction

    Thank you for responding, i meant you have a lot of versions right? so when i select any version it shows different show case images. supposedly for this current selected version. so i select legacy which i downloaded. then i downloaded an image from the show case ones, dragged it into comfy. but the workflow has a diffusion model loader with totally different model. i really want to use your workflow for legacy version because overall quality is incredible. and all the workflows i tried so far gives inferior quality to the show cases you posted. thanks again man really appreciate it <3

    Adel_AI
    Author
    Oct 19, 2025· 2 reactions

    @TheHorrorStoryMaker I think I can pinpoint the issue. You noticed that I use an image for latent injection, and if you’re not using the same one, your results will inevitably differ. Starting from an empty latent never yields the same level of detail or rendering quality. The image that provides the latent isn’t included in my workflow, which explains the gap between my output and yours.

    Latent input plays a crucial role in the rendering process. To improve your results, use my final image—or something visually similar—as your latent source. That should bring your renderings much closer to mine and significantly enhance their quality

    TheHorrorStoryMakerOct 19, 2025· 1 reaction

    @Adel_AI i had no idea. will do that from now on. one other thing kindly. with the legacy version do you still recommend using the --clipLCLIPGFullFP32_zer0intVisionCLIPL-- file? or use the one in the checkpoint same as vae? thanks again so much. you're very kind.

    Adel_AI
    Author
    Oct 19, 2025· 2 reactions

    @TheHorrorStoryMaker use Clip L and the VAE recommended for Flux1.Dev, no need to look for exotic versions, as I did at the beginning

    koopa990Oct 20, 2025· 1 reaction
    CivitAI

    Does know or have a workflow specifically made for Fluxmania? I cant use the standard Flux D templates cause they DO NOT require negative text clip.

    Adel_AI
    Author
    Oct 20, 2025· 1 reaction

    use my workflows

    koopa990Oct 20, 2025· 1 reaction

    @Adel_AI Wan refine fluxmania"?

    Adel_AI
    Author
    Oct 20, 2025· 1 reaction

    No, this is a two-pass workflow with refining, you don't need it.

    All my renders include their WF, you just have to drag the image onto the comfyui interface to retrieve the WF.

    If it seems complicated, delete anything that doesn't seem important.

    koopa990Oct 20, 2025· 1 reaction

    @Adel_AI im confused. I dragged the Fluxmania V image to my comfy; says unable to find workflow

    Adel_AI
    Author
    Oct 20, 2025· 1 reaction

    @koopa990 Get a recent one from Kreamania. Later you can change the Kreamania model to version V.

    koopa990Oct 22, 2025· 1 reaction

    @Adel_AI All of these missing nodes and such seems to confuse me. I dont want to use all of these things... unless this workflow only works with fluxmania

    012egisOct 22, 2025· 2 reactions
    CivitAI

    great model (Legacy) and hoping for a chroma version

    Adel_AI
    Author
    Oct 22, 2025· 3 reactions

    A Chroma version will require months of work. Converting all my Fluxmania-component Lora models to Chroma format, is a huge investment of time I don't have.

    Thanks for your feedback.

    012egisOct 22, 2025· 2 reactions

    @Adel_AI I understand. For what it's worth, The flux loras I've trained using sd-scripts still retain character likeness when applied to chroma-hd, without any kind of conversion required

    coochieOct 27, 2025· 2 reactions
    CivitAI

    Could I trouble you for a Q8 version of kreamania? Thank you 🙏

    Adel_AI
    Author
    Oct 27, 2025· 2 reactions

    I'm really sorry, but due to lack of time, this is not possible at the moment.

    gemstonebroOct 29, 2025· 2 reactions
    CivitAI

    Probably best thing so far!

    narcibrookNov 4, 2025· 1 reaction
    CivitAI

    Hello, I like your model, but I'm having trouble removing vertical artifacts, especially on the right edges of the image... do you have any advice? Thanks ^^

    Adel_AI
    Author
    Nov 4, 2025· 1 reaction

    Hey, Yes, indeed, the banding artifact can sometimes appear, especially when using certain LoRA. I've tested several approaches, and this one seems to be the most effective:
    https://civitai.com/articles/16849/smoother-renders-with-fluxmania-a-simple-trick-to-eliminate-banding-and-artifacts
    Thank you for your feedback.

    gemstonebroNov 7, 2025· 3 reactions
    CivitAI

    Discovered far better sampler! Aparently this guide says dpmpp2m sgm is recommended. i Use fp16 and somehow did euler_ancestral linear_quadratic and its insane! its night but i will try test compare tomorrow same prompts etc..

    JohnRohanNov 9, 2025· 2 reactions
    CivitAI

    Works well by itself, but for some reason no Loras will work with this model (kreamania version). Whether I use Loras or not, the image is unchanged. But meanwhile Loras work fine on all my other models. Has anyone else had this problem? Does anyone have any idea what could be going on?

    Adel_AI
    Author
    Nov 10, 2025· 1 reaction

    I use the Kreamania version with both my own LoRAs and others, and there are no problems.

    However, the Fluxmania models (including Kreamania) express little or nothing at all, especially character and celebrity LoRAs.

    JohnRohanNov 11, 2025· 1 reaction

    @Adel_AI it's just bizarre, I have tried other Flux models and they work with Loras. Just not this one. Even if the Lora was weak, I would see at least some difference in the image. I must be missing something but I don't know what.

    Adel_AI
    Author
    Nov 11, 2025· 1 reaction

    @JohnRohan Yes, it's true. Fluxmania models integrate several LoRAs into their merge (Kreamania uses more than 17 LoRAs), resulting in heavily loaded keys that can sometimes introduce artifacts. Therefore, using this type of model leaves a limited scope for LoRAs in inference. Nevertheless, some LoRAs find an interesting field of expression.

    TribalDiffusionNov 14, 2025· 1 reaction

    @JohnRohan Does GGUF: Fluxmania III have this same (personal Lora) problem?
    By GGUF: Fluxmania, I mean this > https://civitai.com/models/1090700/gguf-fluxmania-iii

    JohnRohanNov 23, 2025· 1 reaction

    @GmeJee Sorry took awhile to answer. I just downloaded fluxmania III and tested it out both with and without a Lora using the exact same parameters otherwise. Yes, it responds to Loras.

    Then I tried the same thing with the Kreamania version once again, and sure enough it still didn't respond. I get good results, it just won't react to Loras.

    PrimaveriNov 23, 2025· 1 reaction
    CivitAI

    A Nunchaku version of this checkpoint would be awesome!

    SaoshyantNov 25, 2025· 4 reactions
    CivitAI

    ready for Flux 2, dude? :)

    ironcloudNov 27, 2025· 4 reactions

    @Adel_AI lots of people will be waiting =)

    Adel_AI
    Author
    Nov 27, 2025· 4 reactions

    I can't train Flux2 locally—not even Flux1.

    I used to train LoRAs on Civitai, using Buzz credits, thanks to the generous support from many of you.

    To do it now on ComfyUI Cloud, I’d need a Pro subscription at $100/month.

    Using platforms like Runpod would eat into the credits I reserve for professional work.

    Creating LoRAs is a pleasure, not a profit-driven activity—but spending money on it just doesn’t make sense.

    If Civitai ever offers LoRA training for Flux2, it could become a viable path again.

    Thank you all for your support—it means a lot.

    kennedysworksNov 27, 2025· 3 reactions

    @Adel_AI  Thanks to you, I got to know the possibility for ai art and I would like to work in that field when I finish my studies at the university. Please get coffee. I am always willing to apply

    Adel_AI
    Author
    Nov 27, 2025· 2 reactions

    @kennedysworks What are you studying?

    kennedysworksNov 27, 2025· 2 reactions

    @Adel_AI  I'm majoring in video and film. In Korea, people can't find a job if they can't use AI in videos, movies, and visuals

    kennedysworksNov 27, 2025· 2 reactions

    @Adel_AI  I started ai at a late age while working, but I think this is the future

    Adel_AI
    Author
    Nov 27, 2025· 2 reactions

    @kennedysworks I considered it too, actually — but in the end, I chose a scientific path. Funny enough, my daughter ended up studying film.

    I had the chance to visit Korea (Seoul, to be precise) and I absolutely loved it. The country is stunning, and the people are incredibly kind.

    LeeAeronNov 26, 2025· 1 reaction
    CivitAI

    KreaMania fp16 didn't work on ForgeSD - engine or braking and end without any error or making PC BLUE DEADSCREEN

    Adel_AI
    Author
    Nov 26, 2025· 2 reactions

    Apparently, there are even documented cases of PCs spontaneously combusting thanks to Kreamania. But hey, that's just part of the thrill, right? Who needs safety when you’ve got AI pushing the limits of thermodynamics?

    LeeAeronNov 26, 2025· 1 reaction

    @Adel_AI what termodynamics if I've tried to generate 'on cold', no any temperature heating was at all while this!

    Adel_AI
    Author
    Nov 27, 2025· 1 reaction

    @LeeAeron I was joking, I'm not a Forge user, sorry I can't be of any help.

    Randy_MoreholeDec 15, 2025· 1 reaction

    blue-screen of death is usually a memory problem, or at least it used to be!

    datedman353Jan 10, 2026· 1 reaction

    works well with Forge Neo

    LeeAeronJan 11, 2026· 1 reaction

    @Randy_Morehole have no probs with 99.99% rest models. it's about model prob

    datedman353Jan 16, 2026· 1 reaction

    @LeeAeron what GPU? and what VAE/etc?

    LeeAeronJan 17, 2026· 1 reaction

    @datedman353 man, it's not about gpu or ave/etc. IRest Fllux.1 models (99.9%) works fine, just this makes image into bad. Forget, I'vesimply deleted this model, cuz I don't need it in this view.

    jsnmdrs592Nov 30, 2025· 1 reaction
    CivitAI

    any reason why this loads in invokeai the errors out? Says unexpected keys.. literally was looking at pics 20min was excited to try than that. Prob something I'm doing wrong who knows. Your loras work Checkpoints error

    Adel_AI
    Author
    Dec 1, 2025· 1 reaction

    Sorry, I can't help you, I don't use Invoke.

    mystifyingDec 7, 2025· 1 reaction

    invoke was always an issue with models, because of how they check and verify the model to be used in the ui

    have you tried any other flux model fine-tunes?

    jsnmdrs592Dec 8, 2025· 1 reaction

    @mystifying  Its not the verification. That happens during the adding of the model and 6.9 allows you to fix as well. The model adds to library no issue, it's when I run it is when I get the errors.

    mystifyingDec 8, 2025· 1 reaction

    @jsnmdrs592 the new one invoke version might have gotten past the first part but it sounds like it still connected to that issue, i recall every-time a new model was introduced to the system, tons of people having issues, maybe in a few versions they work out issues i felt were there to force people to to get the paid version, and that team has moved on that was in charge of development during that time

    diffuse11390Dec 2, 2025· 1 reaction
    CivitAI

    I have been unable to train this model on something like FluxTrainer, which works great for Flux-Dev and Flux-Krea. Any reason this might not work, or any work around to make Lora's with it?

    ironcloudDec 11, 2025· 3 reactions
    CivitAI

    @Adel_AI

    Strange, I get error with KREAMANIA fp16 after updated ComfyUI to 0.4. Legacy fp16 works just fine.

    SamplerCustomAdvanced
    FP8 scaledmm failed, falling back to dequantization: Both scale_a and scale_b must be float (fp32) tensors.

    Any Idea? I was going to create an issue in Comfy's Github, but it seems that's the only model that's problematic. Wanna ask you first.

    Adel_AI
    Author
    Dec 11, 2025· 1 reaction

    The response I received via ChatGPT :
    The error you’re seeing is due to changes in ComfyUI 0.4. The FP8 scaled_mm operation now requires both scale_a and scale_b to be FP32 tensors. In earlier versions, workflows could pass FP8 or int scales without issue, but the new runtime enforces stricter type checks.

    Workarounds:

    Update your custom nodes/wrappers (many have already patched this).

    If you maintain code, cast the scales before the call:

    scale_a = scale_a.to(torch.float32)

    scale_b = scale_b.to(torch.float32)

    Or temporarily roll back to ComfyUI 0.3.x until the nodes are updated.

    It’s not a bug in the model itself, but a compatibility change introduced in ComfyUI 0.4.

    ironcloudDec 11, 2025· 1 reaction

    @Adel_AI I see~ I guess I'll let the comfy team know about this. thx~

    ironcloudDec 11, 2025· 2 reactions

    Curious, what's Kreamania differ from Legacy though (since Legacy works perfectly fine), and I'm using the fp16 version not the fp8, why does it complain about fp8?....Obviously I'm not technically savvy enough to get all these.... >_< I created an issue on Comfy's Github...hopefully this is an easy fix.

    Adel_AI
    Author
    Dec 11, 2025· 1 reaction

    @ironcloud There's very little difference between Legacy and Kreamania (3% Krea + 3 LoRA). I ​​suspect the conversion from FP8 to FP16 is the cause of this incompatibility with comfyui 0.4.

    ouzhen123456990Dec 14, 2025· 2 reactions

    @Adel_AI Legacy and Kreamania的生成差别我觉得很大,Kreamania的图片质量和效果基本是一个档次的提升。当然,我测试的结果是这样觉得的。所以很可惜我这里Kreamania已经没法用了。

    Adel_AI
    Author
    Dec 14, 2025· 2 reactions

    @ouzhen123456990 是的,我同意,构图相同,但做工和细节却截然不同。

    sharifahnasuha193587Dec 19, 2025· 1 reaction

    @ironcloud any fix?

    ironcloudDec 20, 2025· 1 reaction

    @sharifahnasuha193587 Nop, not yet, there are many things going on with ComfyUI now, especially with all the new things introduced that come along with many issues and user feedback...a single model incompatibility is probably the least of their concern at the moment. But you could comment on https://github.com/Comfy-Org/desktop/issues/1476, mention that you're sharing the same issue, more people talk about it could help getting their attention.

    ironcloudJan 21, 2026· 1 reaction

    @sharifahnasuha193587 @Adel_AI it's working now, with comfyui v0.9.1

    ouzhen123456990Dec 12, 2025· 3 reactions
    CivitAI

    你的模型我感觉是全网最好的。非常棒!

    Adel_AI
    Author
    Dec 12, 2025· 1 reaction

    感谢您的反馈

    brownbagel0Dec 28, 2025· 1 reaction
    CivitAI

    Does not want to support my LoRAs. Generations turn out all black.

    Soup_FusionFeb 11, 2026· 3 reactions
    CivitAI

    hey, i know you're working on ZiT/ZiB rn but i was wondering when we could expect a Fluxmania 2? it was my personal favorite for flux 1 and i can't wait to see what you can do with Klein.

    Adel_AI
    Author
    Feb 11, 2026· 3 reactions

    Very disappointed with Flux2-Dev it's nothing like the quality of Flux2.Pro. BFL seems to focus solely on its proprietary model. I haven't tested Klein, but I assume it's of the same quality as Dev.

    I think the future lies with Z Image and Qwen Image; I'm currently looking at those models. Thanks for your feedback and appreciation of my work.

    Soup_FusionFeb 11, 2026· 3 reactions

    @Adel_AI so does that mean there won't be a Fluxmania 2? really sad if that's the case but i respect your position. I'd recommend trying Klein out, i was also disappointed at flux2dev, mainly for it being so unnecessarily bloated for what it offers, but Klein 9b is actually quite good. i was all in on z-image but ZiB kinda disappointed me. 9b is just giving me better quality out of the gate compared to ZiB.

    Checkpoint
    Flux.1 D

    Details

    Downloads
    10,248
    Platform
    CivitAI
    Platform Status
    Available
    Created
    8/12/2025
    Updated
    5/16/2026
    Deleted
    -

    Files

    fluxmania_kreamania.safetensors

    Mirrors

    fluxmania_kreamania.safetensors

    Mirrors

    Available On (3 platforms)

    Same model published on other platforms. May have additional downloads or version variants.