Hey everybody!
I'm really excited to share my new model with you all!
I decided to separate this checkpoint from the rest of my models because it's built in a very different way from the others and I think it's the best model I've made so far. I merged 3 models (Plant Milk: Walnut, Vete ┃ Dreamland ILL, and Copper Thyme) then fine-tuned that model with 500 images using OneTrainer. It wasn't very stable though so I merged it back in with the model I had trained it on and refined it by merging again with an older checkpoint I had abandoned.
The style is very different from what I intended but I really like the results. Comtesse Raffinée translate to English as Refined Countess. I named it that because the style reminded me of Duchesse but it's more elegant, refined and flexible. Although it has a dramatic and cinematic style of its own it still works well with artist and style tags as well as LoRAs.
Image Settings
Sampler: DPM++ 2M Karras / Euler A/ Restart
Steps: 25–50 (base), 30–50+ for Hi-Res Fix
CFG: 3–7
Clip Skip 2
Description
FAQ
Comments (43)
Can't wait to check this out after work! Sample images look awesome!
Thanks! I hope you like it
Your showcase images are always so badassily beautiful. Methinks it's worth it to come up with a new checkpoint each time, just to showcase them. ;)
Do you have a specific pitch for this model (like, Duchesse geared for artistic ILL, and Chaotix for stylistic range), or it's freeform art-sans-borders?
(I'm wondering if googling Raffinee will give me any answers 😅)
Raffinee is french for refined. Which I guess is the pitch for this one. I wasn't sure how to do the sales pitch, but what I like about it is that it has a strong style but it's still very flexible. It does most things well from what I've seen so far, with the occasional problem of a missing finger or toe.
@Chumple Ok! that sounds promising (especially given the context of what I know of your standards from your previous checkpoints...)
Now I have to try it out for myself; see if I agree ^_^
And thx for the explanation
(I think I need to learn French, you're not the only one on this site suddenly pulling out Frenchesian terminology, and anyway I'll definitely sound more like an Artiste (TM)
@ravemry9 I don't even speak French, I just kind of fell into naming my checkpoints French by accident. I got the name Melange from GPT and Duchesse was playing off DucHaiten but I like that it makes my models sound fancier and classier than I am 😅.
I hope you like the model!
@ravemry9 I updated the description to add some of the sales pitch. Thanks for the idea!
@Chumple A pleasure to be of service to Her Highness the Comtesse in any way.
As long as my part can be in English :P
I'm so sad.
At first, they said if I knew some basic English words, I could start drawing, and I was studying really hard. Now, French? What's next, Chinese?
@vibhisana Chinese probably would be very helpful to learn for the newer models 😅. But I'm not smart enough for that so I'll stick with knowing one language and googling the rest
@vibhisana Yeah, I feel for you.
I have it even worse, I was hoping if I become one of those devil-may-care Artiste types I wouldn't even need to know English...
(the biggest change of the "AI Art revolution" is that now artistry works thru language and you can't flunk basic schooling anymore) :D
Hey! I see in the model notes
<Steps: 25–50 (base), 30–50+ for Hi-Res Fix>
I always just hiresfix on the same step settings, out of ignorance. Is the additional steps a thing?!
And can anyone point me to a (non-video) primer for hires fix settings? I've been experimenting, I have like 5 articles drafted on my test results, but never saw (good) data/tips anywhere. But I had no clue where to look! :-/
I'm no expert (on anything really 😅) but I usually do 50 steps for Hi-Res fix. I think it helps with refining details but I've never done any real A/B testing. If I like an image I might do a low noise with DPM++2M, but if there's things I want to change I sometimes adjust the prompt, increase the noise and use Euler A.
@Chumple Hm. Got it, makes sense.
Why do you prefer EulerA when you're trying for changes?
(I guess I'd think yeah, it's more predictable so more reliably stick to the base image composition while editing it. Or maybe does it handles changes better? or needs less steps or smth?)
@ravemry9 Ancestral samplers don't converge they inject noise every step so you can have more changes at a lower noise when using it. It can also cause more problems though, extra fingers, changing the shape of the head, or adding an extra nose or eyebrow.
DPM++2M doesn't inject that noise. It converges which means it only removes noise to get to a final image. So it won't cause those deformities unless the noise is higher but it also won't make other changes you might want to make during the hi-res fix.
@Chumple Oh! very useful pointer, I really need to be doing my hires fixing like this too.
Do you have a preferred denoise level for when you like the image and are only trying to polish it without changes? I start at 0.40, but really prefer lower (0.25 thereabouts) to keep the original look. But I've always used EulerA, and anyway i'm a noob.
@ravemry9 I don't have a preferred denoise level but I usually start at 0.4 as a default. But if there are details I really like or image is a portrait or something like that I lower the noise to 0.2 or lower if I have to. I generate at home so I sometimes do a bunch of img2img upscales with different noise levels or changes to the prompt until I get something I like
Here are my two cents. I saw you asked before how to maintain composition when upscaling, but I was busy testing something at the time (which failed), so I didn't reply.
First, it seems you're using the online generator, which is quite different from ComfyUI, which I use. So ultimately, this probably won't help at all.
Regarding the steps, after 30 steps, there's limited changes for details; after 40 steps, it's almost meaningless. But I've heard that dpm_2(a) works very well even up to 60 steps; you could give it a try.
In short, regarding composition, low resolution->high denoise. I usually use 0.75 denoise on 1.50x upscaling (832*1216->1248*1824), combined with a dpmpp series sampler, to obtain rich details and a clear, sharp output. Recently, I tried 1.75x (1456*2128) and 2x (1664*2432) upscaling and found that 0.60 denoise is the optimal choice.
There was a blurring issue at low denoise, possibly because I used different samplers when upscaling, but I'm currently happy with the settings. Believe it or not, in many cases, 0.75 denoise helpes to correct extra/missing fingers.
Let me give you an interesting example: See the long black thing above the white-haired girl with the red mask? That's actually a masked head. I forgot to use "solo" in the prompt, so the AI drew two girls and three heads. I loved the composition so much that I first tried high denoise, then gradually lowered it, and finally removed the extra head when 0.50 denoise.
I don't use upscale models, and in my 20+ tests, they gave different results and were not better. I would like to hear your opinions.
Regarding the test I mentioned earlier, I tried replacing the 2.00x upscaling with 1.15x upscaling * 5 (2.011) combined with low deniose, hoping to obtain rich details while maintaining the composition, but it failed. The problem isn't that "if the distance between each step is short enough, the AI can walk steadily," but rather that "if the total distance is long enough, the AI will eventually fall." In short, the AI doesn't know how to correctly draw at high resolution, and always ends up with extra limbs.
You mentioned five articles, which I'd love to read. I've always enjoyed manipulating different parameters; it's like gambling—you never know which combination will win you the lottery.
@vibhisana Thanks! your answer is packed with lots of useful info.
Despite the civit on-site generator having different settings than Comfy, I think a lot of it is relevant. I'll need to spend time experimenting before I can provide a more solid response.
You say you enjoy playing around with gen parameters it's like gambling,and yeah me too -- this type of experimentation is a major part of what I love with AI gen. But IMO it's much cooler than gambling, it's really like engineering/architecture, except the fractal nature of AI token parsing makes the engineering much less predictable. Infinite fun.
@ravemry9 Yes, it's way cooler than gambling. I was addicted to the Diablo series and Path of Exile for years, and those gear-farming games were essentially gambling. Last year, when I started drawing, it suddenly dawned on me that in games I killed/destroyed things for fun, while in drawing I created beautiful things. txt2img has truly made me a better person, and I've never looked back.
Sometimes, less predictable is good, like this and this. In my prompt, the boy should hold flowers, front view, at middle of frame, girl at left side of frame. The result is totally irrelevant, but what can I say? They're happy! God of AI works in mysterious ways. AI understands what beauty is; I'm simply working with it.
@vibhisana Oh, well put about the fun of AI being in creation, and in beauty.
"The God of AI" , lol. It's definitely true AI has a mind of its own and some of our best results are unplanned.
I don't rmemeber which ones now, but at least twice i got amazing gens that didn't replicate when i realized i had misspelled important parts of the prompt. Fixing the typos took all the magic out of the image.
Those images you link are beautiful works of art. Hearing the story behind how you genned them is really interesting.
@ravemry9 , Let me show you another miracle created by AI. Once I wanted to draw one girl, play piano on beach, ocean, whale jumping in background. If all goes well, it sounds good, right? I cut and pasted some words to adjust the sequence, didn't realize I missed the beach and ocean parts, and ran the generator directly. This is the result. AI turned the impossible into beauty, I lack that creativity, that vision, and that taste. This is the 24th pics we create together, and since then, I've known I'm merely a humble explorer, searching for undiscovered beauty created by AI.
@vibhisana Damn that image with the piano is awesome! I usually do a 2x upscale and img2img pass with my images and then use the SeedVR2 upscaler to get them to 4k.
@ravemry9 Sometimes typos can lead to really interesting results. Like in this image I accidentally type "moth" instead of "mouth." 😅. I've also had similar happy accidents when using detailers/inpainting. This image is my favorite example of that
@Chumple , I'm thinking about a lazy solution recently: just do 1.50x upscaling, then 4x upscaling with model (4992*7296), LoL.
@vibhisana Not a bad idea really. Detailers would be faster for me at 1,5x than they are at 2x
@vibhisana I know I'm late to replying this comment... but hey I've seen your awesome whale-piano post and loved it; it's great to hear the backstory now :D
@Chumple wow those two are great examples, 2 very different ways AI gen mess-ups can make the results so magnificent.
(It really does work parallel to gambling, vibhisana's def right. But... somehow, when I find these happy accidents and then look at the rest of the creator's gallery, even the non-messups are good. Hmmm, I wonder what that means... ^_^)
@ravemry9 , I tried heun+beta recently, 5 steps is enough to "peek" the composition, it has much clearer output than euler, and usually more details. You should give it a try.
@vibhisana Will do. Thx.
Great work, very good and detailed, I'll keep trying.
Thank you!!
it gets very impressive standalone results. creator also has this amazing article where they showcase tags/styles with prompts.
really cool checkpoint.
Thanks! I appreciate the kind words
A great style and very responsive. 10/10 A+
Wow! Thank you so much!
well, pretty great !
Thank you!
@Chumple ngl one of the best IL checkpoint, its my main thx very much!
@Reckoo That's awesome to hear! I'm so glad you like it
@Chumple whats the recommend resolution for base generating?
@Chumple like literally i like prompting long using 60 steps with 7cfg, i put 10 loras at the same time, and this model can still handle it pretty well, like i tried to generate same seed same lora chain but its just not as creative as this one does, looking forward in the future. Been comparing it to others such as wai illustrious, one0bession, amix from phinjo, etc... this one handle the details without losing the detail significantly and it does work with a bunch of loras which is few tuned models sometimes din't pop it out as much as this...
@Reckoo I usually generate at 1216x832 or 1024x1024 for the txt2img. Thanks! I usually use a CFG of 3 or 4 and not many loras so it's good to know that it can handle all of that
All of my output is coming out almost solid black using the generator.
Sorry, there's nothing I can do about it. The entire site has been having major issues since they rolled out civit red






