If you've found value in the model, please consider offering your support. Your contribution will be devoted entirely to advancing the SDXL community.
🙋🏼‍♂️ join us (discord) ㅤ|ㅤ 🛒 purchaseㅤ |ㅤ 🌱 donate
AlbedoBase XL (SFW&NSFW)
The refiner is unnecessary, and VAE is included.
GOAL
Stable Diffusion XL has 3.5 billion parameters (excluding the Refiner), which is about 3.6 times more than the SD v1.5 version. I believe that this is not just a number, but a number that can lead to a significant improvement in performance.
It has been a while since we realized that the overall performance of SD v1.5 has improved beyond imagination thanks to the explosive contributions of our community. Therefore, I am working on completing this AlbedoBase XL model in order to optimally reproduce the performance improvement that occurred in v1.5 in this XL version as well.
My goal is to directly test the performance of all Checkpoints and LoRAs that are publicly uploaded to Civitai, and merge only the resources that are judged to be optimal after passing through several filters. This will surpass the performance of image-generating AI of companies such as Midjourney.
As of now, AlbedoBase XL v3.1 Large has merged about 200 selected checkpoints and 251 LoRAs.
LOG
v3.1-Large
• Merged over 50 selected latest versions of SDXL models using the recursive script employed in V3.
The spec grid(370.7 MB): download


v3-mini
I sincerely apologize for keeping you waiting for such a long time.
I’ve been dealing with some personal matters, and while working on the new version, I also faced health issues. Even as I write this, I am still struggling with these challenges.
I felt it wouldn’t be enough to just provide a brief update, so I kindly ask for your understanding as I share this more detailed message.
Since the release of version 2.0, I have been dedicating myself to studying deep learning independently. I don’t have a formal degree, and apart from a modest aptitude for programming, only my background is in the arts. As a result, I lack the mathematical and scientific foundation to achieve significant breakthroughs, given the time and effort I’ve invested. Despite this, the experience of immersing myself in this self-directed study and research has been an invaluable treasure in my life.
Recently, I stumbled upon an idea that could potentially be a significant breakthrough. After reworking hundreds of formulas and methods since version 2.0, I managed to develop a rather intriguing and successful algorithm. The model merging process was based on SDXL1.0 and SD1.5, along with other carefully curated models. These were categorized into five primary classifications: “ANIME,” “REALISM,” “ARTISTIC,” “NSFW,” and “BASE,” and fed into the merging algorithm as datasets. This approach has resulted in some fascinating outcomes.
However, as challenging as the algorithm development was, nothing has been as daunting as the performance testing phase. My physical and mental health deteriorated significantly during this period, to the point where I realized I couldn’t continue this work on my own. This is what ultimately led me to decide to release this version.
And now, I am thrilled to announce the release of the much-anticipated AlbedoBaseXL V3 Mini version. While this model is a smaller-scale merge, it is not limited to any specific area and performs remarkably well across various domains. It has the potential to serve as a new base model for SDXL1.0. (For reference, my merging algorithm is not a “linear merging,” so it can essentially be considered a new fine-tuned model.)

This model, along with the existing AlbedoBase models, is versatile and surpasses all previous versions in every aspect. (The NSFW content, while not extreme, offers a broader range of expression compared to previous versions like v2.1. A dedicated NSFW merging model will be released in the future.)
On another note, I’ve noticed that many shared models have recently begun to adopt licenses that prohibit merging or external commercialization. This has been a disappointment, as it has prevented me from using some truly excellent models for merging.
I would like to express my heartfelt gratitude to the model developers who have provided free licenses, allowing their high-quality models—products of considerable time and effort—to be used for merging.
I’ll be back soon.
I eagerly look forward to your performance tests across a wide range of areas, including ANIME, REALISM, ARTISTIC, 2.5D, 3D, and NSFW.
As model developers, we only plant the seeds. It is ultimately you, the model users and artists, who cultivate them and bring forth the flowers and fruits.
Thank you, as always.
For those who wish to support my work with a small financial contribution, please consider using the links below. I am currently unable to secure employment and face an uncertain future regarding my livelihood.
The spec grid(380.5 MB): download


v2.1
Re-merge and adjust v0.1 to 2.0 using new merging algorithm and formula.

The spec grid(424.5 MB): download

v2.0
I'd like to thank everyone who helped me on the AlbedoBase XL Pre side. Without you guys, the release date would probably have been much later. Thank you so much!

I have written a custom script to converge the existing AlbedoBase XL models into one. Intricately aligning the row and column weights of all U-NET and CLIP blocks according to a unique formula of mine.
If you encounter a bug in image generation (if nothing is generated), please switch to CLIP SKIP 2 or modify the prompt slightly! There may be combinations of prompts that CLIP does not recognize. In that case, you can change the order of words, use different words, or, most simply, change the CLIP SKIP. I will gradually work on resolving these issues in the future like v1.3.
The spec grid(403.5 MB): download
v1.3
In order to illustrate the quality associated with the model's randomness, I standardized the seed value at '9' for all showcase images intended for sampling and proceeded with their immediate generation.
Especially with this version, due to the significant impact of negative prompts, leaving the negative prompt field empty is likely to produce the nice quality.
The spec grid(438.7 MB): download

As you can see, as the number of Steps increases, it becomes available for all samplers, and the quality also improves.
Due to the effect of the LoRA I developed and merged, as described below, using sentence-form prompts rather than tag (a list of words) prompts is directly related to improving quality.
I merged 45 checkpoints and 7 LoRAs. After that, I merged AlbedoBase v0.4 and v0.3 in order, less than 0~5%, to reawaken the diluted merged models that had become outdated.Â
Among the 7 LoRAs, one is created by me. It involves analyzing and annotating captions for a total of 174 high-quality pictorial photos using GPT4-V. Merging this LoRA resulted in astonishingly clear images and an impressively excellent understanding of prompts.

My self-created LoRAs are exclusively available for purchase to my Ko-fi supporters at the Creative level or higher.
v1.2
Merged the 22 latest checkpoints.
The spec grid(565.6 MB): download
v1.1
Stabilized.
More detailed.
If you think you are an advanced user, I recommend version 1.0. If version 1.0 finds the right settings, it can output much more vivid works.
The spec grid(349.7 MB): download
v1.0
Merged 106 LoRAs.
Merged 19 Checkpoints.
The model can produce different results depending on the settings you choose, so it's important to check the spec grid before using it.
I have found that using a few specific negative prompts can help to resolve the issue of asymmetrical eyes or pixelated images. The Spec Grid may vary depending on your CPU or GPU device, so please use it as a general reference. Experiment with a few negative prompts to improve the quality(ex; strabismus). I found that it is difficult to satisfy all settings equally as the number of merged LoRA increases. However, I would like you to focus on this advantage in version 1.0, as it can produce works of amazing quality in various aspects with the right settings. I will be back with a more stable version in the future.
You can find useful setting values in the showcase or by searching for others.
As always, it's best to leave the negative prompt blank for the best results.
This v1.0 was a lot of work, so I'm taking a break for a bit. I hope you'll enjoy using the model, and if you merge it, please share it in Civitai for free. That way, we can all keep improving it.
The spec grid(479.4 MB): download
v0.4
Merged 132 LoRAs.
Merged 4 Checkpoints.
The spec grid: download
v0.3
Improved in all of the samplers.
Achieved lifelike realism.
Stabilized.
The spec grid: download
v0.2
Significant improvements in clarity and detailing.
Improved hand and foot implementation.
Major aesthetic improvements; composition, abstraction, flow, light and color, etc.
v0.1
After appropriate fine-tuning on the SDXL1.0 model, meticulously and purposefully merge over 40+ high-quality models publicly available on Civitai.
Testing has primarily focused on ensuring maximum quality with the minimum number of prompt tokens, and it has not been confirmed how much the quality can improve when using a large number of tokens. (Please conduct your own tests and share the results)
Typically, the most beautiful results are achieved at the midpoint between reality and animation.
Nevertheless, when using an appropriate prompt, there is generally nothing it cannot express. (I assert that it possesses abundant value as a foundational model that surpasses others in merging. However, please keep in mind that this is currently v0.1)
Description
v1.3
In order to illustrate the quality associated with the model's randomness, I standardized the seed value at '9' for all showcase images intended for sampling and proceeded with their immediate generation.
Especially with this version, due to the significant impact of negative prompts, leaving the negative prompt field empty is likely to produce the best quality.
The spec grid(438.7 MB): download
As you can see, as the number of Steps increases, it becomes available for all samplers, and the quality also improves.
Due to the effect of the LoRA I developed and merged, as described below, using sentence-form prompts rather than tag (a list of words) prompts is directly related to improving quality.
I merged 45 checkpoints and 7 LoRAs. After that, I merged AlbedoBase v0.4 and v0.3 in order, less than 0~5%, to reawaken the diluted merged models that had become outdated.Â
Among the 7 LoRAs, one is created by me. It involves analyzing and annotating captions for a total of 174 high-quality pictorial photos using GPT4-V. Merging this LoRA resulted in astonishingly clear images and an impressively excellent understanding of prompts.
My self-created LoRAs are exclusively available for purchase to my Ko-fi supporters at the Creative level or higher. I plan to release more and more updates in the future. The prices range from $10 to $50.
FAQ
Comments (31)
This model is unbelievable! I made a youtube video about your model! Thank you so much, 1.3 is downright gorgeous!
The BEST SDXL Model for realism right now! Just updated today! - YouTube
I've watched it all the way through. Thank you very much! I've given it a like. Rooting for you!!
@albedobond Thank you so much! If it is ok with you, I don't have a ton more SD videos planned, but I would like to continue using your model, linking folks over here to your civitai and promoting your work a bit.
@hodak2522Â I'm just grateful!
hey, nsfw images containing male characters require additional loras and are extremely difficult to generate. any plans to add such photos to the training?
Ah! I didn't think of that part. - I am foolish. After checking, I'll try to merge a self-made LoRA of Male's in the next version. Thank you for letting me know!
hey! any news regarding that?
@clchbn First, I am testing a new combination method, and if you have time, please test it here and give me your feedback so that I can progress to the next stage more quickly ;) I do want to release it somehow before Christmas.
@albedobond sure thing! I actually started testing it. It is beautiful, but I see no difference in terms of nsfw male photos. any specific aspects you feel the new model should excel in, so I should test them?
@clchbn Thank you! It would be really helpful if you could upload the images there and report any shortcomings or bugs in the post comments 🙏
If you have a specific set of images that you would like to merge into my model, please send them to my email. After reviewing them, I will create a LoRA and proceed with the merging.
Email: [email protected]
imma send u a link to google drive if thats cool
@charnelofmaw Yeah! That would be really cool
I have been playing around with creating models -- I have an AMD 7900xtx with 24gb vram. Still when I try to train a model using your SDXL model as base I run out of vram sadly. Still playing with settings trying to get it right.
Your work is still amazing and loving it. I will keep trying to get it to work so I can make a youtube video about it. I can train into normal SD 1.5 models without issue. While I think your SDXL work is amazing and definitely am all for you keeping up the great work, it would be really awesome if at some point you had a similar type model in SD 1.5 available (obviously it would not be as good as the SDXL model).
what resolution you need?
@KINGLIFERÂ Any size of images is fine as long as there are at least 15 images of 1024x1024 size!
Tip for Beginners:
1. Use sentence-form prompt.
2. Do not use strong and many negative prompts.
3. Do not use CLIP skip 2 ~
4. Adjust the sampler, CFG, and steps properly.
5. Use the specific, long sentence positive prompts.
6. Check the showcase and gallery’s image settings.
7. Enjoy your exploring time.
3. Do not use CLIP skip 2?how to close clip skip 2?
@wpw0213727Â Then, You should use other services or local Webui(Comfyui, Foocus, etc) Currently this Civitai generating service doesn't support changing the CLIP SKIP :(... But It's okay now, AlbedoBase XL v2.0 officially supports CLIP SKIP 2!
How come I am getting this message when I try to use this model? :(
AssertionError: We do not support vanilla attention in 1.13.1+cu117 anymore, as it is too expensive. Please install xformers via e.g. 'pip install xformers==0.0.16'
If you are running SD locally or on colab you will need to add --xformers in the webui-user.bat file arguments
How can I use it locally on my PC?
install stable-diffusion-webui from automatic1111 github. place AlbedoBase XL in it's models/stable-diffusion folder, then it will be selectable as a checkpoint in the webui
Everyone, please help with testing the next version here! (AlbedoBase XL Pre)
hey man ...
watch my image and prompt ...
seams your oversized prompts are not that highlclass you thinking ?!?
sry i play a-lot with all models and only three creators here on Civitai know what they doo ...
iam to smart for all here .. sry again
nerver seen 300 tokens :D
This model is absolutely insane. I'm throwing 5 LORAs at it all at 0.9+ with a few of them being sliders set really high and my images still aren't breaking. It's pulling off what other models can't even come close to without going insane and generating nightmares.
This is an actual gamechanger IMO and I spend well over 30 hours in SD a week.
Do you mind sharing what sliders you are using? I am moving from sd1.5 and the hair sliders I used got me an exact image I want to transfer over to SDXL.
Thank you.
@KINGLIFERÂ Sure!
In a single gen I was using:
neg4all_bdsqlsz_xl_v7 (negative LORA)
extremely detailed (slider)
confused (slider)
custom character LORA
ahegao_xl LORA
I know there was a gen I did using one more slider but can't remember which to track it down. Also, I experimented for probably another hour or two, and I was able to get some bad gens. So it might just be that my customer character LORA is trained enough that it kept things congruent. In either case I typically have bad artifacts with just two LORAs set high, let alone 2 set high and 2-3 sliders also set to strength 3.
First of all, thank you for letting me use your model for free. It is a very excellent model and it seems to respond well even if I write a long sentence. But there is one problem. When I make a human image, if I make 10 images, all 10 images have their eyes closed. I have to make 30 images to find an image with open eyes. So it takes too much time and when I write (open eye) 30 times in the prompt, then the image with open eyes is created, but the image is not just open eyes, but “very wide open eyes”. I wrote this to let you know if you are still developing this excellent model.
Can I know the prompt? I, as well as other users, didn't encounter such an issue. If there is a problem, it might be due to sensitivity towards specific words in the prompt...?
@albedobond Thank you for your reply. I tested it again and found that if the sentence is too short or the scene description is lacking, the image comes out with closed eyes. For example, when I write simple and vague prompts like “a city and a person in cyberpunk style” or “a dancing dancer”, it seems to happen. So I think I need to put the concept or theme of the image I imagine in detail with various elements when I enter the prompt.
Details
Available On (1 platform)
Same model published on other platforms. May have additional downloads or version variants.














