Flux and Qwen have no clue what a vagina looks like, and I was disappointed with the other LoRAs that were vagina-related. The main problem I found was that they didn't train on the pose, so while it could draw a pussy if the situation arose, she'd never put herself in that position. While attempting to train the main LoRA, I have been repeatedly foiled by Flux/Qwen's attempts to hide my favourite pixels.
Note, for compatibility, this is a unet-only LoRA. It won't affect the meaning of words, and should be safe to combine with other LoRAs. It also happens to make the number of fingers on hands more accurate for reasons I don't know. As a result, these tags aren't as fine-tuned as the main LoRA. The goal of this LoRA is to seamlessly integrate with your workflow for any style you like.
For those wondering about the main LoRA, it's this one:
https://civarchive.com/models/903490/grabby-realistic-tentacles
This model was entirely trained on images that I personally hand-generated. You can use it without crediting me, including for commercial reasons. Just let me know if you do, I'm curious.
Tags:
[x] photorealistic: It is just like real life!
[x] semirealistic: I can tell this isn't real, if I look hard.
[x] cartoonish: This isn't real, obviously
[x] vagina: recommended over 'pussy' to avoid felines entering the chat
[x] pussy: I didn't want to disappoint people who prefer this word
[x] topless: #FreeTheNipple
[x] bottomless: She isn't wearing bottoms
[x] naked: She isn't wearing anything!
[x] open crotch: There's a convenient hole in this garment
[x] full body shot: You can see her whole body!
[x] frontal shot: Facing directly towards you
[x] crotch shot: The whole point of this photo is to capture the coochie
[x] male pov shot: WIP, but ideally your eyeballs took this photo
[x] three-quarter angle shot: She is facing around 45deg away from the camera
[x] three-quarter shot: AKA cowboy shot, but avoids the invasive cowboy hat
[x] medium shot: You can see like half of her
[x] torso shot: Crop out the face and knees
[x] dry: Not wet
[x] wet: Not dry
[x] cum: Cinnzeo's secret sauce
[x] masturbating: WIP, hands are closer than usual to the clitoris
[x] penis in vagina: Ideally.
[x] sex machine: Automation improves efficiency
[x] slime clothes: Best if used with the main LoRA
[x] <COLOR> tentacles: Best if used with the main LoRA
[x] tentacle in vagina: Best if used with the main LoRA
[x] tentacle in ass: Best if used with the main LoRA
[x] suspended by tentacles: Best if used with the main LoRA
[x] legs up: Her legs are spread open and up
[x] legs wide open: Her legs are spread open normally
[x] on her back: She is resting on her back, not doing crunches.
[x] laying on a <THING>: Resting comfortably.
[x] tiny breasts: Flat chested
[x] small breasts: Slightly smaller than average, but nice and perky
[x] medium breasts: My wife has these, and so do most women
[x] large breasts: My wife wants these, and so do most women
[x] huge breasts: This is utterly impractical, these are ridiculous, unrealistic knockers, you need medical help to walk around with these gazongas.
Description
FAQ
Comments (31)
2.2gb? A vagina is so huge?
There's a load of other pron related concepts in this lora. It huge, but it's awesome.
@ragnaroook is right, it's actually just the non-tentacle subset of images from my main LoRA. This isn't just a LoRA for vaginas, it's a LoRA for making Qwen NSFW. Plus, if you've got Qwen running, you have to have a good sized disk. It takes so bloody long to train these, I'm not going to sacrifice accuracy in the name of disk space. Training the LoRA took up 315GB of disk space on my training rig.
@turiyag however a high rank lora also consumes more VRAM during inference
@civitaiclimatic796289 Well, oddly, it shouldn't. So, a LoRA is basically a change made to a model, it doesn't make the model itself bigger. It essentially tells the base model "change this weight from 1.03 to 2.34, and this weight from 1.33 to 0.65".
BEFORE inference, the Low Rank matrix is expanded to the dimensions of the target parameter shape, and it permutes the parameters. So a larger rank LoRA will take more time to load the model, since there's more math to do. But once the model is loaded, it shouldn't take up more or less RAM.
But what I THINK is happening, is that this LoRA is a BF16 format. So when it's applied to smaller quants of the base model, like INT4, then it casts them up to FP8 or BF16, which uses twice or 4x as many bits. I don't see more VRAM usage on my machine when I use the base FP8 model (https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/blob/main/split_files/diffusion_models/qwen_image_fp8_e4m3fn.safetensors). I haven't personally tried nunchaku, but when people complain about VRAM issues, they seem to mention that quant a lot.
I just use it for the more accurate fingers 😋
Qwen Lora is already pretty awesome, and can do a lot of things labeled "Best if used with the main LoRA", but still I'm tempted to ask - is there a version of that lora for qwen as well?
Give me like a week, the training dataset on the main LoRA has around 10 000 images now, so to get through a single epoch it takes about 8h. I don't need as many epochs as other LoRAs, because of my huge dataset, but it needs a few to settle down and stop doing freaky things.
@turiyag Thank you very much. (And I'm kind jealous, my computer needs like 8 hours to get a lora out of 200 pictures 😋)
@ragnaroook I don't really mind my computer grinding away for 8h, I sleep most nights, so it can grind away while I'm resting. Qwen is crazy, it's actually too large to fit entirely on my GPU for training, so my CPU has to help with some of the layers.
All of the training data are images that I have personally generated. Which means that the main thing that consumes my compute time are actually generating images. It's much much easier now that I have LoRAs for each architecture I've loved, but initially, back in the days of SD1, there was no prompt for a yellow tentacle. So I had to hand-build the training data in Photoshop, but I wanted the entire dataset to be from my own art, not anyone else's. Obviously it's a fusion of my art and the base training data from the models I use to help, but when I'm making new training data (last week I was trying to teach it the concept of octopus-like tentacles with suckers) I have to keep Photoshop open, because new concepts that no model knows are not easy to generate with AI. I've got trained YOLO models for detecting faces, boobs, tentacles, feet, hands, and for finding when an arm stops being an arm and suddenly and inexplicably becomes a tentacle. I train a lot of LoRAs that never leave my computer, because they have like 10 training images, and they've been trained for like 3 epochs, so they can barely do the new concept.
BUT, on the flipside, generating new training data and sorting and labelling it involves looking at a lot of tentacle porn, so it's not soooooo awful, haha.
@turiyag Thanks for the insights. I'll give my rig another run, trying to give qwen Loras a go
This Lora is giving me non-realistic images (mostly semi-realistic, sort of rendered images.) Even when using "semirealistic" and "cartoonish" in the negative prompt and photorealistic in the positive prompt. Using res_multistep and simple as sampler and scheduler.
Sorry, I haven't figured out what's causing that. The source images I feed into the training look a lot more realistic. If you are using ComfyUI, you can take the output image, and pass it through PonyRealism (that's been my favorite semi-realistic to realistic converter) with just a denoise of like 0.1 or 0.2, and it will make it realistic. Hopefully in v2 I'll have that figured out.
This LoRA is so big, it can't be loaded with nunchaku, even on a 4090. I'm getting OOM most of the time.
I can run it on Nunchaku with a 4060/16GB Vram, I had to lower "num_blocks_on_gpu" to 30 in the Nunchaku Qwen-Image DiT Loader. Wasn't sure whether it was because of the Lora or Chrome eating all Ram, but it works.
I can load it on nunchaku fully when using another card for windows but no other loras fit
@Omegaxx That's odd. A LoRA shouldn't take up VRAM. It permutes the source model, it doesn't add to it. I don't have this issue in ComfyUI. I'm seeing a recurring theme of people disliking the LoRA size, I'll look into reducing it.
@turiyag No, it it rather like separate model which not only takes space but requires its own compute you can merge lora into model when it no longer takes space but if loaded in usual way it sits separately and eats VRAM
@Omegaxx I just double-checked with ComfyUI locally, and the LoRA does increase my RAM usage, but not my VRAM at all. Generating an image without any LoRA, and with the LoRA takes just as long. The whole concept of a LoRA is to be able to fine-tune without fully fine-tuning. Maybe nunchaku is doing it strangely? I got a LoRA to consume VRAM, but only when I generated two images in a single workflow, with Illustrious, presumably because ComfyUI realizes that it has the space in VRAM to two copies of Illustrious, so rather than re-applying the LoRA every time the workflow executes, keeps two copies around. I'm just using all standard built-in nodes.
OH! I bet I know what's happening. This LoRA is intended to permute OG Qwen, and Nunchaku is a quantized model. I'm betting that since this LoRA is BF16, it's taking your FP4/INT4 or otherwise quantized model, and it's blowing it up to BF16, so like 4x-ing the RAM it needs. Does it still eat VRAM when you use a non-nunchaku model?
@turiyag Qwen behaves very weird in general as it does not report RAM usage properly, like 4 bit quantized model takes only 8 GB but my vram consumption remains similar as with 8 bit model.
when it OOMs it doe not even show amy considerable RAM use changes juts fails to allocate something and crashes.
loras do increase ram usage but it is not easy to see, it works fine until you load too many loras and everything crashes
as I see you can load about 2GB worth of loras at most.
also depends on image size a bit
but generally there is "plenty of memory" all the time
so to test your therosy you could try juts load several big size loras for test and see if it will OOM as if they just modify model them no matter how many you load it would have no effect on memory use, but in my experience loras consume extra memory and run in parallel to model
Finally, I can say goodbye to the dark and bumpy crotch!
The OG Qwen has very traumatizing hallucinations of the nethers.
Well that colosal size Lora but i don't get if it can do anything other than juts plain pussy?
like no gape streching big insertions piercings when Using it with qwen image edit it defaults whatever I give to it to the plain thing in default state, always the same no matter context or character
Sorry, I didn't have any crazy stuff in the training set. I want to keep it realistic, nothing cartoonish, but if you have some sample images as example that represent what you'd like it to do, let me know and I'll make a handful for the next version.
Do you use 64 linear rank to get the lora to be 2.25gb? What type of gpu did you use to train this on? Did you use AI toolkit or Musubi Tuner or other?
It took FOREVER. I eventually got it to work with Musubi tuner. The actual training only took like 16h, on my RTX 5090 with 32GB of VRAM, but GETTING IT TO WORK was the literal worst. I had to drop from Windows to WSL, I had to buy a Patreon sub to SECourses to get the files for his Qwen image training tutorial, and his settings absolutely obliterated my model's coherence, because his learning rate was far too strong. He was training on just 30 images, so he didn't have the kind of dataset I wanted to throw at it. I have around 1000 coochie images I used to train this, but I also needed to feed it non-sexual images and text images to keep it stable. Qwen is super picky. If you just give it 10 images, you'll make a completely inflexible LoRA that absolutely will recreate that dude's face, but if you want it to be smart, to be good, to understand what you MEAN, you need to follow the same training process the OG devs used, which involves having Qwen-VL describe your images, and Qwen-VL has no idea what sex is. You need long form QWEN SOURCED descriptions, you need to train at 1.8 megapixels, it's so finicky. Which is super sad, because the base model is so so capable. It also needs my PC to have NOTHING running, or the training goes from 3s/it to like 87s/it.
The key is that you need Qwen to make the image descriptions, and then you edit those long form descriptions. It can't handle tags like other models, it just goes crazy.
@turiyag Wrote an LLM tool that creates prompts from images quite well actually. Atm training on 750 images on a 3090 rtx. Using 3bit ARA. Just made it fit in 24gb of memory. I can't increase from 16 to 32 in Linear Rank though. And I am guessing that is what you did?
I am also training on pony images which are 1 megapixel. the first lora I made came out excellent, but only used 200ish images.
@Lady_Valeria Qwen is super picky about training labels, I've been struggling for so long to get it working, I eventually read the entire Qwen-Image paper, and they trained it with descriptions from Qwen-VL. The way I finally got it to work is with Qwen/Qwen3-VL-4B-Instruct, I give it each image in my dataset, and ask it to describe it, here is my current prompt below. But after Qwen has labelled them all, it gets everything the LoRA is for wrong, haha, because Qwen-VL hasn't been trained on what it means for a tentacle to be in a vagina, so I then go through every caption manually and correct each caption, and cut out the headers before training, it's pretty funny to see the output of Qwen-VL, since it uses analogies and euphemisms to describe things, presumably because the training data never got to see the word "nude" or "naked" or "vagina". It's like "a glossy tendril is near her lower torso and is excreting a thick white syrup". But Qwen-VL handles describing most things very nicely, like the background and her face/skin and most clothing:
```
Describe this image in these sections. Be succinct, each section should be a maximum of 30 words long. Avoid negative descriptions like "no shirt" or "without fear", and use positive descriptions like "topless" or "brave".
TONE: The aesthetic tone and lighting (ex. moonlit ethereal misty forest ambiance)
SETTING: The location the image is set in (ex. A brightly lit opulent hallway with clean marble floors)
SHOT: Shot scale (ex. three-quarter shot, bust shot, dutch angle wide shot, or close up shot)
POSE: The pose of the central subject, include arm and leg positions (ex. standing at attention, legs together, one arm in a military salute, the other holding a rifle)
CLOTHING: The clothing, or lack thereof, of the central subject (ex. wearing a white crop top and blue pleated skirt with long black leggings and combat boots), if the subject is partially nude, describe that (ex. topless with green shorts).
EXPRESSION: The facial expression of the central subject (ex. gently smiling)
NATIONALITY: A one-word nationality for the central subject (ex. Canadian), always pick a nation, not a general class.
DETAILS: Without repeating earlier descriptions, include any other important information here. This section should be at least 40 words long. (ex. She is surrounded by thick blue tentacles that are wrapping around her legs and torso. One tentacle is in her vagina. The cave walls are wet and slimy. A small stream of muddy water flows from the background, pooling at her feet.)
SUBJECT: Without repeating earlier descriptions, include a full description of the central subject. (ex. A cute Neko catgirl with 4ft long red hair and pale skin with vibrant red lipstick. She is filthy from playing in the mud.).
```
I'm hand-labelling training data tonight, and I generate all the training data myself, and label it myself, and I learned early along, that if an image is "almost good enough" then it is NOT GOOD ENOUGH DON'T PUT IT IN YOUR TRAINING DATA. And I was just labelling her pose, since I'm trying to teach the model the word "clitoris", and my label is:
"Legs parted, she masturbates with one hand, fingers adorned with a silver diamond ring. Her middle finger is rubbing her clitoris. Her pointer finger and ring finger rest on her labia. Her pinky curls away with pleasure. Her thumb is above on her thigh and FUCK THE THUMB IS ON THE WRONG SIDE"
This image sadly needs to be discarded.
What is V? OK.
And now: what is A?
It should have basic A capabilities. It's just the non-tentacle data from the main LoRA, it's not actually specifically a vagina dataset. I'm not just personally an A man myself, and I hand-label all my training data, so I just don't feel like hand-labelling an A-specific dataset. But there are plenty of training images with A.
Details
Available On (1 platform)
Same model published on other platforms. May have additional downloads or version variants.



