CivArchive
    Preview 74735

    Check my exclusive models on Mage: ParagonXL / NovaXL / NovaXL Lightning / NovaXL V2 / NovaXL Pony / NovaXL Pony Lightning / RealDreamXL / RealDreamXL Lightning

    Recommendations for using the Hyper model:
    Sampler = DPM SDE++ Karras or another / 4-6+ steps
    CFG Scale = 1.5-2.0 (
    the lower the value, the more mutations, but the less contrast)

    I also recommend using ADetailer for generation (some examples were generated with ADetailer, this will be noted in the image comments).

    This model is available on Mage.Space (main sponsor).
    You can also support me directly on Boosty.

    Realistic Vision V6.0 (B2 - Full Re-train) Status (Updated: Apr. 4, 2024):
    - Training Images: +3400 (B1: 3000)
    - Training Steps: +724k (B1: 664k)
    - Approximate percentage of completion: ~30%

    All models, including Realistic Vision (VAE / noVAE) are also on Hugging Face

    Please read this! How to remove strong contrast.

    To make the image less contrasty you can use LoRA [Detail Tweaker LoRA] in a negative value.

    Orange Color = Optional

    I use this template to get good generation results:

    Prompt:

    RAW photo, subject, 8k uhd, dslr, soft lighting, high quality, film grain, Fujifilm XT3

    Negative Prompt:

    • (deformed iris, deformed pupils, semi-realistic, cgi, 3d, render, sketch, cartoon, drawing, anime), text, cropped, out of frame, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, UnrealisticDream

    • (deformed iris, deformed pupils, semi-realistic, cgi, 3d, render, sketch, cartoon, drawing, anime, mutated hands and fingers:1.4), (deformed, distorted, disfigured:1.3), poorly drawn, bad anatomy, wrong anatomy, extra limb, missing limb, floating limbs, disconnected limbs, mutation, mutated, ugly, disgusting, amputation, UnrealisticDream

    Euler A or DPM++ SDE Karras

    CFG Scale 3,5 - 7

    Hires. fix with 4x-UltraSharp upscaler

    Denoising strength 0.25-0.45

    Upscale by 1.1-2.0

    Clip Skip 1-2

    ENSD 31337

    Thanks to the creators of these models for their work. Without them it would not have been possible to create this model.

    HassanBlend 1.5.1.2 by sdhassan

    Uber Realistic Porn Merge (URPM) by saftle

    Protogen x3.4 (Photorealism) + Protogen x5.3 (Photorealism) by darkstorm2150

    Art & Eros (aEros) + RealEldenApocalypse by aine_captain

    Dreamlike Photoreal 2.0 by sviasem

    HASDX by bestjammer

    Analog Diffusion by wavymulder

    WoopWoop-Photo by zoidbb

    Life Like Diffusion by lutherjonna409

    AbsoluteReality by Lykon

    CyberRealistic by Cyberdelia

    Analog Madness by CornmeisterNL

    A-Zovya Photoreal by Zovya

    ICBINP - "I Can't Believe It's Not Photography" by residentchiefnz

    epiCRealism by epinikion

    Juggernaut by KandooAI

    Description

    NOT FOR TRAINING!

    Model for inpainting and outpainting.
    Make sure that the model name has the ending -inpainting.

    • vae-ft-mse-840000-ema-pruned.ckpt included

    FAQ

    Comments (241)

    adryFeb 5, 2023· 1 reaction
    CivitAI

    Where is the VAE? Is it integrated in the model and gets selected automatically? There is no option to select it in automatic1111 webinterface settings...

    SG_161222
    Author
    Feb 5, 2023· 2 reactions

    Hi. In version 1.3, VAE is integrated into the model and works automatically.

    LayscreamFeb 5, 2023· 2 reactions
    CivitAI

    when I run 1.3-inpainting...

    RuntimeError: Error(s) in loading state_dict for LatentDiffusion:

    size mismatch for model.diffusion_model.input_blocks.0.0.weight: copying a param with shape torch.Size([320, 9, 3, 3]) from checkpoint, the shape in current model is torch.Size([320, 4, 3, 3]).

    SG_161222
    Author
    Feb 5, 2023· 1 reaction

    I will try to solve this problem.

    SG_161222
    Author
    Feb 5, 2023· 2 reactions

    Try downloading the config file and put it in the folder with the model, make sure that the model has the name ending -inpainting and that the configuration file has the same name as the model.

    highly_illiquidFeb 5, 2023

    @SG_161222 I can't get the models to load in A1111. I've downloaded the model and the configuration file, and I put them in the correct file location.

    I named the files the following:

    realisticVisionV13_v13Inpainting-inpainting.safetensors

    realisticVisionV13_v13Inpainting-inpainting.yaml

    What am I doing wrong? If the files are named in correctly, can you type it out correcly for me here? Thanks!

    SG_161222
    Author
    Feb 5, 2023· 1 reaction

    @donald_diffusion Hi. The files are named correctly. Try running A1111 with webui.bat

    Once you have tried it, let me know.

    SG_161222
    Author
    Feb 5, 2023· 1 reaction

    @donald_diffusion You can also shorten the name a bit to "Realistic_Vision_V1.3-inpainting.safetensors" and "Realistic_Vision_V1.3-inpainting.yaml"

    highly_illiquidFeb 6, 2023· 1 reaction

    @SG_161222 got it working! Thanks

    LayscreamFeb 6, 2023

    @SG_161222 i just changed the name to realisticVisionV13_v13-Inpainting.ckpt . But ... File "D:\AI\stable-diffusion-webui\repositories\stable-diffusion\ldm\models\diffusion\ddpm.py", line 448, in init

    super().__init__(conditioning_key=conditioning_key, args, *kwargs)

    TypeError: DDPM.__init__() got an unexpected keyword argument 'finetune_keys'

    SG_161222
    Author
    Feb 6, 2023

    @Layscream Did you update Automatic1111?

    SG_161222
    Author
    Feb 6, 2023

    @Layscream You can also try to perform these actions:
    Try running A1111 with webui.bat
    You can also shorten the name a bit to "Realistic_Vision_V1.3-inpainting.safetensors" and "Realistic_Vision_V1.3-inpainting.yaml"

    LayscreamFeb 6, 2023

    @SG_161222  thank u! i will update and try again~

    jimmyjazzjazzFeb 10, 2023

    @SG_161222 Where do you get the config file??

    jimmyjazzjazzFeb 10, 2023

    I have the same exact error and can't seem to get it to work. Using Easy Diffusion.

    SG_161222
    Author
    Feb 11, 2023
    SirhFeb 5, 2023· 1 reaction
    CivitAI

    I'm having serius problem with this model, it always generate ugly malformed faces, even if i put all my input efforts in prompting nice faces

    bbackagainnFeb 6, 2023

    ive found having strange dimensions to your photos influences this heavily. stick to standard widths and heights

    SG_161222
    Author
    Feb 6, 2023

    Hi. What GUI are you using? What size image are you trying to generate?

    SirhFeb 19, 2023

    @SG_161222 AUTOMATIC1111 512x512

    SG_161222
    Author
    Feb 19, 2023

    @Sirh You can also try version 1.4

    The link is in the first post: https://boosty.to/evgk.132

    sniparsFeb 5, 2023· 2 reactions
    CivitAI

    Can you help me please, did a clean install and added the model, I keep getting this error when trying to generate an image, also made sure everything was named correctly. Much appreciated!

    File "C:\stable-diffusion-ui\stable-diffusion\env\Lib\site-packages\torch\nn\modules\module.py", line 1497, in load_state_dict

    raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(

    RuntimeError: Error(s) in loading state_dict for UNet:

    size mismatch for model1.diffusion_model.input_blocks.0.0.weight: copying a param with shape torch.Size([320, 9, 3, 3]) from checkpoint, the shape in current model is torch.Size([320, 4, 3, 3]).

    SG_161222
    Author
    Feb 6, 2023

    Hi. This is a very strange situation. Some people have my model working, and some people don't have it working. Then we will try to solve this problem. What version of Python do you have installed? Have you updated the Automatic1111 GUI?

    maribaraFeb 5, 2023· 2 reactions
    CivitAI

    i am pretty sure ive done everything correctly (downloaded the model, and put the model in the right folder) however whenever i try to generate images, it is "in queue" and never generates. is it because its not ckpt file? or could there be another issue? I'm sorry, im very new to this

    SG_161222
    Author
    Feb 6, 2023

    Hi, if you need a model to generate images then use version 1.3, if you need to use inpaint or outpaint then download 1.3-inpainting. And I want to ask. What GUI are you using?

    loboFeb 6, 2023· 1 reaction

    What are you using Automatic1111, NKMS? post more info

    chrissy22Feb 6, 2023· 1 reaction
    CivitAI

    hi i created a model with the previous version just fine, however with this version im getting an error. this version was working fine with the right yaml, but when creating a model i get this error : Exception training model: 'Given groups=1, weight of size [320, 9, 3, 3], expected input[2, 4, 56, 64] to have 9 channels, but got 4 channels instead'.

    any ideas ?

    my model spent all days training 4000 images, and when it came time to do the last quick swoop of everything before generating the ckpt i got this error. is there anyway i can fix that issue, or at least use the images generated with another model ? so i didnt waste all this time for nothing

    thanks :D

    SG_161222
    Author
    Feb 6, 2023

    Hi. Sorry for the inconvenience. What model did you use for training?

    chrissy22Feb 6, 2023

    i used your model -inpainting and trained it with my custom pictures (which ive done yesterday) with the non inpainting model and it worked fine. i did the exact same thing and got this error

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 You can make an inpaint version of the model from your model, I can send you a post with information on how to do it.

    chrissy22Feb 6, 2023

    @SG_161222 i dont have a model, i trained my images in dreambooth, i would love to be able to somehow pick back up from the generated images i was able to make and not restart the whole generation progress. also when crashing it generated a ckpt file which appears to be bigger than realisticvision13-inpainting, which should mean my data is there, but i cant do anything with it

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 Can you upload the ckpt file to Google Drive?

    chrissy22Feb 6, 2023

    @SG_161222 uploading

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 Started downloading. So you can't open the model?

    chrissy22Feb 6, 2023

    @SG_161222 no it doesnt open since it only went thru the first part and longest part of generation. its only missing the last part before it generates a ckpt. which is why its confusing that i even have one.

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 I downloaded the model and just opened it in Automatic1111. I did not see any errors. Have you updated Automatic1111? :)

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 I can also say that inpainting models do not lend themselves to training.

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 So you have to train the regular version of the model and then combine it with the inpaint version to get the inpaint model with your new data.

    chrissy22Feb 6, 2023

    @SG_161222 does it mean i spent the whole day training my images for nothing ? :/ im updating now and trying

    SG_161222
    Author
    Feb 6, 2023· 1 reaction

    @chrissy22 Yes, it was a waste of time :(

    Sorry for the delayed answers. I use a translator and it doesn't always translate the way I want it to.

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 It's my fault for not marking that the inpainting model is untrainable.

    chrissy22Feb 6, 2023

    @SG_161222 but how did you manage to open it ? i can not open it and have the latest version :/

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 Try deleting the .yaml file and running Automatic1111 via webui.bat

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 But it is still useless, because the inpainting model has a different structure and most likely all your data is not in it.

    chrissy22Feb 6, 2023

    @SG_161222 alright thank you, also one more question, i trained a model on the non -inpainting version and it worked great, however i can only get closeup pictures, i have added 280 pictures to train, all different angles, some full body, some face only, why is it that is only gives close up results (for example boobs to face, sometimes waist to face ) i can not get a proper full body result, what kind of pictues should i add to the training ? if i crop the picture to 512-512 i can not get the full body in the pictures that i want to train, this is complicated. thanks

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 Perhaps for this you should make a text description in .txt format for photos. txt file must have the same name as the photo.
    It's a pretty broad topic.

    chrissy22Feb 6, 2023

    @SG_161222 i have no idea how to do it easily, i type my prompt in dreamboot and forget about it, im new to this lol now i have to spend all night retraining :( thanks for your help tho

    SG_161222
    Author
    Feb 6, 2023

    @chrissy22 Try with a few images :)

    Also don't forget to specify in "Class Prompt" the value [filewords] so that the model is trained based on the text description.

    moogooFeb 7, 2023· 2 reactions
    CivitAI

    Hi dev just letting you know i am getting error message when trying to load your model in to invokeai

    the error message is-

    raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(

    RuntimeError: Error(s) in loading state_dict for LatentDiffusion:

    size mismatch for model.diffusion_model.input_blocks.0.0.weight: copying a param with shape torch.Size([320, 9, 3, 3]) from checkpoint, the shape in current model is torch.Size([320, 4, 3, 3]).

    it won't let me even change to your model in the model manager, I am a noob so don't really understand any of this.

    thanks

    stablydiffusedFeb 7, 2023· 1 reaction

    What version? I was able to load the 1.3 .ckpt version in InvokeAI version 2.5.0 without any issues.

    Triceratops9Feb 7, 2023· 1 reaction
    CivitAI

    Super noob question. I know how to train and create a .cpkt file and then launch Automatic 1111 with that. How do I use this file? And how is it different than a .cpkt file? (Or what am I searching for in google to find out?) Thank you!

    KurtFlandersFeb 7, 2023· 1 reaction

    Use it the exact same way as .ckpt files. SafeTensor is just safer for the end user,

    shrrgsaFeb 7, 2023· 1 reaction
    CivitAI

    hello sir what how can i use the vae from huggingface i didnt use vae included bc it woudnt let me train using it

    SG_161222
    Author
    Feb 7, 2023· 1 reaction

    Hi. You need to download the VAE file then put it in this path *\models\VAE if you are using Automatic1111. Then start Automatic1111 and go to Settings -> Stable Diffusion -> in SD VAE select vae-ft-mse-840000-ema-pruned.ckpt -> Apply Settings

    ShquonkFeb 7, 2023· 4 reactions
    CivitAI

    Any way to train this on the fast-dreambooth colab? I keep getting errors when trying to convert to diffusers.

    SG_161222
    Author
    Feb 7, 2023

    You can find diffusers on Hugging Face: https://huggingface.co/SG161222/Realistic_Vision_V1.3

    ShquonkFeb 7, 2023· 1 reaction

    @SG_161222 thanks! Amazing model btw

    Can someone please explain for dummies that how you can use this model with dreambooth? Tried the version also from Huggingface but still errors. Other models work well so what should be done for this model prior to dreambooth?

    bbackagainnFeb 9, 2023· 1 reaction
    CivitAI

    It seems when i make an image with two people it applies the same description to both, even when they are described separate and very different from each other in the prompt.

    i.e if i say one is japanese, they both will be, or if one is supposed to be a policeman, the others outfit will take on police-like properties. how do i weigh against that in the prompt so that images with two figures can have different looks.

    SG_161222
    Author
    Feb 9, 2023· 2 reactions

    I think this is a problem with many models.
    You can use the inpainting version to put a mask on the person you want to change (not a 100% way).

    bbackagainnFeb 9, 2023

    @SG_161222 does this model use parenthesis and brackets to weigh positively and negatively? or another way perhaps?

    SG_161222
    Author
    Feb 9, 2023· 1 reaction

    @bbackagainn Yes, brackets can be used with this model.

    BlacksmithFeb 9, 2023

    That's quite common. I think it happens with all the models.

    Physical_Arugula_115Feb 11, 2023· 3 reactions
    CivitAI

    Dreambooth:

    I got error while trying to train with this model: https://huggingface.co/SG161222/Realistic_Vision_V1.3

    Error:

    Traceback (most recent call last):

    File "main.py", line 907, in <module>

    if trainer.global_rank == 0:

    NameError: name 'trainer' is not defined

    Any ideas why I got this error?

    Physical_Arugula_115Feb 11, 2023· 1 reaction

    @SG_161222 Thank you for your reply! I'm using JoePenna Repo on Runpod and everything mentioned in that post seems to be ok. I can run 1.2 with no issues: https://huggingface.co/ykurilov/realistic_vision_diff

    Is there some changes from 1.2 to 1.3 affecting to this?

    SG_161222
    Author
    Feb 11, 2023· 1 reaction

    @Physical_Arugula_115 Most likely a bug in the model, I'll try to fix it in a future update.
    The new version will be finished soon and after the tests will be uploaded here and on Hugging Face.

    Physical_Arugula_115Feb 11, 2023· 2 reactions

    @SG_161222 Thank you! You have done great work with the model!

    Kalllop55Feb 16, 2023· 1 reaction
    CivitAI

    I am new to this, can I create a custom model and use it in tons of different scenarios?

    SG_161222
    Author
    Feb 16, 2023

    Yes, you can do that, for example using Dreambooth or some of its counterparts.

    efraxFeb 16, 2023· 1 reaction
    CivitAI

    I love it but idkw it keeps adding "analog style, modelshoot style, nsfw, nudity" to the beginning of my prompts. anyway to avoid this?

    SG_161222
    Author
    Feb 16, 2023

    Can you tell me more about the problem? :)

    efraxFeb 16, 2023

    @SG_161222 when i generate something it doesn't add to the prompt perse it add to the metadata in the image i have one that starts "analog style, modelshoot style, nsfw, nudity analog style, modelshoot style, nsfw, nudity (Masterpiece photograph:1.4),of (Ultrarealistic:1.3), High Detail RAW photo Professional full body shot of...." let me know where i can upload images and they conserve metadata to show you some examples of the problem

    SG_161222
    Author
    Feb 16, 2023

    @efrax What GUI do you use?

    efraxFeb 16, 2023

    @SG_161222 automatic1111 web ui. i thought it was maybe the version but i re download it and same problem

    SG_161222
    Author
    Feb 16, 2023

    @efrax Very strange, I have now checked the metadata and there is no such thing, only the prompt, the negative prompt and the generation parameters. You can upload pictures to https://postimages.org/

    SG_161222
    Author
    Feb 16, 2023

    @efrax In the metadata I found that you use some template that includes these very words (analog style, modelshoot style, nsfw, nudity analog style, modelshoot style, nsfw, nudity), maybe that's the reason?

    efraxFeb 16, 2023

    @SG_161222 no, here some more without template https://postimg.cc/gallery/jrRV2mF

    SG_161222
    Author
    Feb 16, 2023

    @efrax I found the source of the problem. It's the model-keyword extension.

    efraxFeb 16, 2023

    @SG_161222 how can i fix it? i tested with realisticVisionV13_v13 and realisticVisionV13_v13VAEIncluded both safetensors

    SG_161222
    Author
    Feb 16, 2023

    @efrax Try disabling this extension in the Extensions tab of your Automatic1111, then apply these settings using the Apply button and reload the UI

    efraxFeb 16, 2023· 1 reaction

    @SG_161222 it looks like its solved. You are a god. how did you find out without knowing i have that extension?

    SG_161222
    Author
    Feb 16, 2023· 7 reactions

    @efrax Just assumed, downloaded this extension and checked it myself and it gave the same result as yours :)

    leonakaFeb 16, 2023· 3 reactions
    CivitAI

    inpaint doesn't work with controlnet in automatic1111. Has anyone experienced this?

    RuntimeError: Given groups=1, weight of size [320, 4, 3, 3], expected input[2, 9, 80, 120] to have 4 channels, but got 9 channels instead

    SG_161222
    Author
    Feb 16, 2023· 1 reaction

    This may be a problem with version 1.3. I am now preparing a new version for release. After successful tests I will post it together with the inpainting version.

    Alex225Feb 16, 2023· 2 reactions
    CivitAI

    How do I train this model with my photos?

    SG_161222
    Author
    Feb 16, 2023

    You will be able to train the model in your photos with Dreambooth. There are many videos on Youtube on this topic. If you have any problems, feel free to contact me :)

    AztecsFeb 16, 2023· 2 reactions
    CivitAI

    Hey I want to make it a live model, how can I do it ? Like I want to generate images as your model does on mage space ? How do I do that ? I want to make a website like mage with your model only, please guide. Would love to support you as well.

    18394Feb 16, 2023· 2 reactions
    CivitAI

    Can't train this model on fast-dreambooth

    conversion error

    SG_161222
    Author
    Feb 17, 2023· 2 reactions

    You can try to do it with the new version of the model (it should have less problems, but is still being tested): https://boosty.to/evgk.132

    18394Feb 17, 2023· 1 reaction

    @SG_161222 Thank you, i will try it for you <3

    18394Feb 17, 2023· 1 reaction

    @SG_161222 Just want to tell you, your new version of your model is amaaaaaaaaazinnnnnnnnggggggg!!!!!!!!!! keep doing great mate! <3

    SG_161222
    Author
    Feb 17, 2023· 1 reaction

    @lysella Thank you so much for the review :)

    omenizerFeb 19, 2023· 1 reaction

    @lysella could you train the new version? I know these errors, so many models have them :(

    18394Feb 19, 2023

    @omenizer I haven't try it yet

    argus2968Mar 18, 2023· 1 reaction

    Do the following: Checkpoint merger > Primary model A set to v1.5pruned > Secondary model B set to RealisticVision > Set multiplier to 0.99.
    It will now identify the model as v1.5 and not have conversion errors, Works to convert all models if they need it. You can also use other models as the base A model if you know they work without error in dreambooth, I'm just using 1.5 as an example. When multiplier is set to 0.99 there will be practically no effect to model B visually.

    ZcatFeb 17, 2023· 2 reactions
    CivitAI

    I'm getting odd results, tan or wood like colors, with just rand text or just photo like pictures that have nothing to do with the prompt. Anyone know what kind of error that would be?

    SG_161222
    Author
    Feb 17, 2023

    Hi, what GUI are you using?

    girlswithafrosFeb 24, 2023

    I am using NMKD GUI and getting similar results.

    SG_161222
    Author
    Feb 24, 2023

    @girlswithafros What version are you using? The model should run stable on version 1.9.1

    reddotFeb 17, 2023· 3 reactions
    CivitAI

    I don't think the model is working, I'm trying to write a prompt and I'm only getting portraits that don't have anything to do with the prompt, I'm using automatic111

    SG_161222
    Author
    Feb 17, 2023

    What version of the model are you using?

    reddotFeb 17, 2023

    @SG_161222 1.3 the newest one

    SG_161222
    Author
    Feb 18, 2023

    @reddot You can try the new version of the model, which is still being tested and will be improved in the future (should work stable and without errors): https://boosty.to/evgk.132

    PolygonFeb 17, 2023· 1 reaction
    CivitAI

    If I have the VAE already selected in A1111 do I need to turn it off for this model? If so, is there any way we could get a version without the VAE?

    SG_161222
    Author
    Feb 17, 2023

    Hi, you can test the new version of the model, you can find the download link on this page (this model does not have VAE included): https://boosty.to/evgk.132
    Important note: "RAW photo" in the prompt may degrade the result.

    SG_161222
    Author
    Feb 17, 2023

    "If I have the VAE already selected in A1111 do I need to turn it off for this model?" - it does not matter.

    GG_SystemFeb 18, 2023
    CivitAI

    Any idea why i always get multiple copies of the same person , i only want 1 person in the picture , not 3 of the same person

    SG_161222
    Author
    Feb 18, 2023

    Which graphical interface do you use? In which resolution do you generate the image?

    BlacksmithFeb 18, 2023· 4 reactions

    Bigger resolution usually clones the subject in all models. Maybe that's your problem.

    westingtylerFeb 18, 2023· 2 reactions

    i installed and use ControlNet to control the framing and pose, and it save a LOT of trial and error getting it to have the subject the way I want.

    westingtylerFeb 18, 2023· 2 reactions

    @Blacksmith yeah, I think these base models are trained on 512 or 768 resolution images, so going above 768 will start to "tile" the contents it seems.

    hollowmilkyFeb 20, 2023

    @westingtyler can these image be upscaled to higher definition like 1920pixels with good details

    westingtylerFeb 18, 2023· 8 reactions
    CivitAI

    fantastic model and great with hands, but it wants to default to people wearing shorts or skirts, showing their legs and knees. even "49-year-old man school teacher" puts him in a mini-skirt unless I clearly specifi (jeans pants:1.3) and even then it's hit or miss. apart from that, great model.

    SG_161222
    Author
    Feb 18, 2023

    You can also try version 1.4 :)

    The link is in the first post: https://boosty.to/evgk.132

    MarkDopplerFeb 18, 2023· 11 reactions
    CivitAI
    This model is incredible! I have created a web application to be able to train an AI with photos of faces, and be able to use them with this model! https://www.extendimage.ai/train
    SG_161222
    Author
    Feb 18, 2023

    Thank you for your feedback. I am very glad that you like this model, I will continue to work on it.

    danGorstMar 11, 2023

    @MarkDoppler why is the price of training not listed upfront?

    CuboneFeb 19, 2023· 2 reactions
    CivitAI

    Still my favorite model ever. I saw a link to test the 1.4 version but CKPT only, can you make it safetensors? Thank you

    SG_161222
    Author
    Feb 19, 2023

    Wait a while, as soon as I upload the model, I'll tell you.

    SG_161222
    Author
    Feb 19, 2023
    CuboneFeb 19, 2023· 1 reaction

    @SG_161222 Lovely. Thank you

    CuboneFeb 20, 2023· 1 reaction

    @SG_161222 I've been testing it for a few hours and for now I notice that the 1.4 tend to produce empty/white background more often than the 1.3. It's not really a problem but I have to write a prompt about the background too if I use the 1.4.

    SG_161222
    Author
    Feb 20, 2023

    @Tacite Good, thanks for letting me know about it :)

    pangpang0112Feb 20, 2023· 3 reactions
    CivitAI

    Dear Sire, im with diffusionbee on macbook. and every time i tried to import the model(1.3 or 1.4test ckpt), it gives me the following error info: Error Traceback (most recent call last):

    File "convert_model.py", line 28, in <module>

    KeyError: 'state_dict'

    [50607] Failed to execute script 'convert_model' due to unhandled exception!

    SG_161222
    Author
    Feb 20, 2023

    Perhaps this information can help you solve the problem: https://github.com/divamgupta/diffusionbee-stable-diffusion-ui/issues/372

    thesingularartFeb 23, 2023

    Try using a icognito mode

    sdsdcs12qwMar 3, 2023· 1 reaction

    I also used diffusionbee first, performance was not great so I switched to google colab, just follow this tutorial - https://saashed.com/how-to-run-stable-diffusion-for-free-on-google-colab-in-less-than-10-minutes/ I don't understand why people still continue to use stable diffusion on their own GPU when clearly it is not well optimized to run these. With colab, you get free decent GPU access for few hours every day. That should be good for most casual users.

    theonewhodiffusesFeb 20, 2023· 1 reaction
    CivitAI

    How do I post an image here?

    SG_161222
    Author
    Feb 20, 2023

    You can attach it when you want to leave a review :)

    theonewhodiffusesFeb 20, 2023

    Thank you! Can you do multiple reviews?

    SG_161222
    Author
    Feb 21, 2023

    @theonewhodiffuses Perhaps I misunderstood. But you can leave multiple reviews.

    ritcher1Feb 25, 2023· 1 reaction
    CivitAI

    Really a great blend, one of the best !

    Can you upload a smaller pruned fp16 version, without baked vae, for even greater flexibility ?

    SG_161222
    Author
    Feb 25, 2023

    Hi! Unfortunately I won't be able to cut VAE from version 1.3, but you will be able to test version 1.4 beta (without VAE). https://huggingface.co/SG161222/Realistic_Vision_V1.4

    ritcher1Feb 25, 2023

    That's bad. I guess I'll wait for the 1.4 final version: will it be vae baked too ?

    Can you make the actual v1.3 smaller, a 2gb pruned fp16 safetensor version ?

    Anyway, I remember of a script that can remove the vae from vae baked models:

    https://github.com/Akegarasu/sd-webui-model-converter

    "convert/copy/delete any parts of model: unet, text encoder(clip), vae"

    SG_161222
    Author
    Feb 26, 2023

    @ritcher1 Version 1.4 will be without VAE.
    I will try to remove VAE from 1.3 and reduce the model size.

    ritcher1Feb 26, 2023· 1 reaction

    @SG_161222 Thank you very much for your efforts. Waiting for any news.

    SG_161222
    Author
    Feb 26, 2023

    @ritcher1 I uploaded the pruned model to the website. I tried to remove the VAE, but I don't think it worked.

    ritcher1Feb 26, 2023· 1 reaction

    @SG_161222 Thanks, just downloaded. I'd change the filename in realisticVision_v13_pruned to remind the users what they got.

    SG_161222
    Author
    Feb 26, 2023

    @ritcher1 Renaming the file does not help in this case :)

    ritcher1Feb 26, 2023

    @SG_161222 It's said that the pruned versions aren't so good in creating new mixes, so knowing if a model is full or pruned could help.

    isocornyFeb 26, 2023· 1 reaction
    CivitAI

    I'm new around here. What folder inside of model folder do I place the download file? I am using stable-diffusion-webui. Thanks and sorry for the n00b question

    SG_161222
    Author
    Feb 26, 2023· 2 reactions

    stable-diffusion-webui\models\Stable-diffusion

    isocornyFeb 26, 2023· 1 reaction

    @SG_161222 I was just about to post this as a reply. Thanks so much, works like a charm

    CuboneFeb 26, 2023· 1 reaction
    CivitAI

    Where can I find R-ESRGAN General WDN 4xV3 please?

    SG_161222
    Author
    Feb 26, 2023· 1 reaction

    Hi. If you use Automatic1111:
    Settings -> Upscaling -> Turn on the R-ESRGAN General WDN 4xV3 -> Apply Settings -> Restart Automatic1111

    CuboneFeb 26, 2023· 1 reaction

    Thanks @SG_161222 

    CuboneFeb 27, 2023· 2 reactions
    CivitAI

    Any tips on getting better eyes/pupils with this model? Most of the time, even if I follow the high-res fix tips, I just get very low detailed ones

    SG_161222
    Author
    Feb 27, 2023· 1 reaction

    You can use this information: How to Fix Face and Hands with Inpaint

    magalie301443783Feb 27, 2023· 1 reaction
    CivitAI

    Я тебя добавил VK, хочу тебе сообщение отправить.

    kleerp2018Mar 2, 2023· 1 reaction
    CivitAI

    Curious: Do most people on here generate images locally on their computers or do they use a generator website?

    sdsdcs12qwMar 2, 2023

    some on colab or similar websites, others on personal computers.

    kleerp2018Mar 3, 2023· 1 reaction

    @sdsdcs12qw I run codeformer and gfpgan on colab - not the fastest, but it works. You can run models on colab? Is that entirely pc dependent? I guess I'm not sure what resources colab is using or if they're on some server in google-land. I've used openart, dreamlikeart, getimg, dall-e....not even sure what other sites are good. But I would like to know how the colab thing works for these models if you know...thanks so much for replying!

    sdsdcs12qwMar 3, 2023· 6 reactions

    @kleerp2018 Yeah, you can easily run this or any custom models on google colab. A free colab account is fine too. Very decent performance and can use it for roughly 3-4 hours daily max before google takes away the GPU. Just follow this guide to use colab with stable diffusion https://saashed.com/how-to-run-stable-diffusion-for-free-on-google-colab-in-less-than-10-minutes and this one to download civitai models to colab https://saashed.com/how-to-easily-download-use-custom-stable-diffusion-models-from-civitai-in-google-colab/

    kleerp2018Mar 5, 2023· 1 reaction

    @sdsdcs12qw Thanks so much! I assumed it was running off one's own gpu so I never tried anything more than gfpgan and codeformer. I'll give it a try! Thanks again!

    SejfMar 11, 2023

    @kleerp2018 I run on my own pc, I have a 2060 with 6gb vram you can run stable diffusion webui on your pc if you have at least 4gb vram with the --medvram parameter.
    https://github.com/AUTOMATIC1111/stable-diffusion-webui

    Boxer766Mar 16, 2023

    local on my laptop with 16gb vram

    rogueAIMar 18, 2023· 1 reaction

    I generate images locally on a 3080 TI with 12GB of VRAM. It bogs down the video card when generating, but it only takes a few seconds or a couple minutes to generate a single image depending on the resolution and number of steps.

    AlexparkerMar 5, 2023· 3 reactions
    CivitAI

    Thanks for this Great Model !!!

    and please,why does the my character always have to be facing the front and looking at the camera(viewer)?What can i do to change that?should i use some specific words in the prompt?like what?

    thanks again !!!peace

    SG_161222
    Author
    Mar 5, 2023· 1 reaction

    Hi, thank you. I will try to fix this problem in a future update. For now you can experiment with the following words in the prompt: [front view|side view]

    RandmeistMar 6, 2023· 6 reactions

    Try not to use "close-up" and "portrait", use just a photo and maybe these, inbrackets:

    Wide-Angle Shot - Pull the camera back from subject

    Ultra-Wide Angle - Pull the camera very far back from subject

    Satellite View - Pull the camera into the stratosphere

    Eye-Level Shot - Camera even with head of figure

    Far-Shot Angle - Figure looks very small/distant

    Medium-Shot Angle - Looking up at figure from knee level (child gazes up at parent)

    Ground-Shot Angle - Looking down at figure from knee level (parent gazes down at child)

    Low-Angle Shot - On ground, looking up at figure (puppy gazes up at human)

    Full-Shot Angle - Your target in full view

    Full-Body Shot - Head-to-toe view, like full-shot angle

    Glamour Shot - Different zooms, but always flattering to figure

    Cinematic Still Shot - Emphasis on context/background
    Also you can use controlnet with depth/hed/openpose models to catch up a pose that you like from any photo

    AlexparkerMar 6, 2023· 1 reaction

    @SG_161222 thanks pal!looking forward to your update!

    AlexparkerMar 6, 2023· 1 reaction

    @RoninRex Really helpful !thanks man! 

    175737Mar 9, 2023· 2 reactions

    OpenPose Editor + ControlNet can give you a bit more control on poses and head positioning. Something to look into.

    You can also try giving the prompt specific features to focus on. I've used (back, butt:1.3) before. A bit crude lol but it works. The 1.3 weight helps to focus the output.

    You could also try adding these as the first few words in your positive prompt: (1/4 profile shot) or (profile shot) or even (editorial photograph)

    ing170Mar 10, 2023

    @ovalshrimp 👍👍👍

    xin811Mar 10, 2023

    @ovalshrimp cool

    BedlamTheBardMar 7, 2023· 1 reaction
    CivitAI

    I can't get it to make a photo where it's a dark bedroom. No matter what I do there's lots of sunlight. Trying to make it look like a selfie taken at night. Any help?

    SG_161222
    Author
    Mar 8, 2023

    This LoRA can help you with that: https://civitai.com/models/13941/epinoiseoffset

    175737Mar 8, 2023· 1 reaction

    One trick you could use to darken a picture is controlnet. It's a little complicated but it will work. First, generate your image then send it to img2img. Within img2img, open controlnet and import that image you just generated into the controlnet panel. Set the preprocessor to depth and the model to depth. Leave all the other settings alone in controlnet. Now, grab a picture of a random black background from online and add that to img2img pane at the top of the webui. You'll be removing your picture that was added there and replacing it with the black background. You still want your original picture in controlnet, though. So, black picture up top, your generate image down below in controlnet. Now, Click generate. You should see a dark version of your image. You can then adjust the denoising strength up or down to fine tune it. Lowering it will make it darker. This method actually works great in tandem with the offset noise lora.

    gaydiffusionMar 8, 2023· 30 reactions
    CivitAI

    "Commercially licensed" for fantasy ai but you can't even credit all the model makers who did the actual work for you, how do you know if you can even license this in the first place considering that? I suppose if I make a merge but forget to credit you and then commercially license it in turn, that's fine? lmfao

    SG_161222
    Author
    Mar 9, 2023

    The model is still free and under the creativeml-openrail-m license.

    About the models I listed in the description: I merged my model for myself and did not plan to release it to the public, so I can not remember which models were used and which were not.

    If you have any more questions, I'll try to answer them.

    twistedellieMar 10, 2023

    @SG_161222 If it falls under creativeml-openrail-m license that means anyone can still commercially use the model. So what's with fantasy AI's claim about commercial use exclusivity? Do they only mean image generation sites but not individuals?

    SG_161222
    Author
    Mar 10, 2023· 2 reactions

    @twistedellie Hi! This will only affect image generation services. You can still generate images that you can sell, and it doesn't limit the use of the model in any way.

    If you have any more questions, feel free to ask :)

    twistedellieMar 10, 2023· 2 reactions

    @SG_161222 Ah, that clears up a lot. You probably want to put that in your description to dissolve many people's worries. Thanks for answering. ^^

    Syn_tzuMar 8, 2023· 5 reactions
    CivitAI

    Can anyone advise me. I've downloaded all model of 1.3. I load them into InvokeAI successfully. However rendering and image always produces something from an entirely different genre. Example, prompt = portrait of a woman. Results are images of money, or a piece of furniture, a vase or anything from the 1800 time period. Help!

    xektopMar 28, 2023

    I am not sure about how it's called in invoke AI, but in Automatic1111 the option is called CFG and you have to put it higher number so it generates things closer to your prompt. If you use low CFG your prompt may get completely ignored.

    ReeserhhMar 9, 2023· 2 reactions
    CivitAI

    So exactly is fantasy.ai claiming rights to a model merged so many times which models weresed aren't even remembered?

    SG_161222
    Author
    Mar 9, 2023

    The model is still free and under the creativeml-openrail-m license.

    About the models I listed in the description: I merged my model for myself and did not plan to release it to the public, so I can not remember which models were used and which were not.

    If you have any more questions, I'll try to answer them.

    fbmacMar 9, 2023· 4 reactions

    There are models on fantasy.ai that don't even disclose what they merged, let's not punish @SG_161222 for being honest. Fantasy know what they are buying, and they are the ones that should deal with it. It's impossible to check the license on this merges of merges of merges.

    If we do everything by the book, none of the popular models can be hosted commercially by anyone.

    IANAL, but my suggestion to fantasy.ai and anyone thinking about doing this is use trademarks instead of copyright. You should probably avoid trained copyrighted models like dreamlike and novel ai, or consider getting a license from them. If you can't see if the stuff you merged has it, they probably can't either. But test how your model react to their trigger words to be sure.

    7727Mar 9, 2023· 3 reactions

    @fbmac I’m sorry, but if a company like fantasy.ai is going to charge anything at all, they should be training their own models from scratch. I’ve also never understood the Dreamlike license thing, seems like an honor code thing and not actually enforceable after a simple PNG info removal.

    Just trying to host free models and showcasing the names of them on their homepage is kinda a joke. The ONLY people that have any clue what those model names are, are not the target market for paid Stable Diffusion. This shit is already free, and totally untraceable when it comes to the model used. Licenses on any of these models are an unenforceable joke, ESPECIALLY merges.

    As a photographer though, I could see training a model on my own work and charging for that with a license, but first it’d have to be good enough to warrant payment over the endless free options, and then I’d have to distribute somewhere other than Civitai so that I could actually sell it initially. And even then, the first person to buy it would just upload it here. And next it’s 10 layers deep in a merge someone else is trying to sell.

    People are just desperately trying to figure out how they can monetize this before someone else and throwing horrible websites up. Like, we get it Fantasy.ai, you know how to change fonts.

    wendaleMar 9, 2023· 1 reaction
    CivitAI

    I saw the latest update today (March 9th), but they are all v1.3, so what is updated? Where can I see update records?

    SG_161222
    Author
    Mar 9, 2023

    Hi. The description has been updated.

    wendaleMar 9, 2023

    @SG_161222 emmm... where is it? I don't see anything different in the description than before. :(

    SG_161222
    Author
    Mar 9, 2023

    @wen834 New information in the text about the "Official Partner" :)

    ManuManuMar 15, 2023· 1 reaction

    @SG_161222 publish your v1.4 here. community will be happy

    SG_161222
    Author
    Mar 15, 2023

    @ManuManu I'll do it tomorrow :)

    skullzy77Mar 17, 2023

    @SG_161222 Still doing it?

    SG_161222
    Author
    Mar 17, 2023· 1 reaction

    @skullzy77 I won't be able to do it yet until my weekend. Version 1.4 is available on Hugging Face: https://huggingface.co/SG161222/Realistic_Vision_V1.4_Fantasy.ai

    skullzy77Mar 17, 2023

    @SG_161222 Oh nice! Thanks

    7727Mar 9, 2023· 42 reactions
    CivitAI

    Lmao. “Licenses.”

    The second you put a model on Civitai or HF for download, you can kiss that license goodbye.

    The ONLY way anyone is ever going to monetize their model, is if it’s exclusively locked down within their own webui environment that keeps record of generations.

    The second you give that model out to the public, it’s open season. You cannot prove what I used to make my own images locally, especially once scrubbed of any PNG info.

    Unless you’re training models from scratch/only merging with other models YOU made from scratch, again, “licensing” it is a joke. You also need to be the exclusive owner of any content you train the model on to truly be the license owner. And even then, you better keep that source ckpt on lockdown before it gets merged a million times without evidence.

    Training a model from scratch on someone else’s work doesn’t count either. Which is what all of the “big” models are.

    I laughed when I uploaded an embedding of a person here and it asked me about licensing. Imagine owning a commercial license for an embedding of someone else’s face. This is all legal *make-believe*.

    Please stop putting license talk in these uploads until they’re actually enforceable.

    davizcaMar 9, 2023

    Same thought and if these people keep doing these licenses jokes, don't worry, more people will jump with better models and will wipe them out, period.

    In fact a lot of "licensed" model are using protogen models (among others in the mix), this is a fucking joke. I will only respect licenses for those who really took time to drew these images and train them as a model, if not, they're just training over MJ/SD generations so please, as an artist in the industry, this is total lame and you should start looking more at the community instead of trying to make as muas as $$$ possible.

    ManuManuMar 10, 2023

    People are desperate to make money, smh.

    "Look guys I merged 10 different models in civitAi but I forgot the rest of them, so now if you use the my checkpoint please pay three fiddy $$".

    It's like he's owning all of the images that all of the models that were trained upon.

    SG_161222
    Author
    Mar 10, 2023· 5 reactions

    @ManuManu Now show me where I demand money to use the model?

    LoveLunaMar 27, 2023

    Problem is the training data is a legal quagmire. For this same reason, industry (film, entertainment) is unable to use AI generated content for commericial production until these legal issues are sorted out (if ever).

    7727Mar 27, 2023

    @Fyyre already hitting commercial markets actually https://www.engadget.com/levis-will-supplement-human-models-with-ai-generated-fakes-190011557.html

    The AI company they’re using looks to be creating their own training data at a quick glance

    https://lalaland.ai

    1111231Mar 11, 2023
    CivitAI

    exclusive 1 star for exclusive fatnasy.ai

    anonyvpn001962Mar 13, 2023· 2 reactions
    CivitAI

    why many peopel use SD1.5 instead of 2.0?

    ManuManuMar 13, 2023· 6 reactions

    1.5 is way better than 2.0

    openglsMar 14, 2023· 18 reactions

    Short answer: people proompt for coom and SD 2.x has no nudity, so people proompt in 1.5 instead.

    Long answer: SD 1.5, being a 512x512 base model, has smaller VRAM requirements compared to 2.x, so naturally more people can use it and therefore why it is so popular. SD 2.1768 produces photo-realistic outputs with far better fidelity, but it's a 768x768 base model, meaning it has more robust VRAM requirements. Never mind the fact that StabilityAI took some... questionable decisions in the direction of SD 2.0: they removed the ability to prompt for artists art styles, celebrities and nudity, which the community did not take lightly. They later tried to circumvent the backlash by reintroducing nudity to SD 2.1, but that was too little; too late. All in all, I find 1.5 amazing for illustrations, paintings, drawings, anime, because you can prompt for an artist and have the art in their style (like the infamous Greg Rutkowski), but for photo-realistic outputs, everything looks kind of fake, and this mix is no exception. Meanwhile, I find 2.1 awful for art/illustration, but infinitely better for photo-realism, and its finetunes (SciFi v2.1, Illuminati Diffusion) are far more photo-realistic (although these finetunes are all SFW-only).

    anonyvpn001962Mar 14, 2023

    @opengls thanks for your reply.

    Rin_chanMar 18, 2023· 6 reactions

    @opengls Even 2.1 literally plasts random splotches of color to censor nipples, and good luck having actually nude humans, so nope. And camera models, art styles AND famous people likenesses removed, what's left, landscapes? Nobody cares for AI-created paintings of random grass fields, AI exists to make things you can't do otherwise, fun stuff like picture of US presidents as Warhammer marines or Oscar nominees drawn by Van Gogh, this kind of stuff...

    vuzzMar 26, 2023

    @anonyvpn001962 It is not just about nudity and artist styles, SD 2.x models appear to be way less steerable when using unusually complex prompts (such as verses, songs etc.). Overall, less degrees of (artistic) freedom, and probably more fawning to corporate interests.

    Rin_chanMar 26, 2023

    @vuzz I've read about it and the reason is simple, they used OpenClip as a base, which works better BUT is limited to open source/non-copyrighted work so say good bye to stuff while trying to clear reputation of being zOMG evil copyright infringers. The weirder choice is the ridiculous NSFW limit. But that is avoided in custom models, the lack of things like modern artist, photographer and device names in dataset is not. And the 1.4/1.5 models of SD, DALL-E 2, Midjourney etc. all use regular Clip thingy, which is suckier but since it's closed source they don't actually know what's inside though peopel figured out by trial and error.

    Easy way to check is inputting "photo by [photographer name]" and see if it shows photo typical for them, or photographic equipment, i.e. Steve McCurry will show women in gritty Afghanistan photos. He's in dataset/ "photo by Kim Anderson" shows a photography camera. So the name isn't in the data. Interesting case is "by Annie Leibovitz" shows her face in 1.5, meaning dataset knows her as celebrity and not a photographer.

    They ALSO have different "weight" to them so some artists, styles, names, whatever are pulling the picture heavier. There's a huge database of artists people made with how they influence images. But all of that is reverse engineering, we don't have the data CLIP has, we wing it and guess. For 2.0, OpenClip data IS known so devs find it easier to train and to know what's up in general. It SHOULD have been upgrade, but it wasn't.

    From what I gather, 2.0 DOES train on artists and photographers, celebrities etc., it just doesn't label them so doesn't "know" that by prompts. I don't even know if it can be fixed with custom datasets.

    What I tried to explain in this long ass post is that it's more of a case of "good intentions" paving road to hell than StabilityAI just randomly middle-fingering customers to please corporate overlords. TBH, none of this should be even possible to copyright at all, except for code of programs/methods themselves and stuff like sampling algorithms that can be patented (and has academic documentation for them). After all, AI is learning on things, not copying them. You can't really ban learning, that would open a bad can of worms nobody wants to go into, like can you now persecute artists from learning art they seen in gallery? The argument "but AI fast" doesn't hold legal candle here.

    But hey, we've seen worse abuse of law, and I can't blame devs of Stable Diffusion for wanting to be more transparent and placate the idiots in parliaments and various judges, especially after those keep proving themselves being incompetent boomers who don't understand business and copyright 101. See Zuckerberg in American parliament thingy being asked dumbest questions since Rob Halford was trying to exhale backwards in court to prove rock doesn't make kids off selves smh. It's PMRC and Napster stuff all over again.

    vuzzMar 27, 2023

    @Rin_chan Thank you for the elaborate and concise development, much appreciated.

    I wasn't thinking about "copyright" when mentioning corporate needs, actually, it transcends AI, we could call it "political" or "ideological", even "philisophical" as in neoplatonism : What I see is a more general pattern of data suppression / misrepresentation / distorsion, repeated each time it doesn't fit some preset narrative. Getting into the details, and examples, would bring us far from the specific question in this thread.

    Just to answer your point, Rin_chan, learning suppression, as well as its corollary, burning books, (and sometimes people) have been attempted many times in history, and the worst is that as tactical devices they have worked efficiently in delaying or derailing societal processes.

    As a recent example of un-learning, just consider the slow (3 generations) eviction of the study of the ancient authors and "dead languages" from the western curriculum. A whole database of political, religious and behavioral experiences got slowly out of view of the commoner, pushed into the realm of "specialists". The consequences are both tragic... and quite lucrative, so as a rule of thumb you'll meet some people very willing to go that way and open that can of worms.

    vuzzMar 27, 2023

    @Rin_chan According to this paper, OpenCLIP aint that bad at all. https://arxiv.org/pdf/2209.06103v1.pdf

    Dwells on the importance of dialing in class information, more than details, this is an useful hint.

    "Our analysis shows that: (i) most of the classes in popular zero-shot benchmarks are observed (a

    lot) during pre-training; (ii) zero-shot performance mainly comes out of models’ capability of recognizing class labels, whenever they are present in the text, and a significantly lower performing capability of attribute-based zeroshot learning is only observed when class labels are not used; (iii) the number of the attributes used can have a significant effect on performance, and can easily cause a significant performance decrease."

    anonyvpn001962Mar 14, 2023· 6 reactions
    CivitAI

    why many peopel use SD1.5 instead of 2.0?

    rasterizerMar 14, 2023· 3 reactions

    2.0 purged quite a lot of keywords especially those related to specific artists and NSFW content, so while it might be "better" if you're creating what Stability creator wants you to do, it's less than ideal otherwise.

    leveroz22Mar 28, 2023

    because not all models and extensions can work on 2.0

    SantaonholidaysMar 14, 2023
    CivitAI

    What the fuck happened to the Publisher? Reddit got deleted.1.4 Version is no longer existant.The Images here are gone also

    yourowndesiresMar 15, 2023
    CivitAI

    i create a perfect character but now how do i make more pictures with the same character? i train the machine?

    NjakkiMar 15, 2023· 4 reactions

    You can either grab the seed from the generation with png analyzer, and continue tweaking (although that will not be completely replicable always, since if you change enough keywords, it will change), or the better option, which is to take it into img2img.

    RalFingerMar 17, 2023· 35 reactions
    CivitAI

    I know you are mad because of fantasy.ai, but rating this model 1/5 stars is just more work for us moderators. Can you please find a different way of venting? We have enough to moderate and don´t need that kind of nonsense on top, thank you!

    harmonicdiffusionMar 17, 2023· 16 reactions

    let people express their mind like a free society.

    denislMar 17, 2023· 11 reactions

    perhaps you should stop moderating because it's a valid concern. We're not hating personally on the creator and in fact want them to get paid so I don't see why you're blocking all bad reviews. If users are unhappy with the model or its conditions, users are unhappy and by moderating this hard, you're undermining your own credibility.

    denislMar 17, 2023· 6 reactions

    maybe you can add a warning sign behind the true score for this model saying that the average may not be representative of the model's generation quality as it is also judged by its licensing conditions.

    RalFingerMar 17, 2023· 6 reactions

    @denisl just take a look at the comments of the 1/5 ratings, it's not related to the model. Nothing gets removed or moderated, all the reviews and comments are still there. The rating is just not counting.

    gaydiffusionMar 17, 2023· 4 reactions

    Reddit gonna Reddit. At least have the dignity of leaving commentary in a Comment, not a Review. No need to tank a rating, which should be the quality of a model, not its usage.

    denislMar 18, 2023· 6 reactions

    @RalFinger it is related to the model, more specifically to the license conditions behind it. By the same logic, in the future, you shouldn't allow to take into account downvotes on some model with some ridiculous license that makes the use of the model tricky. So yeah, I think you should let the rating count.

    SG_161222
    Author
    Mar 18, 2023· 3 reactions

    @denisl What license do you keep talking about?
    And how does this license (creativeml-openrail-m) prevent you from using the model?
    Or are you talking about some obscure license that doesn't exist here? :)

    GodbearmaxMay 14, 2023

    @denisl It very much depends on what they want here. Allow ratings just to have them? No matter what gets rated on? Or ratings regarding only the quality of the model....I very much would also want to have ratings for the modelquality but on the other hand ratings dont matter that much anyway

    RalFingerMay 14, 2023

    @Godbearmax the rating system will get an overhaul within the next weeks. If you like to leave feedback or how to improve the site, please use https://civitai.canny.io/feature-requests

    mramer723Mar 17, 2023· 8 reactions
    CivitAI

    Thank you for keeping this model free, please don't change! Love you!

    SG_161222
    Author
    Mar 18, 2023· 3 reactions

    Thank you :)

    This model have been and will be free.

    I already have version 2.0, but it's in the process of testing now. I will be able to release this model here soon.

    Have a great day.

    RandmeistMar 19, 2023

    @SG_161222 А есть в ней nsfw фильтр или нет ?

    SG_161222
    Author
    Mar 19, 2023

    @RoninRex Realistic Vision 2.0 основана на SD1.5, так что nsfw фильтра нет :)

    RandmeistMar 19, 2023

    @SG_161222 а, я думал основа 2.0) из улучшений просто больше база фото на которых была тренировка?

    SG_161222
    Author
    Mar 19, 2023

    @RoninRex Я пока не могу позволить себе тренировку. Realistic Vision V2.0 это смесь Realistic Vision V1.3 и Realistic Vision V1.4 :)

    eabdullaha92154Mar 18, 2023· 13 reactions
    CivitAI

    fuck fantasy.ai

    vlkMar 20, 2023· 1 reaction
    CivitAI

    Thank you for the pruned version of the model! Now I can keep more ControlNet models installed on Paperspace.

    TokendreamsMar 20, 2023· 2 reactions
    CivitAI

    What is the difference between the normal model and the in-painting model? I am very interested in knowing how to train an optimized model for in-painting.

    SG_161222
    Author
    Mar 20, 2023· 1 reaction

    Hi! The inpainting version of the model handles inpaint and outpaint operations much better. The inpainting model is not trainable as far as I know.

    I think you need to train a normal model and then make an inpainting version out of it. There is a guide on this topic: https://www.reddit.com/r/sdforall/comments/zyieht/how_to_turn_any_model_into_an_inpainting_model/

    frayoshiMar 20, 2023· 3 reactions
    CivitAI

    I'm confused.. license declared here says one thing.. and then there's a link to the stable-diffusion-license in there too... which one is the real one? is there a downloadable pdf with the license to study?
    Thanks!

    SG_161222
    Author
    Mar 21, 2023

    The real license is "creativeml-openrail-m"

    Alex225Mar 24, 2023· 3 reactions
    CivitAI

    Why when I use this model to train my face or another image the file gets 1.99gb instead of getting bigger? Also, I can't train two faces in my model at different stages. why would it be? I use FastDreambooth

    92056Mar 24, 2023· 1 reaction
    CivitAI

    What's the goal/improvements of V1.4?

    SG_161222
    Author
    Mar 24, 2023

    The goal was to improve skin texture. But version 2.0 is coming out soon (this Sunday). There are more improvements compared to version 1.3. You can see it here.

    ManuManuMar 24, 2023

    @SG_161222 show samples of hands & fingers please

    92056Mar 24, 2023· 1 reaction

    @SG_161222 Excellent work! The skin is definitely more realistic and less airbrushed

    Checkpoint
    SD 1.5

    Details

    Downloads
    20,438
    Platform
    CivitAI
    Platform Status
    Available
    Created
    2/5/2023
    Updated
    5/13/2026
    Deleted
    -
    Trigger Words:
    analog style
    modelshoot style
    nsfw
    nudity

    Files