CivArchive
    TTPLanet_SDXL_Controlnet_Tile_Realistic - v2.0_fp16
    NSFW
    Preview 12165916
    Preview 9784776
    Preview 10097060
    Preview 9779584
    Preview 9781816
    Preview 9779585
    Preview 9779610
    Preview 9779646
    Preview 9779781
    Preview 9779889

    Thanks for your attention.

    contact me if you want, discord with "ttplanet", Civitai with "ttplanet"

    you can also join the group discussion with QQ group number: 294060503

    update for rank256 version, please notice it will save the VRAM, but it will decrease the Quality!!!

    please refer to the following pic to decide which version you will use:

    the logo of Farai, V2>V1>rank256

    the material of carbon fiber. V2=V1>rank256


    update Preprocessor for comfyui

    you can find it here:https://civarchive.com/models/426501/60sec-process-for-4k-resolution-t2i-with-rtx4090-and-tile-model

    in the workflow package

    Latest Update 2024/4/13:

    Here's a refined version of the update notes for the Tile V2:

    -Introducing the new Tile V2, enhanced with a vastly improved training dataset and more extensive training steps.

    -The Tile V2 now automatically recognizes a wider range of objects without needing explicit prompts.

    -Strong text reorganization, can keep the most clear text with style transfer process.

    -I've made significant improvements to the color offset issue. if you are still seeing the significant offset, it's normal, just adding the prompt or use a color fix node.

    -The control strength is more robust, allowing it to replace canny+openpose in some conditions.

    If you encounter the edge halo issue with t2i or i2i, particularly with i2i, ensure that the preprocessing provides the controlnet image with sufficient blurring. If the output is too sharp, it may result in a 'halo'—a pronounced shape around the edges with high contrast. In such cases, apply some blur before sending it to the controlnet. If the output is too blurry, this could be due to excessive blurring during preprocessing, or the original picture may be too small.

    Enjoy the enhanced capabilities of Tile V2!

    This is a SDXL based controlnet Tile model, trained with huggingface diffusers sets, fit for Stable diffusion SDXL controlnet.

    1. It is original trained for my own realistic model used for Ultimate upscale process to boost the picture details. with a proper workflow, it can provide a good result for high detailed, high resolution image fix.

    2. As there is no SDXL Tile available from the most open source, I decide to share this one out.

    update for style change application instruction and upscale simple work flow:

    update the style change workflow for comfyui:

    https://openart.ai/workflows/gJQkI6ttORrWCPAiTaVO

    Part1 for style and background change application:

    1. Open a A1111 webui.

    2. select a image you want to use for controlnet tile

    3. remember the setting is like this, make 100% preprocessor is none. and control mode is My prompt is more important.

      :

    4. type in the prompts in positive and negative text box, gen the image as you wish. if you want to change the cloth, type like a woman dressed in yellow T-shirt, and change the background like in a shopping mall,

    5. Hires fix is supported!!!

    You will get the result as below:

    Part2 for ultimate sd upscale application

    Here is the simplified workflow just for ultimate upscale, you can modify and add pre process for your image based on the real condition. In my case, I usually make a image to image with 0.1 denoise rate for the real low quality image such as 600*400 to 1200*800 before I through it into this ultimate upscale process.

    Please add IPA process if you need the face likes identical, please also add IPA in the raw pre process for low quality image i2i. Remember, over resolution than downscale is always the best way to boost the quality from low resolution image.

    https://civarchive.com/models/333060/simplified-workflow-for-ultimate-sd-upscale

    This is a SDXL based controlnet Tile model, trained with huggingface diffusers sets, fit for Stable diffusion SDXL controlnet.

    1. It is original trained for my own realistic model used for Ultimate upscale process to boost the picture details. with a proper workflow, it can provide a good result for high detailed, high resolution image fix.

    2. As there is no SDXL Tile available from the most open source, I decide to share this one out.

    image/pngimage/pngimage/pngimage/pngimage/pngimage/png

    • Developed by: TTPlanet

    • Model type: Controlnet Tile

    • Language(s) (NLP): No language limitation

    Uses

    • Important: Tile model is not a upscale model!!! it enhance or change the detail of the original size image, remember this before you use it!

    • This model will not significant change the base model style. it only adding the features to the upscaled pixel blocks....

    • --Just use a regular controlnet model in Webui by select as tile model and use tile_resample for Ultimate Upscale script.

    • --Just use load controlnet model in comfyui and apply to control net condition.

    • --if you try to use it in webui t2i, need proper prompt setup, otherwise it will significant modify the original image color. I don't know the reason, as I don't really use this function.

    • --it do perform much better with the image from the datasets. However, everything works fine for the i2i model and what is the place usually the ultimate upscale is applied!!

    • --Please also notice this is a realistic training set, so no comic, animation application are promised.

    • --For tile upscale, set the denoise around 0.3-0.4 to get good result.

    • --For controlnet strength, set to 0.9 will be better choice

    • --For human image fix, IPA and early stop on controlnet will provide better reslut

    • --Pickup a good realistic base model is important!

    • blurry recovery:

    • image/jpegimage/png

    • cloth change but keep the pose and person:

    • Besides the basic function, Tile can also change the picture style based on you model, please select the preprocessor as None(not resample!!!!) you can build different style from one single picture with great control!

      image/png

    Bias, Risks, and Limitations

    • No commercial USE!!!! Do not use it for adult content

    Recommendations

    • Use comfyui to build your own Upscale process, it works fine!!!

    • Special thanks to the Controlnet builder lllyasviel Lvmin Zhang (Lyumin Zhang) who bring so much fun to us, and thanks huggingface make the training set to make the training so smooth.

    Model Card Contact

    contact me if you want, discord with "ttplanet", Civitai with "ttplanet"

    Description

    Here's a refined version of the update notes for the Tile V2:

    -Introducing the new Tile V2, enhanced with a vastly improved training dataset and more extensive training steps.

    -The Tile V2 now automatically recognizes a wider range of objects without needing explicit prompts.

    -I've made significant improvements to the color offset issue. if you are still seeing the significant offset, it's normal, just adding the prompt or use a color fix node.

    -The control strength is more robust, allowing it to replace canny+openpose in some conditions.

    If you encounter the edge halo issue with t2i or i2i, particularly with i2i, ensure that the preprocessing provides the controlnet image with sufficient blurring. If the output is too sharp, it may result in a 'halo'—a pronounced shape around the edges with high contrast. In such cases, apply some blur before sending it to the controlnet. If the output is too blurry, this could be due to excessive blurring during preprocessing, or the original picture may be too small.

    Enjoy the enhanced capabilities of Tile V2!

    ![TBT9$5UL`53RKP`85JXIZ_H.jpg](https://cdn-uploads.huggingface.co/production/uploads/641edd91eefe94aff6de024c/yS1ax7FWZS7b5Zz1co8_b.jpeg)

    ![Q5A0[{{0{]I~`KJFCZJ7`}4.jpg](https://cdn-uploads.huggingface.co/production/uploads/641edd91eefe94aff6de024c/HMGmYz7IiLSqfoiMgcmgU.jpeg)

    FAQ

    Comments (59)

    justinnovelty716Apr 13, 2024· 1 reaction
    CivitAI

    牛牛牛

    ChenkinApr 13, 2024· 1 reaction
    CivitAI

    伟大的工作

    EricRollei21Apr 14, 2024· 2 reactions
    CivitAI

    Thank for you this tile ControlNet. I can not tell the difference between the 1st and 2nd version, sometimes 1st version looks better particularly with faces/eyes. What is the suggested applied strength and time steps for tiled upscaling?

    ttplanet
    Author
    Apr 14, 2024· 1 reaction

    v1 and v2 has different parameters, if setup correctly V2 will show much better result. V2 has much less error on material texture, and has the nsfw capabilities.

    yifang0826Apr 15, 2024
    CivitAI

    请问是否可以用diffusers加载呢?

    MIHU2077Apr 15, 2024· 2 reactions
    CivitAI

    放大是怎么使用呢?

    olivettyApr 15, 2024· 2 reactions
    CivitAI

    Hello Trainer! Most of the time V1 gives much better results! I am using GaussianBlur at radius 3 in Comfy for comparison. TileV2 is stronger, so might be just a change of weights that are needed. Will explore more! Thank you for all your hard work! <3

    ttplanet
    Author
    Apr 15, 2024· 1 reaction

    Thanks I noticed too. because the weight are different, and I modified the sample dataset, so now it will need more accurate adjustment on image preprocess, most blur node in comfyui can’t do it! so I am trying to build a node to do it. Current tile preprocess in comfyui is terrible.

    olivettyApr 21, 2024

    @ttplanet what is the preprocess in A1111? I thought it was only gaussian blur as well? Thanks for looking into it! <3

    ttplanet
    Author
    Apr 22, 2024

    @olivetty Not really, but similar

    NothingButThymeApr 15, 2024
    CivitAI

    can someone please explain how to use this in comfyui?

    ttplanet
    Author
    Apr 18, 2024

    There are multiple way you can use, see the related resource below to use it

    mnemicApr 19, 2024

    @NothingButThyme

    Workflow included in my image in the gallery below.

    Key settings:
    Scaled soft weights: 0.92
    Control Net Strength: 0.9
    End percent: 0.9

    katsuro1626610Apr 19, 2024· 3 reactions
    CivitAI

    Would you like provide how to construct the train data, condition is the blur or the small size image of the target image?

    changpeidong2148Apr 20, 2024· 4 reactions
    CivitAI

    不会用,也没更详细的说明

    ZharulApr 21, 2024· 5 reactions
    CivitAI

    what a wonderful gift. This model is just awesome. It works beautifully with high noise samplers to make even more details and keeping the picture composition and format much better than Controlnet CANNY or DEPTH. thank you!

    greasebigApr 25, 2024

    no, it is not work on canny

    ttplanet
    Author
    Apr 26, 2024

    @greasebig you don't need a canny preprocessor for this one

    kiryanton930Apr 21, 2024· 4 reactions
    CivitAI

    I still don't understand how it works

    olivettyApr 21, 2024

    Just like regular tile for 1.5 ? Just it needs a gaussian blurred image going in, so either preprocess in A1111 or use blur node in comfy? :)

    ttplanet
    Author
    Apr 22, 2024

    @olivetty slightly different logic on preprocessor, I built a comyui node. For now you can use regular blur

    olivettyApr 25, 2024

    @ttplanet Nice, where can we find the node?

    kiryanton930Apr 25, 2024· 1 reaction

    @olivetty I get a full prompt image in each tile. Always! SD upscale/Ultimate SD upscale/ControlNet Tile - useless nonsense.

    olivettyApr 26, 2024· 1 reaction

    @kiryanton930 That's not how this works, if you push the denoise to much it'll give you full prompt for each, it's better to just have detail and style tokens when doing upscales and keep the denoise between 0.2-0.4 depending on the scene, it should work! @ttplanet we would love that node you built!

    kiryanton930Apr 26, 2024

    @olivetty At 0.2-0.4 there is no effect. If more, then the effect appears in the form of additional faces, but this is not the effect that I expect

    MysticDaedraJun 13, 2024· 1 reaction

    @olivetty At denoise 0.25 it creates ghost images. Setting the denoise any lower entirely defeats the purpose of using a controlnet tile model when upscaling: to introduce further details as the resolution increases.

    ttplanet
    Author
    Jun 14, 2024

    @MysticDaedra controlnet can provide the possibility to set high denoise to adding detail or change the content. you are tight

    olivettyJun 17, 2024

    @MysticDaedra Weird, mine works perfectly, and trust me, I know what this is for :) I can go as high as 0.6-7 if I want and it does a good job still. Depends on my needs.

    greasebigApr 25, 2024· 4 reactions
    CivitAI

    i find the model can't follow canny and openpose. why?

    greasebigApr 25, 2024· 3 reactions
    CivitAI

    for "cloth change but keep the pose and person". did you use IPA?

    and while doing blurry recovery, i find it can't recover so well like you post above

    like this image : https://imgur.com/a/9emyUEj

    use your model and workflow, i can only get this result: https://imgur.com/a/srMRERF

    aiaicaptainApr 28, 2024

    that does not seem to work. your output image looks like an upscale only. i tried and got a better result tho not vey convincing. https://imgur.com/VOFVALG

    ttplanet
    Author
    May 6, 2024

    no, all the samples are without IPA

    ttplanet
    Author
    May 6, 2024

    for the picture you posted with not enough effective pixels can't be recovered.

    greasebigMay 8, 2024

    "no, all the samples are without IPA". I find it hard to reproduce your work on "cloth change but keep the pose and person". that means most of results i got using your workflow can't keep the pose and person

    ttplanet
    Author
    May 9, 2024

    @greasebig refer to the review pic below, you will see the test with different cloth, face, hair color, race.....

    tudi99wpmApr 28, 2024· 3 reactions
    CivitAI

    根本不会用,不知道如何安装到sd UI当中

    hjwnbshenMay 7, 2024

    这个只能在comfy上用吧

    destiny77021513May 8, 2024

    放進Controlnet照作者的截圖指示一樣跑就好了 不用預處理器 就那張"1.Open a A1111 webui." 下面那張截圖

    destiny77021513May 8, 2024

    @hjwnbshen 作者有給Web UI的指示 我也實際改提示詞跑一堆結果出來了 

    PATAPON07May 2, 2024· 3 reactions
    CivitAI

    will you publish your PVC models soon?

    ttplanet
    Author
    May 4, 2024

    not yet...:)

    chuckchen51May 5, 2024
    CivitAI

    It's amazing for portraits. Great work!

    One quick question: does it work with non-portrait images?

    ttplanet
    Author
    May 5, 2024· 1 reaction

    Yes

    KosiiqMay 15, 2024· 1 reaction

    I use it on Landscape images. It works great!

    AsensMay 6, 2024
    CivitAI

    不知道什么原因,无法运行,在使用这个contronet模型后,总是在采样器报错。是我2070 8g显卡带不动吗?

    ttplanet
    Author
    May 6, 2024

    看看你是不是中文路径了

    AsensMay 7, 2024

    @ttplanet 好的 我再试试 谢谢回复

    AsensMay 8, 2024· 1 reaction

    @ttplanet 我成功了,不是中文路径的问题,是内存不够的问题,我把虚拟内存搞到40g就可以运行了了,显存刚刚够,流程参考了你的,但有些不一样,你流程里的一些面板我的confyui加载不出来。

    ttplanet
    Author
    May 9, 2024· 1 reaction

    8G带不动的,因为sdxl大模型6.3+cn2.4

    AsensMay 10, 2024

    @ttplanet 我成功了,哈哈,能搞,webui开fp8的精度就行,我不知道confyui怎么生效的,但是也能跑成功,刚开始运行大模型的时候内存会要求很高,16g不够,把虚拟内存加上一共开到40g才不爆内存,后面KS采样器渲染用了6.8g+的显存,时间要一两分钟。

    ccrn9527May 10, 2024
    CivitAI

    很好,目前比较下来我感觉质量比sd1.5那个官方模型更好

    ccrn9527May 11, 2024
    CivitAI

    有个建议,能否给这个tile模型加两个滑块,1滑块是形状的权重,2滑块是颜色的权重(如果允许用户上传自定义色彩图那就更牛逼啦)

    ttplanet
    Author
    May 11, 2024

    webui有作者自己的colorfix你可以调用,comfyui直接用其他颜色匹配就行了

    ccrn9527May 23, 2024

    @ttplanet 哈哈,你说的虽然没错,但终归是不如集成在一个工具里那么好使的

    3dartalMay 15, 2024· 3 reactions
    CivitAI

    Is there any suggestions about How to use in Invokai?

    MysticDaedraMay 24, 2024
    CivitAI

    You say to use tile_resample... any clue where this can be downloaded?

    ttplanet
    Author
    May 24, 2024

    webui, tile_resample, for comfyui, I thought my preprocessor is combined in advanced contorlnet node

    eman23noJun 5, 2024
    CivitAI

    New to this. Would you kindly tell me which folder of webui to start with? Place it where after download? Xiexie

    cptdanJun 9, 2024· 2 reactions

    models/ControlNet