CivArchive
    Preview 22439926

    📖 Introduction

    Kolors is a large-scale text-to-image generation model based on latent diffusion, developed by the Kuaishou Kolors team. Trained on billions of text-image pairs, Kolors exhibits significant advantages over both open-source and proprietary models in visual quality, complex semantic accuracy, and text rendering for both Chinese and English characters. Furthermore, Kolors supports both Chinese and English inputs, demonstrating strong performance in understanding and generating Chinese-specific content. For more details, please refer to this technical report.

    🚀 Quick Start

    Requirements

    • Python 3.8 or later

    • PyTorch 1.13.1 or later

    • Transformers 4.26.1 or later

    • Recommended: CUDA 11.7 or later

    1. Repository cloning and dependency installation

    apt-get install git-lfs
    git clone https://github.com/Kwai-Kolors/Kolors
    cd Kolors
    conda create --name kolors python=3.8
    conda activate kolors
    pip install -r requirements.txt
    python3 setup.py install
    
    1. Weights download(link):

    huggingface-cli download --resume-download Kwai-Kolors/Kolors --local-dir weights/Kolors
    

    or

    git lfs clone https://huggingface.co/Kwai-Kolors/Kolors weights/Kolors
    
    1. Inference:

    python3 scripts/sample.py "一张瓢虫的照片,微距,变焦,高质量,电影,拿着一个牌子,写着“可图”"
    # The image will be saved to "scripts/outputs/sample_test.jpg"
    

    Using with Diffusers

    Please refer to https://huggingface.co/Kwai-Kolors/Kolors-diffusers.

    📜 License&Citation

    License

    Kolors are fully open-sourced for academic research. For commercial use, please fill out this questionnaire and sent it to [email protected] for registration.

    We open-source Kolors to promote the development of large text-to-image models in collaboration with the open-source community. The code of this project is open-sourced under the Apache-2.0 license. We sincerely urge all developers and users to strictly adhere to the open-source license, avoiding the use of the open-source model, code, and its derivatives for any purposes that may harm the country and society or for any services not evaluated and registered for safety. Note that despite our best efforts to ensure the compliance, accuracy, and safety of the data during training, due to the diversity and combinability of generated content and the probabilistic randomness affecting the model, we cannot guarantee the accuracy and safety of the output content, and the model is susceptible to misleading. This project does not assume any legal responsibility for any data security issues, public opinion risks, or risks and liabilities arising from the model being misled, abused, misused, or improperly utilized due to the use of the open-source model and code.

    Citation

    If you find our work helpful, please cite it!

    @article{kolors,
      title={Kolors: Effective Training of Diffusion Model for Photorealistic Text-to-Image Synthesis},
      author={Kolors Team},
      journal={arXiv preprint},
      year={2024}
    }
    

    Acknowledgments

    • Thanks to Diffusers for providing the codebase.

    • Thanks to ChatGLM3 for providing the powerful Chinese language model.

    Contact Us

    If you want to leave a message for our R&D team and product team, feel free to join our WeChat group. You can also contact us via email ([email protected]).

    Description

    VAE
    Kolors

    Details

    Downloads
    78
    Platform
    CivitAI
    Platform Status
    Available
    Created
    8/2/2024
    Updated
    9/27/2025
    Deleted
    -

    Files

    kolorsVAE_fp32.bin

    Mirrors

    Huggingface (64 mirrors)
    CivitAI (1 mirrors)