README.md

February 23, 2025 · View on GitHub

Control Color: Multimodal Diffusion-based Interactive Image Colorization

S-Lab, Nanyang Technological University 

Control Color (CtrlColor) achieves highly controllable multimodal image colorization based on stable diffusion model.

Region colorization Iterative editing

:open_book: For more visual results and applications of CtrlColor, go checkout our project page.


:mega: Updates

  • 2024.12.16: The test codes (gradio demo), colorization model checkpoint, and autoencoder checkpoint are now publicly available.

:desktop_computer: Requirements

  • required packages in CtrlColor_environ.yaml
# git clone this repository
git clone https://github.com/ZhexinLiang/Control-Color.git
cd Control_Color

# create new anaconda env and install python dependencies
conda env create -f CtrlColor_environ.yaml
conda activate CtrlColor

:running_woman: Inference

Prepare models:

Please download the checkpoints of both colorization model and vae from [Google Drive] and put both checkpoints in ./pretrained_models folder.

Testing:

You can use the following cmd to run gradio demo:

python test.py

Then you will get our interactive interface as below:

:love_you_gesture: Citation

If you find our work useful for your research, please consider citing the paper:

@article{liang2024control,
  title={Control Color: Multimodal Diffusion-based Interactive Image Colorization},
  author={Liang, Zhexin and Li, Zhaochen and Zhou, Shangchen and Li, Chongyi and Loy, Chen Change},
  journal={arXiv preprint arXiv:2402.10855},
  year={2024}
}

Contact

If you have any questions, please feel free to reach out at zhexinliang@gmail.com.