ComfyUI IC-Custom Node
September 3, 2025 ยท View on GitHub
A custom node for ComfyUI that integrates IC-Custom model for high-quality image customization and generation.
โจ Features
- ๐จ High-Quality Image Generation: Powered by FLUX.1-Fill-dev and IC-Custom models
- ๐ผ๏ธ Image Customization: Generate customized images based on reference images
- ๐ฏ Flexible Generation Modes: Support for position-free and position-precise generation
- โ๏ธ Advanced Controls: Configurable guidance scale, inference steps, and seed control
- ๐ Optimized Performance: Model quantization and offloading for better memory efficiency
๐ฆ Installation
Step 1: Install the Node
# Navigate to ComfyUI custom_nodes directory
cd ComfyUI/custom_nodes
# Clone the repository
git clone https://github.com/HM-RunningHub/ComfyUI_RH_ICCustom
# Install dependencies
cd ComfyUI_RH_ICCustom
pip install -r requirements.txt
Step 2: Download Required Models
Create the following directory structure in your ComfyUI models folder:
Main Models
FLUX.1-Fill-dev Model:
- Download: FLUX.1-Fill-dev
- Files:
ae.safetensors,flux1-fill-dev.safetensors - Location:
ComfyUI/models/black-forest-labs/FLUX.1-Fill-dev/
IC-Custom Model:
- Download: IC-Custom
- Files: All files from the repository
- Location:
ComfyUI/models/IC-Custom/
FLUX Redux Model:
- Download: FLUX.1-Redux-dev
- File:
flux1-redux-dev.safetensors - Location:
ComfyUI/models/IC-Custom/
CLIP Models
SigLIP Model:
- Download: siglip-so400m-patch14-384
- Files: All files from the repository
- Location:
ComfyUI/models/clip/siglip-so400m-patch14-384/
XFlux Text Encoders:
- Download: xflux_text_encoders
- Files: All files from the repository
- Location:
ComfyUI/models/clip/xflux_text_encoders/
CLIP Vision Models
CLIP ViT Large:
- Download: clip-vit-large-patch14
- Files: All files from the repository
- Location:
ComfyUI/models/clip_vision/clip-vit-large-patch14/
SigCLIP Vision:
- Download: sigclip_vision_patch14_384
- File:
sigclip_vision_patch14_384.safetensors - Location:
ComfyUI/models/clip_vision/
๐ Usage
Basic Workflow
- Add Model Loader: Add "RunningHub ICCustom Loader" node to your workflow
- Add Sampler: Add "RunningHub ICCustom Sampler" node and connect the pipeline output
- Configure Inputs:
- Connect reference image
- Set prompt text
- Configure generation parameters
- Optionally add target image and mask for precise control
Example Workflow
[Reference Image] โ [ICCustom Loader] โ [ICCustom Sampler] โ [Save Image]
โ
[Prompt Input]
Generation Modes
- Position-Free: Generate without target constraints (no mask required)
- Position-Precise: Generate with specific target positioning (requires mask)
โ๏ธ Parameters
- Prompt: Text description for the generated content
- Guidance: Controls adherence to prompt (default: 40.0)
- True GS: Additional guidance parameter (default: 3.0)
- Steps: Number of inference steps (default: 25)
- Seed: Random seed for reproducible results
๐ง Requirements
- GPU Memory: 16GB+ VRAM recommended
- System RAM: 32GB+ recommended
- Storage: ~100B for all models
- Dependencies: PyTorch, Diffusers, Transformers
๐ License
This project is licensed under the Apache 2.0 License.
๐ References
๐ Acknowledgments
Special thanks to AIwood็ฑๅฑ็ ็ฉถๅฎค (Bilibili) for helping with Windows environment testing and contributing to the installation documentation.
๐ค Contributing
Contributions are welcome! Please feel free to submit issues and pull requests.