๐Ÿง  EVOLVE : Event-Guided Deformable Feature Transfer and Dual-Memory Refinement for Low-Light Video Object Segmentation (ICCV 2025)

March 24, 2026 ยท View on GitHub

News

  • [New] We upload the LLE-VOS qualititative result on Google Drive Link
  • [New] We upload the LLE-DAVIS qualititative result on Google Drive Link

๐Ÿ“Œ Key Features

  • ๐ŸŽฏ Event-guided Deformable Feature Transfer Module

  • ๐Ÿ” Dual-Memory Object Transformer

  • ๐Ÿงฉ Memory Refinement Module


Data preparation & Installation

See Datasets & Installation

Training Command

We trained with four A6000 GPUs, which took around 10 hours on LLE-VOS.

OMP_NUM_THREADS=4 torchrun --master_port 25357 --nproc_per_node=4 evolve/train.py exp_id=[some unique id] model=base data=base
  • Change nproc_per_node to change the number of GPUs.
  • Prepend CUDA_VISIBLE_DEVICES=... if you want to use specific GPUs.
  • Change master_port if you encounter port collision.
  • exp_id is a unique experiment identifier that does not affect how the training is done.
  • Models and visualizations will be saved in ./output/.
  • To load a pre-trained model, e.g., to continue main training from the final model from pre-training, specify weights=[path to the model].

Evaluation Command

python cutie/eval_vos.py dataset=[dataset] weights=[path to model file] model=[small/base]
  • Possible options for dataset: see config/eval_config.yaml.
  • We evaulate our models on base model setting.

Qualititative Result

  • LLE-VOS Dataset

  • LLE-DAVIS Dataset

Citation

@InProceedings{Baek_2025_ICCV,
    author    = {Baek, Jong-Hyeon and Oh, Jiwon and Koh, Yeong Jun},
    title     = {EVOLVE: Event-Guided Deformable Feature Transfer and Dual-Memory Refinement for Low-Light Video Object Segmentation},
    booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
    month     = {October},
    year      = {2025},
    pages     = {11273-11282}
}

References

  • The GUI tools uses RITM for interactive image segmentation. This repository also contains a redistribution of their code in gui/ritm. That part of code follows RITM's license.

  • The interactive demo is developed upon IVS, MiVOS, and XMem.

  • We used framework from CUTIE.

  • Thanks to RTIM and XMem++ for making this possible.