XTrack

July 11, 2025 · View on GitHub

Official implement of XTrack(XTrack: Multimodal Training Boosts RGB-X Video Object Trackers).

:star: :star: :star: ICCV 2025 :star: :star: :star:

Paper: [Preprint].

Data: 🤗 VOT-RGBD2022

Model Weight: 🤗XTrack-Base&OSTrack&DropMAE

Raw Result:Google Drive

Mixture of Modal Experts

meme_pipeline

SOTA Comparison

MethodDepthTrack (F-score↑)DepthTrack (Re↑)DepthTrack (Pr↑)VOT-RGBD2022 (EAO↑)VOT-RGBD2022 (Acc.↑)VOT-RGBD2022 (Rob.↑)LasHeR (Pr↑)LasHeR (Sr↑)RGBT234 (MPR↑)RGBT234 (MSR↑)VisEvent (Pr↑)VisEvent (Sr↑)
XTrack-L64.864.365.474.082.888.973.158.787.865.480.563.3
XTrack-B61.862.061.574.082.188.869.155.787.464.977.560.9
UnTrack61.061.061.072.182.086.964.651.384.262.575.558.9
SDSTrack61.960.961.472.881.288.366.553.184.862.576.759.7
OneTracker60.960.460.772.781.987.267.253.885.764.276.760.8
ViPT59.459.659.272.181.587.165.152.583.561.775.859.2
ProTrack57.857.358.365.180.180.253.842.079.559.963.247.1

Usage

Installation

Create and activate a conda environment:

conda create -n xtrack python=3.7
conda activate xtrack

Install the required packages:

bash install_xtrack.sh

Data Preparation

Download the training datasets, It should look like:

$<PATH_of_Datasets>
    -- LasHeR/TrainingSet
        |-- 1boygo
        |-- 1handsth
        ...
    -- VisEvent/train
        |-- 00142_tank_outdoor2
        |-- 00143_tank_outdoor2
        ...
        |-- trainlist.txt

Path Setting

Run the following command to set paths:

cd <PATH_of_XTrack>
python tracking/create_default_local_file.py --workspace_dir . --data_dir <PATH_of_Datasets> --save_dir ./output

You can also modify paths by these two files:

./lib/train/admin/local.py  # paths for training
./lib/test/evaluation/local.py  # paths for testing

Training

Download the pretrained foundation model as posted above. and put it under ./pretrained/.

bash train_xtrack.sh

You can train models with various modalities and variants by modifying train_xtrack.sh.

Testing

For RGB-T benchmarks

[LasHeR & RGBT234]
Modify the <DATASET_PATH> and <SAVE_PATH> in./RGBT_workspace/test_rgbt_mgpus.py, then run:

bash eval_rgbt.sh

We refer you to use LasHeR Toolkit for LasHeR evaluation, and refer you to use MPR_MSR_Evaluation for RGBT234 evaluation.

For RGB-E benchmark

[VisEvent]
Modify the <DATASET_PATH> and <SAVE_PATH> in./RGBE_workspace/test_rgbe_mgpus.py, then run:

bash eval_rgbe.sh

We refer you to use VisEvent_SOT_Benchmark for evaluation.

Citation

Please cite our work if you think it is useful for your research.



@inproceedings{tan2024xtrack,
  title={XTrack: Multimodal Training Boosts RGB-X Video Object Trackers},
  author={Tan, Yuedong and Wu, Zongwei and Fu, Yuqian and Zhou, Zhuyun and Sun, Guolei and Ma, Chao and Paudel, Danda Pani and Van Gool, Luc and Timofte, Radu},
  booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
  year={2025}
}

Acknowledgment