M3Track: Meta-Prompt for Multi-Modal Tracking

June 15, 2025 · View on GitHub

This work is accepted by IEEE Signal Processing Letter in 2025. Paper

6f7e9ff5939908ed2da185f9f8f4fb9

⭐ We carefully design a dedicated meta-prompt learning solution in the realm of multi-modal tracking, injecting sequence-specific evidence through online adaption.

⭐ We propose a unified multi-modal tracker without the prerequisite of feeding any task priors (notification of task type) in both training and test phases.

image

⭐ This work brings limited degradation on efficiency but consistent improvements

image

⭐ Visiualisation

image

🔋The main file is lib/train/actor/vipt.py (update trigger), lib/models/ThreeMT/ostrack_meta_ptompt.py (inner update), and lib/train/trainers/ltr_trainer.py (backward, outter update)

🔽 Please follow ViPT[https://github.com/jiawen-zhu/ViPT] to create your workspace (conda environment and download the pretrained OSTrack).