📄 Paper | 📦 Models & Results: Google Drive / Baidu Drive / Hugging Face
- [May 1, 2026] The models, raw results, and training logs are now available on Hugging Face 🤗.
- [Apr 13, 2026] Code, models, and raw results are released.
- 🌊 A simple unified multimodal tracking framework for RGB-T, RGB-D, and RGB-E tasks.
- 🚀 Achieves strong performance across multiple multimodal benchmarks.
- ⚡ Highly efficient: only 0.6M trainable parameters and 63.5 FPS.
- 🔍 Highlights the importance of cross-modal alignment in multimodal tracking.
conda env create -f environment.yaml
conda activate seatrack- LasHeR
- RGBT234 (qvsq)
- DepthTrack
- VOT22-RGBD
- VisEvent
Organize datasets as follows:
-- <DATA_PATH>
-- DepthTrack/trainingset
|-- adapter02_indoor
|-- bag03_indoor
|-- bag04_indoor
...
-- LasHeR/trainingset
|-- 1boygo
|-- 1handsth
...
-- VisEvent/trainingset
|-- 00142_tank_outdoor2
|-- 00143_tank_outdoor2
...
cd <PATH_TO_SEATRACK>
python tracking/create_default_local_file.py \
--workspace_dir . \
--data_dir <DATA_PATH> \
--save_dir ./outputOr manually modify:
./lib/train/admin/local.py # paths for training
./lib/test/evaluation/local.py # paths for testingDownload pretrained OSTrack and place it under:
./pretrained/vitb_256_mae_32x4_ep300
./pretrained/vitb_256_mae_ce_32x4_ep300
Then run:
bash train.shModify checkpoint in:
./lib/test/parameter/seatrack.py
Place datasets and the provided list.txt into:
./Depthtrack_workspace/sequences
./VOT22RGBD_workspace/sequences
Modify paths in
./Depthtrack_workspace/trackers.ini
./VOT22RGBD_workspace/trackers.ini
Run evaluation with VOT Toolkit:
bash eval_rgbd.shModify <DATASET_PATH> and <SAVE_PATH> in:
./RGBT_workspace/test_rgbt_mgpus.py
Then run:
bash eval_rgbt.shEvaluation tools:
Modify <DATASET_PATH> and <SAVE_PATH> in:
./RGBE_workspace/test_rgbe_mgpus.py
Run:
bash eval_rgbe.shEvaluation:
If you find this work helpful, please consider citing:
@misc{su2026seatracksimpleefficientadaptive,
title={SEATrack: Simple, Efficient, and Adaptive Multimodal Tracker},
author={Junbin Su and Ziteng Xue and Shihui Zhang and Kun Chen and Weiming Hu and Zhipeng Zhang},
year={2026},
eprint={2604.12502},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2604.12502},
}SEATrack uses code from a few open source repositories. Without the efforts of these folks (and their willingness to release their implementations), SEATrack would not be possible. We thank these authors for their efforts!
If you have any questions, feel free to contact:




