This code has been tested on Ubuntu 16.04, Python 3.6, Pytorch 0.4.1/1.2.0, CUDA 9.0. Please install related libraries before running this code:
pip install -r requirements.txt| Dataset | SiamCAR | |
| OTB100 | Success | 70.0 |
| Precision | 91.4 | |
| UAV123 | Success | 64.0 |
| Precision | 83.9 | |
| LaSOT | Success | 51.6 |
| Norm precision | 61.0 | |
| Precision | 52.4 | |
| GOT10k | AO | 58.1 |
| SR0.5 | 68.3 | |
| SR0.75 | 44.1 | |
| VOT2018 | EAO | 42.3 |
| Robustness | 19.7 | |
| Accuracy | 57.4 | |
| VOT2020 | EAO | 27.3 |
| Robustness | 73.2 | |
| Accuracy | 44.9 | |
| TrackingNet | Success | 74.0 |
| Norm precision | 80.4 | |
| Precision | 68.4 | |
Download the pretrained model:
general_model code: lw7w
got10k_model code: p4zx
LaSOT_model code: 6wer
(The model in google Driver)
and put them into tools/snapshot directory.
Download testing datasets and put them into test_dataset directory. Jsons of commonly used datasets can be downloaded from BaiduYun or Google driver. If you want to test the tracker on a new dataset, please refer to pysot-toolkit to set test_dataset.
python test.py \
--dataset UAV123 \ # dataset_name
--snapshot snapshot/general_model.pth # tracker_nameThe testing result will be saved in the results/dataset_name/tracker_name directory.
Download the datasets:
Note: train_dataset/dataset_name/readme.md has listed detailed operations about how to generate training datasets.
Download pretrained backbones from google driver or BaiduYun (code: 7n7d) and put them into pretrained_models directory.
To train the SiamCAR model, run train.py with the desired configs:
cd tools
python train.pyWe provide the tracking results (code: 4er6) (results in google driver )of GOT10K, LaSOT, OTB, UAV, VOT2018 and TrackingNet. If you want to evaluate the tracker, please put those results into results directory.
python eval.py \
--tracker_path ./results \ # result path
--dataset UAV123 \ # dataset_name
--tracker_prefix 'general_model' # tracker_name
The code is implemented based on pysot. We would like to express our sincere thanks to the contributors.
If you use SiamCAR in your work please cite our papers:
@article{cui2022joint,
title={Joint Classification and Regression for Visual Tracking with Fully Convolutional Siamese Networks},
author={Cui, Ying and Guo, Dongyan and Shao, Yanyan and Wang, Zhenhua and Shen, Chunhua and Zhang, Liyan and Chen, Shengyong},
journal={International Journal of Computer Vision},
year={2022},
publisher={Springer},
doi = {10.1007/s11263-021-01559-4}
}
@InProceedings{Guo_2020_CVPR,
author = {Guo, Dongyan and Wang, Jun and Cui, Ying and Wang, Zhenhua and Chen, Shengyong},
title = {SiamCAR: Siamese Fully Convolutional Classification and Regression for Visual Tracking},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2020}
}
@InProceedings{Guo_2021_CVPR,
author = {Guo, Dongyan and Shao, Yanyan and Cui, Ying and Wang, Zhenhua and Zhang, Liyan and Shen, Chunhua},
title = {Graph Attention Tracking},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2021}
}