Official implementation of E.T.Track. E.T.Track utilized our proposed Exemplar Transformer, a transformer module utilizing a single instance level attention layer for realtime visual object tracking. E.T.Track is up to 8x faster than other transformer-based models, and consistently outperforms competing lightweight trackers that can operate in realtime on standard CPUs.
E.T.Track | The standard attention vs our Exemplar Attention on the right |
---|---|
Install the python environment using the environment file ettrack_env.yml
.
Generate the relevant files:
python -c "from pytracking.evaluation.environment import create_default_local_file; create_default_local_file()"
python -c "from ltr.admin.environment import create_default_local_file; create_default_local_file()"
- Modify local.py.
Modify the path files for the evaluation in
pytracking/evaluation/local.py
- Trained E.T.Track model for inference:
wget https://data.vision.ee.ethz.ch/kanakism/checkpoint_e35.pth -P ./checkpoints/et_tracker/
We evaluate our models using PyTracking.
- Add the correct dataset in
pytracking/experiments/myexperiments.py
(default: OTB-100) - Run
python3 -m pytracking.run_experiment myexperiments et_tracker --threads 0
If you use this code, please consider citing the following paper:
@article{blatter2021efficient,
title={Efficient Visual Tracking with Exemplar Transformers},
author={Blatter, Philippe and Kanakis, Menelaos and Danelljan, Martin and Van Gool, Luc},
journal={arXiv preprint arXiv:2112.09686},
year={2021}
}