MCG-NJU/MixFormerV2

Search factor for MixFormerV2 during training and testing

Kou-99 opened this issue · 4 comments

Kou-99 commented

Thanks for your great work! I notice that the training search factor in yaml file for MixFormerV2-S is 4.5 while the testing search factor is 4.4 on LaSOT and 4.55 for Got10k and TrackingNet. Is there a reason for such inconsistency and will the mismatched factor affect the performance? Additionally, the center jitter for MixFormerV2-S is 4.5 (same as the search factor) will this cause more partially visible targets during training and reduce performance, considering MixFormerV2-S uses a smaller center jitter (4.5) and a larger search factor (5)?

The search factor is a hyper parameter in both training and testing, well as center jitter. We choose the training hyper parameter according to the teacher model setting. The search factors in training and testing are not necessarily to be the same, even not necessarily to be fixed. Search factor only determines how to crop the test frame according to the result box. It may blend the performance slightly but is usually not the key point to the performance level. For specific videos, it can be tuned for a suitable one. It is also even possible to design a smart strategy for search factor during inference, such as dynamic adaptive method.

Kou-99 commented

The search factor is a hyper parameter in both training and testing, well as center jitter. We choose the training hyper parameter according to the teacher model setting. The search factors in training and testing are not necessarily to be the same, even not necessarily to be fixed. Search factor only determines how to crop the test frame according to the result box. It may blend the performance slightly but is usually not the key point to the performance level. For specific videos, it can be tuned for a suitable one. It is also even possible to design a smart strategy for search factor during inference, such as dynamic adaptive method.

Thanks for your quick reply! To make sure I get everything right, I list my understanding below:

  1. The student share the same set of hyper-parameter with the teacher
  2. During training, for Base: template/ search size: 128/ 288, template/ search factor: 2/ 5
  3. During training, for Small: template/ search size: 112/ 224, template/ search factor: 2/ 5
  4. During testing, search factor are carefully tuned for each dataset whereas the search size is fixed to the size used in training

Please correct me if anything is wrong, thanks!

Yes. Except for the Small, search factor is 4.5 during training.

Kou-99 commented

Get it. Thanks for your reply! Good luck with your submission!