About Figure.5 in your paper
Sebastian-X opened this issue · 4 comments
Hello!
I want to implement an Object Confusion experiment as your paper shows in Figure.5, but couldn't find an appropriate way. I tried to use the number of True Positives(TP) and False Positives(FP) to represent (classification √ localization √) and (classification √ localization ×) respectively, but not sure if it makes sense and the results are far beyond from yours. Could you please briefly share experiment details about Figure.5, like what kind of metrics do you use. Thanks!
Loc √ Cls √: correct label and IoU >= 0.5 with groundtruth box
Loc √ Cls ×: wrong label and IoU >= 0.5 with groundtruth box
Loc × Cls √: correct label and 0.3 <= IoU < 0.5 with groundtruth box
Loc × Cls ×: others
Normally, when we calculate AP of a category, only the detection boxes of this specific category are selected to compute IOU with groundtruth. However, your experiment contains 'correct label' and 'wrong label', does it mean all the detection boxes with different classification results are used for calculation?
Sure. We use all detection boxes for calculation.
Thanks for your help!