Unable to replicate model zoo mAP
kjxlstad opened this issue · 0 comments
kjxlstad commented
I've been wanting to try out the RepVGG-A0 backed legacy model. However, whenever I download the weights / checkpoints from the model zoo and use the legacy_v0.x_configs/RepVGG/nanodet-RepVGG-A0_416.yml
I only get 16.3 COCO mAP on the validation data while the benchmark table claims a COCO mAP of 27.8.
I have tried cloning earlier versions of the repository and using older checkpoints to no avail, running tools/test.py --task val
I always seem to get something close to this:
Average Precision (AP) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.163
Average Precision (AP) @[ IoU=0.50 | area= all | maxDets=100 ] = 0.251
Average Precision (AP) @[ IoU=0.75 | area= all | maxDets=100 ] = 0.171
Average Precision (AP) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.085
Average Precision (AP) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.184
Average Precision (AP) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.249
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 1 ] = 0.256
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 10 ] = 0.410
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.432
Average Recall (AR) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.172
Average Recall (AR) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.504
Average Recall (AR) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.643
How can I reproduce your results?