Datasets for semi-supervised video object segmentation
z-jiaming opened this issue · 6 comments
Could you please tell me how to convert json to mask in DAVIS-format, like first_frame_annotations.json?
And how to convert the mask prediction to json for evaluating?
If possible, it would greatly facilitate BURST's research in VOS.
Thanks a lot!!!
Hi, you can inspect the format by opening up the JSON file in Firefox. It will show you how the lists and dictionaries are organized. The format is also documented in ANNOTATION_FORMAT.md. To see how you can parse the RLE-encoded mask into an image see this method: https://github.com/Ali2500/BURST-benchmark/blob/main/burstapi/utils.py#L21
The test set contains some videos which do not have any ground-truth annotations. Consequently, these videos are not present in the first_frame_annotations.json
file.
So, if my predictions.json dosen't contain these removed video, the eval result is also correct?
Yes it should be.
Got it. Thanks!