/DIVOTrack

Primary LanguagePythonOtherNOASSERTION

DIVOTrack: A Novel Dataset and Baseline Method for Cross-View Multi-Object Tracking in DIVerse Open Scenes

This repository contains the details of dataset and the Pytorch implementation of Baseline Method CrossMOT of the Paper: DIVOTrack: A Novel Dataset and Baseline Method for Cross-View Multi-Object Tracking in DIVerse Open Scenes

Abstract

Cross-view multi-object tracking aims to link objects between frames and camera views with substantial overlaps. Although cross-view multi-object tracking has received increased attention in recent years, existing datasets still have several issues, including 1) missing real-world scenarios, 2) lacking diverse scenes, 3) owning a limited number of tracks, 4) comprising only static cameras, and 5) lacking standard benchmarks, which hinder the investigation and comparison of cross-view tracking methods. To solve the aforementioned issues, we introduce DIVOTrack: a new cross-view multi-object tracking dataset for DIVerse Open scenes with dense tracking pedestrians in realistic and non-experimental environments. Our DIVOTrack has ten distinct scenarios and 550 cross-view tracks, surpassing all cross-view multi-object tracking datasets currently available. Furthermore, we provide a novel baseline cross-view tracking method with a unified joint detection and cross-view tracking framework named CrossMOT, which learns object detection, single-view association, and cross-view matching with an all-in-one embedding model. Finally, we present a summary of current methodologies and a set of standard benchmarks with our DIVOTrack to provide a fair comparison and conduct a comprehensive analysis of current approaches and our proposed CrossMOT.

  • Dataset Description
    • Dataset Structure
    • Dataset Downloads
  • Training Detector
  • Single-view Tracking
  • Cross-view Tracking
  • TrackEval
  • Multi_view_Tracking
  • MOTChallengeEvalKit_cv_test

The test result of the cross-view MOT baseline method MvMHAT on the DIVOTrack. test.gif

The ground truth of the DIVOTrak. gt.gif

Dataset Description

We collect data in 10 different real-world scenarios, named: 'Circle', 'Shop', 'Moving', 'Park', 'Ground', 'Gate1', 'Floor', 'Side', 'Square', 'Gate2'. All the sequences are captured by using 3 moving cameras: 'View1', 'View2', 'View3' and are manually synchronized.

In old version, the corresponding scenarios named: 'circleRegion', 'innerShop', 'movingView', 'park', 'playground', 'shopFrontGate', 'shopSecondFloor', 'shopSideGate', 'shopSideSquare', 'southGate'. The corresponding camera named: 'Drone', 'View1', 'View2'.

Dataset Structure

The structure of our dataset as:

DIVOTrack
    └——————datasets
             └——————DIVO
                |——————images
                |        └——————annotations
                |        └——————dets    
                |        └——————train
                |        └——————test
                └——————labels_with_ids
                |        └——————train
                |        └——————test  
                |——————ReID_format
                |        └——————bounding_box_test
                |        └——————bounding_box_train
                |        └——————query        
                └——————boxes.json  

Dataset Downloads

The whole dataset can download from GoogleDrive. Note that, each file needs to unzip by the password. You can decompress each .zip file in its folder after send us (shengyuhao@zju.edu.cn, gaoangwang@intl.zju.edu.cn) the License in any format. After that, you should run generate_ini.py to generate seqinfo.ini file.

Training Detector

The training process of our detector is in ./Training_detector/ and the details can see from Training_detector/Readme.md.

Single-view Tracking

The implement of single-view tracking baseline methods is in ./Single_view_Tracking and the details can see from Single_view_Tracking/Readme.md.

Cross-view Tracking

The implement of cross-view tracking baseline methods is in ./Cross_view_Tracking and the details can see from Cross_view_Tracking/Readme.md.

TrackEval

We evaluation each single-view tracking baseline by ./TrackEval, and the details can see from TrackEval/Readme.md.

Multi-view Tracking

The multi-view tracking results can get from ./Multi_view_Tracking, the details can see from Multi_view_Tracking/Readme.md

MOTChallengeEvalKit_cv_test

The cross-view evaluation can get from ./MOTChallengeEvalKit_cv_test, and the details can see from ./MOTChallengeEvalKit_cv_test/Readme.md

CrossMOT: A Novel Baseline Cross-View Tracking Method with A Unified Joint Detection and Cross-view Tracking Framework

The details of implementation of CrossMOT can see from CrossMOT/Readme.md

Reference

Any use whatsoever of this dataset and its associated software shall constitute your acceptance of the terms of this agreement. By using the dataset and its associated software, you agree to cite the papers of the authors, in any of your publications by you and your collaborators that make any use of the dataset, in the following format:

@article{wangdivotrack,
  title={DIVOTrack: A Novel Dataset and Baseline Method for Cross-View Multi-Object Tracking in DIVerse Open Scenes},
  author={Shenghao Hao, Peiyuan Liu, Yibing Zhan, Kaixun Jin, Zuozhu Liu, Mingli Song, Jenq-Neng Hwang, Gaoang Wang},
  journal={arXiv preprint arXiv:2302.07676},
  year={2023}
}

The license agreement for data usage implies the citation of the paper above. Please notice that citing the dataset URL instead of the publications would not be compliant with this license agreement. You can read the LICENSE from LICENSE.

Contact

If any concerns please contact shengyuhao@zju.edu.cn