/YCB_demo

A demo for understanding the structure and pre-processing of YCB-Video Dataset.

Primary LanguagePython

Demo for YCB-Video Dataset Usage

Adapted from PoseCNN-PyTorch

Running the demo

  • run the following script

    ./experiments/scripts/ycb_video_test.sh
  • on hpc

    sbatch hpc.sh

Information

This is the YCB-Video dataset [1] for 6D object pose estimation. It provides accurate 6D poses of 21 objects from the YCB dataset [2] observed in 92 videos with 133,827 frames.

This project was funded in part by Siemens and by NSF STTR grant 63-5197 with Lula Robotics.

How to Cite

@article{xiang2017posecnn, author = {Xiang, Yu and Schmidt, Tanner and Narayanan, Venkatraman and Fox, Dieter}, title = {PoseCNN: A Convolutional Neural Network for 6D Object Pose Estimation in Cluttered Scenes}, journal = {arXiv preprint arXiv:1711.00199}, year = {2017} }

Basic Usage

The description of the directories in this package:

  1. cameras. The camera parameters used to capture the videos. asus-uw.json for video 0000 ~ 0059, asus-cmu.json for video 0060 ~ 0091.
  2. data. The 92 videos in the dataset.
  3. data_syn. 80,000 synthetic images of the 21 YCB objects.
  4. image_sets. Separation of the videos into training set (train.txt) and the testing set (val.txt, keyframe.txt).
  5. keyframes. Keyframe indexes of the 12 testing videos.
  6. models. 3D models of the 21 YCB objects.
  7. pairs. Stereo pair indexes of the 12 testing videos.
  8. poses. All the 6D poses of the 21 YCB objects in the dataset (quaternion + translation).

References

[1] Y. Xiang, T. Schmidt, V. Narayanan and D. Fox. PoseCNN: A convolutional neural network for 6D object pose estimation in cluttered scenes. In arXiv:1711.00199, 2017.

[2] B. Calli, A. Singh, A. Walsman, S. Srinivasa, P. Abbeel, and A. M. Dollar. The YCB object and model set: Towards common benchmarks for manipulation research. In International Conference on Advanced Robotics (ICAR), pp. 510–517, 2015.