This repository contains the code for the paper VG-TAMP: Visually-Grounded Task-and-Motion Planning in 3D Household Environments.
This work is developed on the top of MMDetection3D toolbox, the model is based on Pointformer, and we edited it to apply on the ALFRED dataset in the paper.
- NVIDIA GPU + CUDA 11.2
- Python 3.8 (Recommend to use Anaconda)
- PyTorch == 1.10.0+cu113
- mmcv-full == 1.3.17
- mmdet == 2.18.1
- mmsegmentation == 0.19.0
- Install dependencies following their guidelines.
- Clone and install mmdet3d in develop mode.
git clone https://github.com/open-mmlab/mmdetection3d.git
cd mmdetection3d
python setup.py develop
- Update the files in this repo into the directories in mmdet3d.
Please refer to alfred for more detail
# Training
bash tools/dist_train.sh configs/pointformer/votenet_ptr_alfred-3d-class.py 8
# Testing
bash tools/dist_test.sh configs/pointformer/votenet_ptr_alfred-3d-class.py checkpoints/votenet_ptr_alfred-3d-class.pth 8 --eval mAP
This code is based on MMDetection3D and Pointformer