This is the pytorch implementation of the paper "Parallel Attention: A Unified Framework for Visual Object Discovery through Dialogs and Queries"
If you use this code in your research, please cite our paper:
@inproceedings{zhuang2018parallel,
title={Parallel attention: A unified framework for visual object discovery through dialogs and queries},
author={Zhuang, Bohan and Wu, Qi and Shen, Chunhua and Reid, Ian and van den Hengel, Anton},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages={4252--4261},
year={2018}
}
Please download the MSCOCO dataset and the GuessWhat?! annotations:
-
MSCOCO dataset http://cocodataset.org/#home
-
GuessWhat?! dataset https://github.com/GuessWhatGame/guesswhat
utils.py: provide necessary functions
main.py: main file, implementing training and testing
read_data.py: self-defined data layer
config.yaml: define the necessary hyperparameters (e.g., data directory, GPU), please modify this file
model.py: define the whole framework
./modules/: defines the attention module and the LSTM module
./data/: provide necessary data used in our experiment
python main.py
Copyright (c) Bohan Zhuang. 2017
** This code is for non-commercial purposes only.
This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version.
This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
You should have received a copy of the GNU General Public License along with this program. If not, see http://www.gnu.org/licenses/.