Skip to content

Latest commit

 

History

History
70 lines (52 loc) · 2.63 KB

README.md

File metadata and controls

70 lines (52 loc) · 2.63 KB

Modeling Explicit Concerning States for Reinforcement Learning in Visual Dialogue

Pytorch Implementation of the paper:

Modeling Explicit Concerning States for Reinforcement Learning in Visual Dialogue.
Zipeng Xu, Fandong Meng, Xiaojie Wang, Duo Zheng, Chenxu Lv and Jie Zhou. In Proccedings of BMVC 2021.

(The Appendix is included in our arXiv version: https://arxiv.org/pdf/2107.05250.pdf.)

This code is adapted from vmurahari3/visdial-diversity, we thank for their open sourcing.

Download data

Download preprocessed dialog data for VisDial v1.0:

sh scripts/download_preprocessed.sh

Download extracted features:

We use bottom-up image features with 10-100 proposals for each image. We use the features provided by Gi-Cheon Kang et al.. We thanks for their release.

Please download the files and put them under data/image_features.

Training

For Supervised Learning pre-training:

SL: Q-Bot

python train_sl.py -useGPU -trainMode sl-qbot -saveName SL_QBot 

SL: A-Bot

python train_sl.py -useGPU -trainMode sl-abot -a_learningRate 4e-4 -lrDecayRate 0.75 -saveName SL_ABot 

For Reinforcement Learning fine-tuning with ECS-based rewards:

python train_rl.py -dropout 0 -useGPU -useNDCG -trainMode rl-full-QAf -startFrom checkpoints/SL_ABOT.vd -qstartFrom checkpoints/SL_QBOT.vd -saveName RL-ECS

Pre-trained checkpoints

Will be released this week.

Reference

@inproceedings{xu2021ecsvisdial,
author = {Xu, Zipeng and Meng, Fandong and Wang, Xiaojie and Zheng, Duo and Lv, Chenxu and Zhou, Jie},
title = {modeling Explicit Concerning States for Reinforcement Learning in Visual Dialogue},
booktitle = {Proceedings of the 32nd British Machine Vision Conference (BMVC)},
year = {2021}
}