Skip to content

Latest commit

 

History

History
43 lines (25 loc) · 2.28 KB

README.md

File metadata and controls

43 lines (25 loc) · 2.28 KB

MM-pretrain

This code is for the paper: Emotion-aware Multimodal Pre-training for Image-grounded Emotional Response Generation. If you use this code or results from our paper, please cite:

@InProceedings{mm-pretrain,
author="Tian, Zhiliang and Wen, Zhihua and Wu, Zhenghao and Song, Yiping and Tang, Jintao and Li, Dongsheng and Zhang, Nevin L.",
editor="Bhattacharya, Arnab and Lee Mong Li, Janice and Agrawal, Divyakant and Reddy, P. Krishna and Mohania, Mukesh and Mondal, Anirban and Goyal, Vikram and Uday Kiran, Rage",
title="Emotion-Aware Multimodal Pre-training for Image-Grounded Emotional Response Generation",
booktitle="Database Systems for Advanced Applications",
year="2022",
publisher="Springer International Publishing",
isbn="978-3-031-00129-1"
}

This work is based on these projects:

The implementation is based on fairseq framework with pytorch. We delete the identification-related information in all the scripts.

./data directory contains scripts describing the formation of datasets.

./extract_features directory contains scripts regarding the collecting and pre-processing of large image datasets.

./model directory contains scripts defining the model structures been studied.

./resnet directory contains the scripts that model, train and predict of our image encoder

./sentiment_predictor directory contains the implementation of Bert based sentiment predictor for the C-I2T task.

./tasks directory contains scripts defining four pre-training tasks and the downstream task.

./scripts directory contains shell scripts related to the pre-training and finetuning of different tasks.