IMPROVING CROSS-MODAL UNDERSTANDING IN VISUAL DIALOG VIA CONTRASTIVE LEARNING
Feilong Chen1,2; Duzhen Zhang2; Xiuyi Chen2; Jing Shi2; Shuang Xu2; Bo Xu2
2022
会议日期2022.5
会议地点Singapore
英文摘要

Visual Dialog is a challenging vision-language task since the visual dialog agent needs to answer a series of questions after reasoning over both the image content and dialog history. Though existing methods try to deal with the cross-modal understanding in visual dialog, they are still not enough in ranking candidate answers based on their understanding of visual and textual contexts. In this paper, we analyze the cross-modal understanding in visual dialog based on the vision-language pre-training model VD-BERT and propose a novel approach to improve the cross-modal understanding for visual dialog, named ICMU. ICMU enhances cross-modal understanding by distinguishing different pulled inputs (i.e. pulled images, questions or answers) based on four-way contrastive learning. In addition, ICMU exploits the single-turn visual question answering to enhance the visual dialog model's cross-modal understanding to handle a multi-turn visually-grounded conversation. Experiments show that the proposed approach improves the visual dialog model's cross-modal understanding and brings satisfactory gain to the VisDial dataset.

内容类型会议论文
源URL[http://ir.ia.ac.cn/handle/173211/51916]  
专题数字内容技术与服务研究中心_听觉模型与认知计算
通讯作者Xiuyi Chen
作者单位1.School of Future Technology University of Chinese Academy of Sciences, Beijing, China
2.Institute of Automation Chinese Academy of Sciences (CASIA
推荐引用方式
GB/T 7714
Feilong Chen,Duzhen Zhang,Xiuyi Chen,et al. IMPROVING CROSS-MODAL UNDERSTANDING IN VISUAL DIALOG VIA CONTRASTIVE LEARNING[C]. 见:. Singapore. 2022.5.
个性服务
查看访问统计
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


©版权所有 ©2017 CSpace - Powered by CSpace