期刊
出版社
ASSOC COMPUTING MACHINERY
DOI: 10.1145/3485042
关键词
Cross-modal image retrieval; re-ranking; question answering
类别
资金
- JSPS KAKENHI [JP17H01744]
In this study, a novel interactive cross-modal image-retrieval method based on question answering is proposed. The method analyzes candidate images and asks users questions to obtain information, leading to the retrieval of desired images even with ambiguous query texts.
Cross-modal image-retrieval methods retrieve desired images from a query text by learning relationships between texts and images. Such a retrieval approach is one of the most effective ways of achieving the easiness of query preparation. Recent cross-modal image-retrieval methods are convenient and accurate when users input a query text that can be used to uniquely identify the desired image. However, in reality, users frequently input ambiguous query texts, and these ambiguous queries make it difficult to obtain desired images. To overcome these difficulties, in this study, we propose a novel interactive cross-modal image-retrieval method based on question answering. The proposed method analyzes candidate images and asks users questions to obtain information that can narrow down retrieval candidates. By only answering questions generated by the proposed method, users can reach their desired images, even when using an ambiguous query text. Experimental results show the proposed method's effectiveness.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据