4.5 Article

Generating Relevant and Informative Questions for Open-Domain Conversations

Journal

ACM TRANSACTIONS ON INFORMATION SYSTEMS
Volume 41, Issue 1, Pages -

Publisher

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3510612

Keywords

Conversational search; neural question generation; open-domain conversations; context modeling

Ask authors/readers for more resources

Recent research emphasizes the importance of mixed-initiative interactions in conversational search. The task of question generation (QG) in open-domain conversational systems aims to enhance human-machine interactions. However, the limited availability of QG-specific data in conversations makes this task challenging. In this study, we propose a context-enhanced neural question generation (CNQG) model that leverages conversational context to predict question content and pattern. We also use multi-task learning with auxiliary training objectives and a self-supervised approach to train our question generator.
Recent research has highlighted the importance of mixed-initiative interactions in conversational search. To enable mixed-initiative interactions, information retrieval systems should be able to ask diverse questions, such as information-seeking, clarification, and open-ended ones. Question generation (QG) of open-domain conversational systems aims at enhancing the interactiveness and persistence of human-machine interactions. The task is challenging because of the sparsity of QG-specific data in conversations. Current work is limited to single-turn interaction scenarios. We propose a context-enhanced neural question generation (CNQG) model that leverages the conversational context to predict question content and pattern, then perform question decoding. A hierarchical encoder framework is employed to obtain the discourse-level context representation. Based on this, we propose Review and Transit mechanisms to respectively select contextual keywords and predict new topic words to further construct the question content. Conversational context and the predicted question content are used to produce the question pattern, which in turn guides the question decoding process implemented by a recurrent decoder with a joint attention mechanism. To fully utilize the limited QG-specific data to train our question generator, we perform multi-task learning with three auxiliary training objectives, i.e., question pattern prediction, Review, and Transit mechanisms. The required additional labeled data is obtained in a self-supervised way. We also design a weight decaying strategy to adjust the influences of various auxiliary learning tasks. To the best of our acknowledge, we are the first to extend the application of QG to the multi-turn open-domain conversational scenario. Extensive experimental results demonstrate the effectiveness of our proposal and its main components on generating relevant and informative questions, with robust performance for contexts with various lengths.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available