3.8 Proceedings Paper

Keep Meeting Summaries on Topic: Abstractive Multi-Modal Meeting Summarization

出版社

ASSOC COMPUTATIONAL LINGUISTICS-ACL

关键词

-

资金

  1. U.S. National Science Foundation [IIP-1631674]
  2. DARPA AIDA Program [FA8750-18-2-0014]
  3. ARL NS-CTA [W911NF-09-2-0053]
  4. Tencent AI Lab Rhino-Bird Gift Fund

向作者/读者索取更多资源

Transcripts of natural, multi-person meetings differ significantly from documents like news articles, which can make Natural Language Generation models generate unfocused summaries. We develop an abstractive meeting summarizer from both videos and audios of meeting recordings. Specifically, we propose a multi-modal hierarchical attention mechanism across three levels: topic segment, utterance and word. To narrow down the focus into topically-relevant segments, we jointly model topic segmentation and summarization. In addition to traditional textual features, we introduce new multi-modal features derived from visual focus of attention, based on the assumption that an utterance is more important if its speaker receives more attention. Experiments show that our model significantly outperforms the state-of-the-art with both BLEU and ROUGE measures.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据