4.7 Article

Enhancing Transformer-based language models with commonsense representations for knowledge-driven machine comprehension

Journal

KNOWLEDGE-BASED SYSTEMS
Volume 220, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.knosys.2021.106936

Keywords

Machine Reading Comprehension; Transformer; Commonsense knowledge; Pretrained language model

Funding

  1. Open Project Foundation of Henan key Laboratory for Big Data Processing AMP
  2. Analytics of Electronic Commerce, China [2020-KF-10]

Ask authors/readers for more resources

The paper introduces three injection methods and a mask mechanism for fine-tuning knowledge-driven MRC tasks, achieving significant performance improvements and demonstrating the effectiveness and robustness of the incremental model.
Compared to the traditional machine reading comprehension (MRC) with limitation to the information in a passage, knowledge-driven MRC tasks aim to enable models to answer the question according to text and related commonsense knowledge. Although pre-trained Transformer-based language models (TrLMs) such as BERT and Roberta, have shown powerful performance in MRC, external knowledge such as unspoken commonsense and world knowledge still cannot be used and explained explicitly. In this work, we present three simple yet effective injection methods integrated into the structure of TrLMs to fine-tune downstream knowledge-driven MRC tasks with off-the-shelf commonsense representations. Moreover, we introduce a mask mechanism for a token-level multi-hop relationship searching to filter external knowledge. We have conducted extensive experiments on DREAM and CosmosQA, two prevalent knowledge-driven datasets. Experimental results indicate that the incremental TrLMs have outperformed the baseline systems by 1%-4.1% with a fewer computational cost. Further analysis shows the effectiveness of the proposed methods and the robustness of the incremental model in the case of an incomplete training set. (c) 2021 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available