4.7 Article

Context-Aware Attentive Multilevel Feature Fusion for Named Entity Recognition

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2022.3178522

关键词

Attention mechanism; multilevel feature extraction; named entity recognition (NER); sequence labeling

资金

  1. National Natural Science Foundation of China [61976102, U19A2065, 61902145]
  2. National Science Foundation (NSF) [IIS-1763365, IIS-2106972]
  3. Hong Kong Baptist University (HKBU) One-Off Tier 2 Start-Up Grant [RCOFSGT2/20-21/SCI/004]
  4. University of California at Davis

向作者/读者索取更多资源

Named entity recognition (NER) is fundamental to information extraction and has attracted widespread attention in the field of natural language processing. Existing methods for NER often fail to integrate semantic and syntactic information and only consider partial features. In this study, a novel multilevel feature fusion model is proposed to capture features from various perspectives and enhance representation learning.
In the era of information explosion, named entity recognition (NER) has attracted widespread attention in the field of natural language processing, as it is fundamental to information extraction. Recently, methods of NER based on representation learning, e.g., character embedding and word embedding, have demonstrated promising recognition results. However, existing models only consider partial features derived from words or characters while failing to integrate semantic and syntactic information, e.g., capitalization, inter-word relations, keywords, and lexical phrases, from multilevel perspectives. Intuitively, multilevel features can be helpful when recognizing named entities from complex sentences. In this study, we propose a novel attentive multilevel feature fusion (AMFF) model for NER, which captures the multilevel features in the current context from various perspectives. It consists of four components to, respectively, capture the local character-level (CL), global character-level (CG), local word-level (WL), and global word-level (WG) features in the current context. In addition, we further define document-level features crafted from other sentences to enhance the representation learning of the current context. To this end, we introduce a novel context-aware attentive multilevel feature fusion (CAMFF) model based on AMFF, to fully leverage document-level features from all the previous inputs. The obtained multilevel features are then fused and fed into a bidirectional long short-term memory (BiLSTM)-conditional random field (CRF) network for the final sequence labeling. Extensive experiments on four benchmark datasets demonstrate that our proposed AMFF and CAMFF models outperform a set of state-of-the-art baseline methods and the features learned from multiple levels are complementary.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据