Lexicon-enhanced transformer with spatial-aware integration for Chinese named entity recognition

Abstract Chinese Named Entity Recognition (CNER) is a fundamental and crucial task in information extraction. In recent years, pre-trained language and lexicon-based models have proven more powerful than the previous character-based models in CNER tasks. However, existing lexicon-enhanced BERT model...

Full description

Saved in:
Bibliographic Details
Main Authors: Jiachen Huang, Shuo Liu
Format: Article
Language:English
Published: Springer 2025-07-01
Series:Complex & Intelligent Systems
Subjects:
Online Access:https://doi.org/10.1007/s40747-025-01953-2
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Abstract Chinese Named Entity Recognition (CNER) is a fundamental and crucial task in information extraction. In recent years, pre-trained language and lexicon-based models have proven more powerful than the previous character-based models in CNER tasks. However, existing lexicon-enhanced BERT models neither integrate lexical knowledge into the fundamental layers of the bidirectional transformer model nor explicitly align character features with lexicon features. In this paper, we propose a spatial-aware lexicon adapter (SALA), a neural adapter capable of dynamically integrating character and lexical representations through spatial-aware attention. SALA is incorporated between the layers of BERT to inject lexical information into the deep contextual representations of corresponding character sequences. The resulting fused vectors are further trained in SALA-BERT to enhance CNER. We evaluate SALA-BERT on various Chinese NER tasks. Compared to previous state-of-the-art models, it achieves comparable or better performance.
ISSN:2199-4536
2198-6053