• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Xiao, Yinlong (Xiao, Yinlong.) | Li, Jianqiang (Li, Jianqiang.) | Zhao, Qing (Zhao, Qing.) | Zhu, Qing (Zhu, Qing.) | Wei, Yu-Chih (Wei, Yu-Chih.)

Indexed by:

EI Scopus

Abstract:

Recently, lexicon-based Chinese Named Entity Recognition (NER) models have achieved state-of-the-art performance by benefiting from the rich boundary and semantic information contained in the lexicon. However, in the Chinese medical domain, it's difficult to obtain the medical lexicon related to the target medical corpus. In this paper, we propose a new paradigm, enhancing Chinese medical NER with Auto-mined Lexicon (ALNER), which alleviates the difficulty of obtaining the medical lexicon by designing a data-driven automatic lexicon construction method. We define medical lexicon construction as a high-quality phrase mining task. We perform secondary annotation on the NER annotated data and use the secondary annotated data to train a deep learning-based phrase tagger. Experimental results show that our method can be combined with different lexicon-based Chinese NER models to improve performance and that the method does not require an external medical lexicon. © 2022 IEEE.

Keyword:

Natural language processing systems Semantics Deep learning

Author Community:

  • [ 1 ] [Xiao, Yinlong]Beijing University of Technology, Faculty of Information Technology, Beijing, China
  • [ 2 ] [Li, Jianqiang]Beijing University of Technology, Faculty of Information Technology, Beijing, China
  • [ 3 ] [Zhao, Qing]Beijing University of Technology, Faculty of Information Technology, Beijing, China
  • [ 4 ] [Zhu, Qing]Beijing University of Technology, Faculty of Information Technology, Beijing, China
  • [ 5 ] [Wei, Yu-Chih]National Taipei University of Technology, Department of Information and Finance Management, Taipei, Taiwan

Reprint Author's Address:

Email:

Show more details

Related Keywords:

Related Article:

Source :

ISSN: 1062-922X

Year: 2022

Volume: 2022-October

Page: 2403-2408

Language: English

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count:

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 1

Affiliated Colleges:

Online/Total:758/5314192
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.