• 综合
  • 标题
  • 关键词
  • 摘要
  • 学者
  • 期刊-刊名
  • 期刊-ISSN
  • 会议名称
搜索

作者:

Hou, Xinyu (Hou, Xinyu.) | Zhu, Cui (Zhu, Cui.) | Zhu, Wenjun (Zhu, Wenjun.)

收录:

EI Scopus

摘要:

In recent times, numerous models tried to enhance the performance of Transformer on Chinese NER tasks. The model can be enhanced in two ways: one is combining it with lexicon augmentation techniques, the other is optimizing the Transformer model itself. According to research, fully connected self-attention may scatter the attention distribution, which is the reason for worse performance of the original Transformer with self-attention. In this paper, we attempt to optimize the Transformer model especially attention layer. Therefore, a novel attention mechanism, Dilated Shift Window Attention, is proposed to address this problem. By using Window Attention, this method improves the model’s capacity to deal local information, meanwhile, the model can still manage long text and long-distance dependencies owing to the Window Dilatation mechanism. Experiments on various datasets also show that DSWA replacing fully connected self-attention improves the model’s performance on the Chinese NER task. Copyright © 2023 by KSI Research Inc. and Knowledge Systems Institute, USA.

关键词:

Software engineering Knowledge engineering

作者机构:

  • [ 1 ] [Hou, Xinyu]School of Computer Science, Faculty of Information Technology, Beijing University of Technology, Beijing, China
  • [ 2 ] [Zhu, Cui]School of Computer Science, Faculty of Information Technology, Beijing University of Technology, Beijing, China
  • [ 3 ] [Zhu, Wenjun]School of Computer Science, Faculty of Information Technology, Beijing University of Technology, Beijing, China

通讯作者信息:

电子邮件地址:

查看成果更多字段

相关关键词:

相关文章:

来源 :

年份: 2023

页码: 51-57

语种: 英文

被引次数:

WoS核心集被引频次:

SCOPUS被引频次:

ESI高被引论文在榜: 0 展开所有

万方被引频次:

中文被引频次:

近30日浏览量: 0

归属院系:

在线人数/总访问数:469/4978882
地址:北京工业大学图书馆(北京市朝阳区平乐园100号 邮编:100124) 联系我们:010-67392185
版权所有:北京工业大学图书馆 站点建设与维护:北京爱琴海乐之技术有限公司