• 综合
  • 标题
  • 关键词
  • 摘要
  • 学者
  • 期刊-刊名
  • 期刊-ISSN
  • 会议名称
搜索

作者:

Bao, Zhenshan (Bao, Zhenshan.) | Liu, Jiayang (Liu, Jiayang.) | Zhang, Wenbo (Zhang, Wenbo.)

收录:

CPCI-S EI

摘要:

As the complexity of processing issues increases, deep neural networks require more computing and storage resources. At the same time, the researchers found that the deep neural network contains a lot of redundancy, causing unnecessary waste, and the network model needs to be further optimized. Based on the above ideas, researchers have turned their attention to building more compact and efficient models in recent years, so that deep neural networks can be better deployed on nodes with limited resources to enhance their intelligence. At present, the deep neural network model compression method have weight pruning, weight quantization, and knowledge distillation and so on, these three methods have their own characteristics, which are independent of each other and can be self-contained, and can be further optimized by effective combination. This paper will construct a deep neural network model compression framework based on weight pruning, weight quantization and knowledge distillation. Firstly, the model will be double coarse-grained compression with pruning and quantization, then the original network will be used as the teacher network to guide the compressed student network. Training is performed to improve the accuracy of the student network, thereby further accelerating and compressing the model to make the loss of accuracy smaller. The experimental results show that the combination of three algorithms can compress 80% FLOPs and reduce the accuracy by only 1%.

关键词:

Model compression Pruning Knowledge distillation CNN Quantization

作者机构:

  • [ 1 ] [Bao, Zhenshan]Beijing Univ Technol, Beijing, Peoples R China
  • [ 2 ] [Liu, Jiayang]Beijing Univ Technol, Beijing, Peoples R China
  • [ 3 ] [Zhang, Wenbo]Beijing Univ Technol, Beijing, Peoples R China

通讯作者信息:

  • [Bao, Zhenshan]Beijing Univ Technol, Beijing, Peoples R China

查看成果更多字段

相关关键词:

来源 :

PROCEEDINGS OF THE 2019 2ND INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND MACHINE INTELLIGENCE (MLMI 2019)

年份: 2019

页码: 3-6

语种: 英文

被引次数:

WoS核心集被引频次: 2

SCOPUS被引频次: 1

ESI高被引论文在榜: 0 展开所有

万方被引频次:

中文被引频次:

近30日浏览量: 2

归属院系:

在线人数/总访问数:420/4954127
地址:北京工业大学图书馆(北京市朝阳区平乐园100号 邮编:100124) 联系我们:010-67392185
版权所有:北京工业大学图书馆 站点建设与维护:北京爱琴海乐之技术有限公司