• 综合
  • 标题
  • 关键词
  • 摘要
  • 学者
  • 期刊-刊名
  • 期刊-ISSN
  • 会议名称
搜索

作者:

Shi, Yong (Shi, Yong.) | Tang, Anda (Tang, Anda.) | Niu, Lingfeng (Niu, Lingfeng.) | Zhou, Ruizhi (Zhou, Ruizhi.)

收录:

EI Scopus SCIE

摘要:

Neural network pruning is a critical field aimed at reducing the infrastructure costs of neural networks by removing parameters. Traditional methods follow a fixed paradigm including pretraining, pruning, and fine-tuning. Despite the close relationship among these three stages, most pruning methods treat them as independent processes. In this paper, we propose a novel two-stage pruning method, which includes pretraining a network that is instructive for subsequent pruning, and a unified optimization model that integrates pruning and fine-tuning. Specifically, in the first stage, we design a group sparse regularized model for pretraining. This model not only safeguards the network from irreversible damage but also offers valuable insights for the pruning process. In the second stage, we introduce an element-wise sparse regularization into pruning model. This model enables us to pinpoint sparse weights more precisely than pretrained network. It automatically derives effective pruning criteria, and omits the step of fine-tuning. To implement the two-stage process in practice, we utilize stochastic gradient algorithm for the pretraining and design a threshold algorithm for pruning stage. Extensive experiments confirm the competitive performance of our proposed method in terms of both accuracy and memory cost when compared to various benchmarks. Furthermore, ablation experiments validate the effectiveness of the proposed pretraining model's guidance for the pruning process.

关键词:

Deep neural networks Model compression Sparse optimization

作者机构:

  • [ 1 ] [Tang, Anda]Univ Chinese Acad Sci, Sch Math Sci, Beijing 100190, Peoples R China
  • [ 2 ] [Niu, Lingfeng]Univ Chinese Acad Sci, Sch Econ & Management, Beijing 100190, Peoples R China
  • [ 3 ] [Shi, Yong]Chinese Acad Sci, Res Ctr Fictitious Econ & Data Sci, Beijing 100190, Peoples R China
  • [ 4 ] [Shi, Yong]Chinese Acad Sci, Key Lab Big Data Min & Knowledge Management, Beijing 100190, Peoples R China
  • [ 5 ] [Zhou, Ruizhi]Beijing Univ Technol, Inst Operat Res & Informat Engn, Beijing 100124, Peoples R China

通讯作者信息:

  • [Zhou, Ruizhi]Beijing Univ Technol, Inst Operat Res & Informat Engn, Beijing 100124, Peoples R China

电子邮件地址:

查看成果更多字段

相关关键词:

相关文章:

来源 :

NEUROCOMPUTING

ISSN: 0925-2312

年份: 2024

卷: 574

6 . 0 0 0

JCR@2022

被引次数:

WoS核心集被引频次:

SCOPUS被引频次: 8

ESI高被引论文在榜: 0 展开所有

万方被引频次:

中文被引频次:

近30日浏览量: 1

归属院系:

在线人数/总访问数:461/4953995
地址:北京工业大学图书馆(北京市朝阳区平乐园100号 邮编:100124) 联系我们:010-67392185
版权所有:北京工业大学图书馆 站点建设与维护:北京爱琴海乐之技术有限公司