• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Shi, Yong (Shi, Yong.) | Tang, Anda (Tang, Anda.) | Niu, Lingfeng (Niu, Lingfeng.) | Zhou, Ruizhi (Zhou, Ruizhi.)

Indexed by:

EI Scopus SCIE

Abstract:

Neural network pruning is a critical field aimed at reducing the infrastructure costs of neural networks by removing parameters. Traditional methods follow a fixed paradigm including pretraining, pruning, and fine-tuning. Despite the close relationship among these three stages, most pruning methods treat them as independent processes. In this paper, we propose a novel two-stage pruning method, which includes pretraining a network that is instructive for subsequent pruning, and a unified optimization model that integrates pruning and fine-tuning. Specifically, in the first stage, we design a group sparse regularized model for pretraining. This model not only safeguards the network from irreversible damage but also offers valuable insights for the pruning process. In the second stage, we introduce an element-wise sparse regularization into pruning model. This model enables us to pinpoint sparse weights more precisely than pretrained network. It automatically derives effective pruning criteria, and omits the step of fine-tuning. To implement the two-stage process in practice, we utilize stochastic gradient algorithm for the pretraining and design a threshold algorithm for pruning stage. Extensive experiments confirm the competitive performance of our proposed method in terms of both accuracy and memory cost when compared to various benchmarks. Furthermore, ablation experiments validate the effectiveness of the proposed pretraining model's guidance for the pruning process.

Keyword:

Deep neural networks Model compression Sparse optimization

Author Community:

  • [ 1 ] [Tang, Anda]Univ Chinese Acad Sci, Sch Math Sci, Beijing 100190, Peoples R China
  • [ 2 ] [Niu, Lingfeng]Univ Chinese Acad Sci, Sch Econ & Management, Beijing 100190, Peoples R China
  • [ 3 ] [Shi, Yong]Chinese Acad Sci, Res Ctr Fictitious Econ & Data Sci, Beijing 100190, Peoples R China
  • [ 4 ] [Shi, Yong]Chinese Acad Sci, Key Lab Big Data Min & Knowledge Management, Beijing 100190, Peoples R China
  • [ 5 ] [Zhou, Ruizhi]Beijing Univ Technol, Inst Operat Res & Informat Engn, Beijing 100124, Peoples R China

Reprint Author's Address:

  • [Zhou, Ruizhi]Beijing Univ Technol, Inst Operat Res & Informat Engn, Beijing 100124, Peoples R China

Show more details

Related Keywords:

Related Article:

Source :

NEUROCOMPUTING

ISSN: 0925-2312

Year: 2024

Volume: 574

6 . 0 0 0

JCR@2022

Cited Count:

WoS CC Cited Count:

SCOPUS Cited Count: 8

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 1

Affiliated Colleges:

Online/Total:530/5316029
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.