收录:
摘要:
Binary neural networks (BNNs) are promising on resource-constrained devices because they reduce mem-ory consumption and accelerate inference effectively. However, they are still potential on performance improvement. Prior studies attribute performance degradation of BNNs to limited representation ability and gradient mismatch. In this paper, we find that it also results from the mandatory representation of small full-precision auxiliary weights to large values. To tackle with this issue, we propose an approach dubbed as Diluted Binary Neural Network (DBNN). Besides avoiding mandatory representation effectively, the proposed DBNN also alleviates sign flip problem to a large extent. For activations, we jointly min-imize quantization error and maximize information entropy to develop the binarization scheme. Com-pared with existing sparsity-binarization approaches, DBNN trains network from scratch without other procedures and achieves larger sparsity. Experiments on several datasets with various networks demon-strate the superiority of our approach. (c) 2023 Elsevier Ltd. All rights reserved.
关键词:
通讯作者信息:
来源 :
PATTERN RECOGNITION
ISSN: 0031-3203
年份: 2023
卷: 140
8 . 0 0 0
JCR@2022
ESI学科: ENGINEERING;
ESI高被引阀值:19
归属院系: