A dynamic CNN pruning method based on matrix similarity.

Citation metadata

Date: Mar. 2021
From: Signal, Image and Video Processing(Vol. 15, Issue 2)
Publisher: Springer
Document Type: Report; Brief article
Length: 220 words

Document controls

Main content

Abstract :

Keywords: Model compression; Feature map; Filter; Network pruning; Similarity measure Abstract Network pruning is one of the predominant approaches for deep model compression. Pruning large neural networks while maintaining their performance is often desirable because space and time complexity are reduced. Current pruning methods mainly focus on the importance of filters in the whole task. Different from previous methods, this paper focuses on the similarity between the filters or feature maps of the same layer. Firstly, cosine similarity is used as the matrix similarity measure to measure the similarity between channels, guiding the network to prune. Secondly, the proposed method is, respectively, applied to filters and feature maps pruning, and the pruning effects in different layers are summarized. Finally, we propose a method to set the pruning rate dynamically according to the situation of each layer. Our method obtains extremely sparse networks with virtually the same accuracy as the reference networks on the CIFAR-10 and ImageNet ILSVRC-12 classification tasks. On CIFAR-10, our network achieves the 52.70% compression ratio on ResNet-56 and increases only 0.13% on top-1 error. Author Affiliation: (1) College of Computer Science and Technology, China University of Petroleum, Qingdao, China (2) School of Mathematics and Statistics, Xi'an Jiaotong University, Xi'an, China (a) smw278@126.com Article History: Registration Date: 07/31/2020 Received Date: 03/25/2020 Accepted Date: 07/13/2020 Online Date: 08/10/2020 Byline:

Source Citation

Source Citation   

Gale Document Number: GALE|A651735679