Filter Pruning via Filters Similarity in Consecutive Layers
Xiaorui Wang (Ping An Technology (Shenzhen) Co., Ltd.); Jun Wang (Ping An Technology (Shenzhen) Co. Ltd.); xin tang (Ping An property&casualty insurance company of China.LTD.); Peng Gao (Ping An Technology); Rui Fang (Ping An property&casualty insurance company of China.LTD.); Guotong Xie (Ping An Technology (Shenzhen) Co. Ltd.)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Filter pruning is widely adopted to compress and accelerate the Convolutional Neural Networks (CNNs), but most previous works ignore the relationship between filters and channels in different layers. Processing each layer independently fails to utilize the collaborative relationship across layers. In this paper, we intuitively propose a novel pruning method by explicitly leveraging the Filters Similarity in Consecutive Layers (FSCL). FSCL compresses models by pruning filters whose corresponding features are more worthless in the model. The extensive experiments demonstrate the effectiveness of FSCL, and it yields remarkable improvement over state-of-the-art on accuracy, FLOPs and parameter reduction on several benchmark models and datasets.