Optimization based Layer-wise Magnitude-based Pruning for DNN Compression
Optimization based Layer-wise Magnitude-based Pruning for DNN Compression
Guiying Li, Chao Qian, Chunhui Jiang, Xiaofen Lu, Ke Tang
Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence
Main track. Pages 2383-2389.
https://doi.org/10.24963/ijcai.2018/330
Layer-wise magnitude-based pruning (LMP) is a very popular method for deep neural network (DNN) compression. However, tuning the layer-specific thresholds is a difficult task, since the space of threshold candidates is exponentially large and the evaluation is very expensive. Previous methods are mainly by hand and require expertise. In this paper, we propose an automatic tuning approach based on optimization, named OLMP. The idea is to transform the threshold tuning problem into a constrained optimization problem (i.e., minimizing the size of the pruned model subject to a constraint on the accuracy loss), and then use powerful derivative-free optimization algorithms to solve it. To compress a trained DNN, OLMP is conducted within a new iterative pruning and adjusting pipeline. Empirical results show that OLMP can achieve the best pruning ratio on LeNet-style models (i.e., 114 times for LeNet-300-100 and 298 times for LeNet-5) compared with some state-of-the- art DNN pruning methods, and can reduce the size of an AlexNet-style network up to 82 times without accuracy loss.
Keywords:
Machine Learning: Neural Networks
Constraints and SAT: Constraint Optimisation
Machine Learning: Feature Selection ; Learning Sparse Models
Heuristic Search and Game Playing: Combinatorial Search and Optimisation
Machine Learning: Deep Learning
Heuristic Search and Game Playing: Heuristic Search and Machine Learning