Journal on Communications ›› 2022, Vol. 43 ›› Issue (4): 114-122.doi: 10.11959/j.issn.1000-436x.2022068

• Papers • Previous Articles     Next Articles

Lost-minimum post-training parameter quantization method for convolutional neural network

Fan ZHANG1, Yun HUANG2,3, Zizhuo FANG4,4, Wei GUO1   

  1. 1 National Digital Switching System Engineering &Technological R&D Center, Zhengzhou 450002, China
    2 Information Engineering University, Zhengzhou 450001, China
    3 Purple Mountain Laboratories, Nanjing 211111, China
    4 School of Cyber Science and Engineering, Southeast University, Nanjing 211189, China
  • Revised:2022-03-18 Online:2022-04-25 Published:2022-04-01
  • Supported by:
    The National Natural Science Foundation of China(61521003)

Abstract:

To solve the problem that that no dataset is available for model quantization in data-sensitive scenarios, a model quantization method without using data sets was proposed.Firstly, according to the parameters of batch normalized layer and the distribution characteristics of image data, the simulated input data was obtained by error minimization method.Then, by studying the characteristics of data rounding, a factor dynamic rounding method based on loss minimization was proposed.Through quantitative experiments on classification models such as GhostNet and target detection models such as M2Det, the effectiveness of the proposed quantification method for image classification and target detection models was verified.The experimental results show that the proposed quantization method can reduce the model size by about 75%, effectively reduce the power loss and improve the computing efficiency while basically maintaining the accuracy of the original model.

Key words: convolutional neural network, batch normalization, simulation input data, dynamic rounding

CLC Number: 

No Suggested Reading articles found!