论文标题

滤波器预先修复,以改善量化的深神经网络的微调

Filter Pre-Pruning for Improved Fine-tuning of Quantized Deep Neural Networks

论文作者

Nishikawa, Jun, Ikegaya, Ryoji

论文摘要

深神经网络(DNN)具有许多参数和激活数据,并且两者都实施昂贵。减少DNN尺寸的一种方法是通过使用低位表达来量化预训练的模型,以进行权重和激活,并使用微调来恢复准确性下降。但是,通常很难训练使用低位表达的神经网络。原因之一是DNN中间层中的权重具有较宽的动态范围,因此当将宽动态范围量化为几个位时,步长变大,这会导致较大的量化误差,并最终导致准确性的较大降级。为了解决这个问题,本文在不使用任何其他学习参数和超参数的情况下做出以下三个贡献。首先,我们分析了批准是如何引起上述问题的,这会打扰量化的DNN的微调。其次,基于这些结果,我们提出了一种称为定量修剪的新修剪方法(PFQ),该方法去除过滤器,这些过滤器会干扰DNN的微调,同时又不影响推断的结果。第三,我们使用建议的修剪方法(PFQ)提出了用于量化DNN的微调工作流程。使用众所周知的模型和数据集的实验证实,所提出的方法的性能具有比传统量化方法(包括微调)相似的模型大小。

Deep Neural Networks(DNNs) have many parameters and activation data, and these both are expensive to implement. One method to reduce the size of the DNN is to quantize the pre-trained model by using a low-bit expression for weights and activations, using fine-tuning to recover the drop in accuracy. However, it is generally difficult to train neural networks which use low-bit expressions. One reason is that the weights in the middle layer of the DNN have a wide dynamic range and so when quantizing the wide dynamic range into a few bits, the step size becomes large, which leads to a large quantization error and finally a large degradation in accuracy. To solve this problem, this paper makes the following three contributions without using any additional learning parameters and hyper-parameters. First, we analyze how batch normalization, which causes the aforementioned problem, disturbs the fine-tuning of the quantized DNN. Second, based on these results, we propose a new pruning method called Pruning for Quantization (PfQ) which removes the filters that disturb the fine-tuning of the DNN while not affecting the inferred result as far as possible. Third, we propose a workflow of fine-tuning for quantized DNNs using the proposed pruning method(PfQ). Experiments using well-known models and datasets confirmed that the proposed method achieves higher performance with a similar model size than conventional quantization methods including fine-tuning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源