Chaim Baskin * 1 Eli Schwartz * 2 Evgenii Zheltonozhskii 1 Natan Liss 1 Raja Giryes 2 Alex Bronstein 1 Avi Mendelson 1 - arXiv
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
UNIQ: UNIFORM NOISE INJECTION FOR N O N - U N I F O R M Q U A N T I Z AT I O N O F N E U R A L N E T W O R K S Chaim Baskin * 1 Eli Schwartz * 2 Evgenii Zheltonozhskii 1 Natan Liss 1 Raja Giryes 2 Alex Bronstein 1 Avi Mendelson 1 A B S T R AC T arXiv:1804.10969v3 [cs.LG] 2 Oct 2018 We present a novel method for neural network quantization that emulates a non-uniform k-quantile quantizer, which adapts to the distribution of the quantized parameters. Our approach provides a novel alternative to the existing uniform quantization techniques for neural networks. We suggest to compare the results as a function of the bit-operations (BOPS) performed, assuming a look-up table availability for the non-uniform case. In this setup, we show the advantages of our strategy in the low computational budget regime. While the proposed solution is harder to implement in hardware, we believe it sets a basis for new alternatives to neural networks quantization. 1 INTRODUCTION of two reduces the amount of hardware by roughly a factor of four. (A more accurate analysis of the effects of quanti- Deep neural networks are widely used in many fields today zation is presented in Section 4.2). Quantization allows to including computer vision, signal processing, computational fit bigger networks into a device, which is especially crit- imaging, image processing, and speech and language pro- ical for embedded devices and custom hardware. On the cessing (Hinton et al., 2012; Lai et al., 2015; Chen et al., other hand, activations are passed between layers, and thus, 2016). Yet, a major drawback of these deep learning models if different layers are processed separately, activation size is their storage and computational cost. Typical deep net- reduction reduces communication overheads. works comprise millions of parameters and require billions of multiply-accumulate (MAC) operations. In many cases, Deep neural networks are usually trained and operate with this cost renders them infeasible for running on mobile de- both the weights and the activations represented in single- vices with limited resources. While some applications allow precision (32-bit) floating point. A straightforward uniform moving the inference to the cloud, such an architecture still quantization of the pre-trained weights to 16-bit fixed point incurs significant bandwidth and latency limitations. representation has been shown to have a negligible effect on the model accuracy (Gupta et al., 2015). In the majority A recent trend in research focuses on developing lighter of the applications, further reduction of precision quickly deep models, both in their memory footprint and computa- degrades performance; hence, nontrivial techniques are re- tional complexity. The main focus of most of these works, quired to carry it out. including this one, is on alleviating complexity at inference time rather than simplifying training. While training a deep Contribution. Most previous quantization solution for neu- model requires even more resources and time, it is usually ral networks assume a uniform distribution of the weights, done offline with large computational resources. which is non-optimal due to the fact these rather have bell- shaped distributions (Han et al., 2016). To facilitate this One way of reducing computational cost is quantization deficiency, we propose a k-quantile quantization method of weights and activations. Quantization of weights also with balanced (equal probability mass) bins, which is partic- reduces storage size and memory access. The bit widths ularly suitable for neural networks, where outliers or long of activations and weights affect linearly the amount of re- tails of the bell curve are less important. We also show quired hardware logic; reducing both bitwidths by a factor a simple and efficient way of reformulating this quantizer * Equal contribution 1 Department of Computer Science, Tech- using a “uniformization trick”. nion, Haifa, Israel 2 School of Electrical Engineering, Tel- Aviv University, Tel-Aviv, Israel. Correspondence to: Ev- In addition, since most existing solutions for quantization genii Zheltonozhskii , Eli struggle with training networks with quantized weights, Schwartz . we introduce a novel method for training a network that performs well with quantized values. This is achieved by Under review by the Systems and Machine Learning (SysML) adding noise (at training time) to emulate the quantization Conference.
UNIQ: Uniform Noise Injection for Quantization ResNet-18 75 ResNet-34 ResNet-50 2xResNet-18 AlexNet 70 MobileNet Accuracy [%] 65 60 Full-precision IQN (Zhou et al., 2017a) XNOR (Rastegari et al., 2016) Apprentice (Mishra & Marr, 2018) QSM (Sheng et al., 2018) 55 Distillation (Polino et al., 2018) QNN (Hubara et al., 2016b) MLQ (Xu et al., 2018) UNIQ (Ours) 50 100 101 102 103 104 105 # Bit operations (billions) Figure 1. Performance vs. complexity of different quantized neural networks. Performance is measured as top-1 accuracy on ImageNet; complexity is estimated in number of bit operations. Network architectures are denotes by different marker shapes; quantization methods are marked in different colors. Multiple instances of the same method (color) and architecture (shape) represent different level of quantization. Notice that our optimized networks are the most efficient ones among all the ones that achieves accuracy 73.4% or lower and are the most accurate ones among all the ones that consume 400 GBOPs or lower. The figure should be viewed in color. noise introduced at inference time. The uniformization to quantize, due to lower redundancy in parameters. trick renders exactly the injection of uniform noise and al- One may argue that due to the non-uniform quantization, it leviates the need to draw the noise from complicated and is harder to implement our proposed solution in hardware. bin-dependent distributions. While we limit our attention Yet, this can be alleviated by using look-up tables for calcu- to the k-quantile quantization, the proposed scheme can lating the different operations between the quantized values. work with any threshold configuration while still keeping Nonetheless, we believe that it is important to explore new the advantage of uniformly distributed noise in every bin. solutions for network quantizations beyond the standard Our proposed solution is denoted UNIQ (uniform noise uniform strategy and this work is doing such a step. injection for non-uniform qunatization). Its better noise modeling extends the uniform noise injection strategy in 2 R E L AT E D WORK (Baskin et al., 2018) that relies on uniform quantization and thus get inaccurate quantization noise modeling in the low Previous studies have investigated quantizing the network bit regime. The accurate noise model proposed here leads weights and the activations to as low as 1- or 2-bit represen- to an advantage in the low bit regime. Measuring the com- tation (Rastegari et al., 2016; Hubara et al., 2016b; Zhou plexity of each method by bit operations (BOPS), our UNIQ et al., 2016; Dong et al., 2017; Mishra et al., 2018). Such strategy achieves improved performance in the accuracy extreme reduction of the range of parameter values greatly vs. complexity tradeoff compared to other leading methods affects accuracy. Recent works proposed to use a wider such as distillation (Polino et al., 2018) and XNOR (Raste- network, i.e., one with more filters per layer, to mitigate the gari et al., 2016). In addition, our method performs well on accuracy loss (Zhu et al., 2016; Polino et al., 2018). In some smaller models targeting mobile devices, represented in the approaches, e.g., Zhu et al. (2016) and Zhou et al. (2017a), paper by MobileNet. Those network are known to be harder a learnable linear scaling layer is added after each quantized
UNIQ: Uniform Noise Injection for Quantization Table 1. Complexity-accuracy tradeoff of various DNN architectures quantized using different techniques. Complexity is reported in number of bit operations as explained in the text. Number of bits is reported as (weights,activations) and model size is calculated as the sum of parameters sizes in bits. Accuracy is top-1 accuracy on ImageNet. For each DNN architecture, rows are sorted in increasing order complexity. Note that for the same architecture with same bitwidths UNIQ may have a different complexity and model size due to fact that, unlike the others, we also quantize the first and the last layers. Compared methods: XNOR (Rastegari et al., 2016), QNN (Hubara et al., 2016b), IQN (Zhou et al., 2017a), Apprentice (Mishra & Marr, 2018), Distillation (Polino et al., 2018), QSM (Sheng et al., 2018), MLQ (Xu et al., 2018). Bits Model size Complexity Accuracy Architecture Method (w,a) [Mbit] [GBOPs] (% top-1) AlexNet QNN 1,2 15.59 15.1 51.03 AlexNet XNOR 1,32 15.6 77.5 60.10 AlexNet Baseline 32,32 498.96 1210 56.50 MobileNet UNIQ (Ours) 4,8 16.8 25.1 66.00 MobileNet UNIQ (Ours) 5,8 20.8 30.5 67.50 MobileNet UNIQ (Ours) 8,8 33.6 46.7 68.25 MobileNet QSM 8,8 33.6 46.7 68.01 MobileNet Baseline 32,32 135.2 626 68.20 ResNet-18 XNOR 1,1 4 19.9 51.20 ResNet-18 UNIQ (Ours) 4,8 46.4 93.2 67.02 ResNet-18 UNIQ (Ours) 5,8 58.4 113 68.00 ResNet-18 Apprentice 2,8 39.2 183 67.6 ResNet-18 Apprentice 4,8 61.6 220 70.40 ResNet-18 Apprentice 2,32 39.2 275 68.50 ResNet-18 IQN 5,32 72.8 359 68.89 ResNet-18 MLQ 5,32 58.4 359 69.09 ResNet-18 Distillation 4,32 61.6 403 64.20 ResNet-18 Baseline 32,32 374.4 1920 69.60 ResNet-34 UNIQ (Ours) 4,8 86.4 166 71.09 ResNet-34 UNIQ (Ours) 5,8 108.8 202 72.60 ResNet-34 Apprentice 2,8 59.2 227 71.5 ResNet-34 Apprentice 4,8 101.6 291 73.1 ResNet-34 Apprentice 2,32 59.2 398 72.8 ResNet-34 UNIQ (Ours) 4,32 86.4 519 73.1 ResNet-34 Baseline 32,32 697.6 3930 73.4 ResNet-50 UNIQ (Ours) 4,8 102.4 174 73.37 ResNet-50 Apprentice 2,8 112.8 230 72.8 ResNet-50 Apprentice 4,8 160 301 74.7 ResNet-50 Apprentice 2,32 112.8 411 74.7 ResNet-50 UNIQ (Ours) 4,32 102.4 548 74.84 ResNet-50 Baseline 32,32 817.6 4190 76.02 layer to improve expressiveness. highly accurate quantized models, but requires training of an additional, larger network. A general approach to learning a quantized model adopted in recent papers (Hubara et al., 2016a;b; Zhou et al., 2016; Most previous studies have used uniform quantization (i.e., Rastegari et al., 2016; Cai et al., 2017) is to perform the for- all quantization bins are equally-sized), which is attractive ward pass using the quantized values, while keeping another due to its simplicity. However, unless the values are uni- set of full-precision values for the backward pass updates. formly distributed, uniform quantization is not optimal. In this case, the backward pass is still differentiable, while Unfortunately, neural network weights are not uniform but the forward pass is quantized. In the aforementioned papers, rather have bell-shaped distributions (Han et al., 2016; An- a deterministic or stochastic function is used at training for derson & Berg, 2018) as we also show in the supplementary weight and activation quantization. Another approach in- material. troduced by Mishra & Marr (2018) and Polino et al. (2018) is based on a teacher-student setup for knowledge distilla- Non-uniform quantization is utilized by Han et al. (2016), tion of a full precision (and usually larger) teacher model where the authors replace the weight values with indexes to a quantized student model. This method allows training pointing to a finite codebook of shared values. Ullrich et al.
UNIQ: Uniform Noise Injection for Quantization (2017) propose to use clustering of weights as a way of (bins) {[ti−1 , ti ]}ki=1 , and let Q = {qi }ki=1 be a set of k rep- quantization. They fit a Gaussian prior model to the weights resentation levels. A quantizer QT,Q is a function mapping and use the cluster centroids as the quantization codebook. each bin [ti−1 , ti ] to the corresponding representation level Xu et al. (2018) also use such an approach but in addition qi . We denote the quantization error by E = X − QT,Q (X). embody incremental quantization both on network and layer The effect of quantization can be modeled as the addition of levels. Other examples of methods that take data distribu- random noise to X; the noise added to the i-th bin admits tion into account are the Bayesian quantization (Louizos the conditional distribution (X − qi )|X ∈ [ti−1 , ti ]. et al., 2017) and the value-aware quantization (Park et al., Most papers on neural network quantization focus on the uni- 2018). In Louizos et al. (2017), sparsity priors are imposed form quantizer, which has a constant bin width ti − ti−1 = on the network weights, providing information as to the bits ∆ and qi = (ti−1 + ti )/2, it is known to be optimal (in the to be allocated per each layer. In Molchanov et al. (2017), sense of the mean squared error EE 2 , where the expectation variational dropout, which also relies on the Bayesian frame- is taken with respect to the density fX ) in the case of uni- work, is proposed to prune weights in the network. Park form distribution. Yet, since X in neural networks is not et al. (2018) propose to leave a small part of weights and ac- uniform but rather bell shaped (Han et al., 2016), in the gen- tivations (with higher values) in full-precision, while using eral case the optimal choice, in the `2 sense, is the k-means uniform quantization for the rest. quantizer. Its name follows the property that each represen- Another approach adopted by Zhou et al. (2017a) learns tation level qi coincides with the i-th bin centroid (mean the quantizer thresholds by iteratively grouping close val- w.r.t. fX ). While finding the optimal k-means quantizer is ues and re-training the weights. Lee et al. (2017) utilize a known to be an NP-hard problem, heuristic procedures such logarithmic scale quantization for an approximation of the as the Lloyd-Max algorithm (Lloyd, 1982) usually produce `2 -optimal Lloyd quantizer (Lloyd, 1982). Cai et al. (2017) a good approximation. The k-means quantizer coincides proposed to optimize expectation of `2 error of quantization with the uniform quantizer when X is uniformly distributed. function to reduce the quantization error. Normally dis- While being a popular choice in signal processing, the k- tributed weights and half-normally distributed activations means quantizer encounters severe obstacles in our problem were assumed, that enables using a pre-calculated k-means of neural network quantization. Firstly, the Lloyd-Max al- quantizer. In Zhou et al. (2017b) balanced bins are used, so gorithm has a prohibitively high complexity to be used in each bin has the same number of samples. In some sense, every forward pass. Secondly, it is not easily amenable to this idea is the closest to our approach; yet, while Zhou et al. our scheme of modeling quantization as the addition of ran- (2017b) force the values to have an approximately uniform dom noise, as the noise distribution at every bin is complex distribution, we pose no such constraint. Also, since cal- and varies with the change of the quantization thresholds. culating percentiles is expensive in this setting, Zhou et al. Finally, our experiments shown in Section 4.3 in the sequel (2017b) estimate them with means, while our method allows suggest that the use of the `2 criterion for quantization of using the actual percentiles as detailed in the sequel. deep neural classifier does not produce the best classifica- tion results. The weights in such networks typically assume 3 NON-UNIFORM Q U A N T I Z AT I O N B Y a bell-shaped distribution with tails exerting a great effect UNIFORM NOISE INJECTION on the mean squared error, yet having little impact on the classification accuracy. To present our uniform noise injection quantization (UNIQ) method for training a neural network amenable to operation Based on empirical observations, we conjecture that the in low-precision arithmetic, we start by outlining several distribution tails, which k-means is very sensitive to, are common quantization schemes and discussing their suitabil- not essential for good model performance at least in classifi- ity for deep neural networks. Then, we suggest a training cation tasks. As an alternative to k-means, we propose the procedure where during training time uniform random ad- k-quantile quantizer characterized by the equiprobable bins ditive noise is injected into weights simulating the quanti- property, that is, P(X ∈ [ti−1 , ti ]) = 1/k. The property −1 zation error. The scheme aims at improving the quantized is realized by setting ti = FX (i/k), where FX denotes network performance at inference time, when regular deter- the cumulative distribution function of X and, accordingly, −1 ministic quantization is used. its inverse FX denotes the quantile function. The rep- resentation level of the i-th bin is set to the bin median, 3.1 Quantization qi = med{X|X ∈ [ti−1 , ti ]}. It can be shown that in the case of a uniform X, the k-quantile quantizer coincides with Let X be a random variable drawn from some distribu- the k-level uniform quantizer. tion described by the probability density function fX . Let T = {ti } with t0 = −∞, tk = ∞ and ti−1 < ti be a set of The cumulative distribution FX and the quantile function −1 thresholds partitioning the real line into k disjoint intervals FX can be estimated empirically from the distribution
UNIQ: Uniform Noise Injection for Quantization of weights, and updated in every forward pass. Alterna- as the network parameters and the quantizer thresholds are tively, one can rely on the empirical observation that the updated. `2 -regularized weights of each layer tend to follow an ap- To overcome this difficulty, we resort to the uniformiza- proximately normal distribution (Blundell et al., 2015). To tion trick outlined in the previous section. Instead of the confirm that this is the case for the networks used in the k-quantile quantizer w0 = Q(w), we apply the equiva- paper, we analyzed the distribution of weights. An exam- lent uniform quantizer to the uniformized variable, ŵ = ple of a layer-wise distribution of weights is shown in the −1 FW (Quni (FW (w))). The effect of the quantizer can be supplementary material. Relying on this observation, we −1 again modeled using noise injection, ŵ = FW (FW (w)+e), can estimate µ and σ per each layer and use the CDF of with the cardinal difference than now the noise e is uniformly the normal distribution (and its inverse, the normal quantile 1 1 distributed on the interval [− 2k , 2k ] (estimating quantiza- function). tion error distribution). Using the fact that applying a distribution function FX of Usually, quantization of neural networks is either used for X to itself results in uniform distribution allows an alter- training a model from scratch or applied post-training as native construction of the k-quantile quantizer. We apply a fine-tuning stage. Our method, as will be demonstrated the transformation U = FX (X) to the input converting it in our experiments, works well in both cases. Our practice into a uniform random variable on the interval [0, 1]. Then, shows that best results are obtained when the learning rate a uniform k-level quantizer (coinciding with the k-quantile is reduced as the noise is added; we explain this by the need quantizer) is applied to U producing Û = Quni (U ); the −1 to compensate for noisier gradients. result is transformed back into X̂ = FX (Û ) using the in- verse transformation. We refer to this procedure as to the uniformization trick. Its importance will become evident in 3.3 Gradual quantization the next section. The described method works well “as is” for small- to medium-sized neural networks. For deeper networks, the 3.2 Training quantized neural networks by uniform basic method does not perform as well, most likely due to noise injection errors arising when applying a long sequence of operations where more noise is added at each step. We found that ap- The lack of continuity, let alone smoothness, of the quan- plying the scheme gradually to small groups of layers works tization operator renders impractical its use in the back- better in deep networks. In order to perform gradual quan- ward pass. As an alternative, at training we replace the tization, we split the network into N blocks {B1 , ..., BN }, quantizer by the injection of random additive noise. This each containing about same number of consecutive layers. scheme suggests that instead of using the quantized value We also split our budget of training epochs into N stages. ŵ = QT,Q (w) of a weight w in the forward pass, ŵ = w+e At the i-th stage, we quantize and freeze the parameters in is used with e drawn from the conditional distribution of blocks {B1 , ..., Bi−1 }, and inject noise into the parameters (W − qi )|W ∈ [ti−1 , ti ] described by the density of Bi . For the rest of the blocks {Bi+1 , ..., BN } neither fW (e + qi ) noise nor quantization is applied. fE (e) = ´ ti f (w) dw ti−1 W This approach is similar to one proposed by Xu et al. (2018). defined for e ∈ [ti−1 − qi , ti − qi ] and vanishing elsewhere. This way, the number of parameters into which the noise The bin i to which w belongs is established according to its is injected simultaneously is reduced, which allows better value and is fixed during the backward pass. Quantization convergence. The deeper blocks gradually adapt to the quan- of the network activations is performed in the same manner. tization error of previous ones and thus tend to converge relatively fast when the noise is injected into them. For The fact that the parameters do not directly undergo quanti- fine-tuning a pre-trained model, we use the same scheme, zation keeps the model differentiable. In addition, gradient applying a single epoch per stage. An empirical analysis updates in the backward pass have an immediate impact on of the effect of the different number of stages is presented the forward pass, in contrast to the directly quantized model, in the supplementary material. This process can be per- where small updates often leave the parameter in the same formed iteratively, restarting from the beginning after the bin, leaving it effectively unchanged. last layer has been trained. Since this allows earlier blocks While it is customary to model the quantization error as to adapt to the changed values of the following ones, the noise with uniform distribution fE (Gray & Neuhoff, 1998), iterative process yields an additional increase in accuracy. this approximation breaks in the extremely low precision Two iterations were performed in the reported experiments. regimes (small number of quantization levels k) considered here. Hence, the injected noise has to be drawn from a poten- tially non-uniform distribution which furthermore changes
UNIQ: Uniform Noise Injection for Quantization 3.4 Activations quantization 4.1 Performance of quantized networks on ImageNet While we mostly deal with weights quantization, activations Table 1 compares the ResNet and MobileNet performance quantization is also beneficial in lowering the arithmetic with weights and activations quantized to several levels complexity and can help decreasing the communication using UNIQ and other leading approaches reported in the lit- overhead in the distributed model case. We observed that a erature. For baseline, we use a full-precision model with 32 by-product of training with noisy weights is that the model bit weights and activations. Note that the common practice becomes less sensitive to a certain level of quantization of not quantizing first and last layers significantly increases of the activations. When training with the gradual process the network complexity in BOPs. Since our model does described above, activations of the fixed layers are quantized quantize all the layers, including the first one, which often as they would be at inference time. The effect of quantized appears to be the most computationally-intensive, it achieve activations on classification accuracy is evaluated in the lower complexity requirements with a higher bitwidth. Note sequel. also the diminishing impact of the weights bitwidth on the BOP complexity as explained hereafter. 4 E X P E R I M E N TA L E VA L U AT I O N We found UNIQ to perform well also with the smaller Mo- bileNet architecture. This is in contrast to most of the pre- We performed an extensive performance analysis of the vious methods that resort to larger models and doubling UNIQ scheme compared to the current state-of-the-art meth- the number of filters (Polino et al., 2018), thus quadrupling ods for neural network quantization. The basis for com- the number of parameters, e.g., from 11 to 44 million for parison is the accuracy vs. the total number of bit oper- ResNet-18. ations in visual classification tasks on the ImageNet-1K (Russakovsky et al., 2015) dataset. CIFAR-10 dataset (Krizhevsky, 2009) is used to evaluate different design 4.2 Accuracy vs. complexity trade-off choices made, while the main evaluation is performed on Since custom precision data types are used for the network ImageNet-1K. weights and activations, the number of MAC operations MobileNet (Howard et al., 2017) and ResNet-18/34/50 (He is not an appropriate metric to describe the computational et al., 2016) architectures are used as the baseline for quan- complexity of the model. Therefore, we use the BOPs tization. MobileNet is chosen as a representative of lighter metric quantifying the number of bit operations. Given the models, which are more suited for a limited hardware setting bitwidth of two operands, it is possible to approximate the where quantization is also most likely to be used. ResNet- number of bit operations required for a basic arithmetic 18/34/50 is chosen due to its near state-of-the-art perfor- operation such as addition and multiplication. The proposed mance and popularity, which makes it an excellent reference metric is useful when the inference is performed on custom for comparison. hardware like FPGAs or ASICs. Both are a natural choice for quantized networks, due to the use of lookup tables We adopted the number of bit operations (BOPs) metric to (LUTs) and dedicated MAC (or more general DSP) units, quantify the network arithmetic complexity. This metric is which are efficient with custom data types. particularly informative about the performance of mixed- precision arithmetic especially in hardware implementations An important phenomenon that can be observed in Table 1 is on FPGAs and ASICs. the non-linear relation between the number of activation and weight bits and the resulting network complexity in BOPs. To quantify this effect, let us consider a single convolutional layer with bw -bit weights and ba -bit activations containing n input channels, m output channels, and k × k filters. The maximum value of a single output is about 2ba +bw nk 2 , which sets the accumulator width in the MAC operations Training details For quantizing a pre-trained model, we to bo = ba + bw + log2 nk 2 . The complexity of a single train with SGD for a number of epochs equal to the num- output calculation consists therefore of nk 2 ba -wide × bw - ber of trainable layers (convolution and fully connected). wide multiplications and about the same amount of bo -wide We also follow the gradual process described above with additions. This yields the total layer complexity of the number of stages set to the number of trainable layers. The learning rate is 10−4 , momentum 0.9 and weight de- BOPs ≈ mnk 2 (ba bw + ba + bw + log2 nk 2 ). cay 10−4 . In all experiments, unless otherwise stated, we fine-tuned a pre-trained model taken from an open-sourced Note that the reduction of the weight and activation bitwidth repository of pre-trained models.1 decreases the number of BOPs as long as the factor ba bw dominates the factor log2 nk 2 . Since the latter factor de- 1 https://github.com/Cadene/pretrained-models.pytorch pends only on the layer topology, this point of diminishing
UNIQ: Uniform Noise Injection for Quantization return is network architecture-dependent. Table 2. UNIQ accuracy on CIFAR-10 for different bitwidth. Another factor that must be incorporated into the BOPs Activation bits calculation is the cost of fetching the parameters from an 4 8 32 an external memory. Two assumptions are made in the Weight approximation of this cost: firstly, we assume that each 2 88.1 90.88 89.14 bits 4 89.5 91.5 89.70 parameter is only fetched once from an external memory; 32 88.52 91.32 92.00 secondly, the cost of fetching a b-bit parameter is assumed to be b BOPs. Given a neural network with n parameters all represented in b bits, the memory access cost is simply nb. Table 3. UNIQ with different quantization methods (ResNet-18 Figure 1 and Table 1 display the performance-complexity top-1 accuracy on CIFAR-10, 3-bit weights) tradeoffs of various neural networks trained using UNIQ and other methods to different levels of weight and activation Quantization method Accuracy Training time [h] quantization. Notice that since we quantize the first and Baseline (unquantized) 92.00 1.42 last layers, our ResNet-34 network has better computational k-quantile 91.30 2.28 k-means 85.80 5.37 complexity and better accuracy compared to all competing Uniform 84.93 5.37 ResNet-18 networks. The same holds for our ResNet-50 compared to all competing ResNet-34. to around 280% increase in training time required for the 4.3 Ablation study k-means quantizer. In addition, k-quantile training time is independent on the number of quantization bins as the Accuracy vs. quantization level. We tested the effect of noise distribution is same for every bin while the other training ResNet-18 on CIFAR-10 with UNIQ for various methods require separate processing of each bin, increasing levels of weight and activation quantization. Table 2 reports the training time for higher bitwidths. the results. We observed that for such a small dataset the quantization of activations and weights helps avoid over- fitting and the results for quantized model come very close Training from scratch vs. fine-tuning. Both training to those with full precision. from scratch (that is, from random initialization) and fine- tuning have their advantages and disadvantages. Training from scratch takes more time but requires a single training Comparison of different quantizers. In the following phase with no extra training epochs, at the end of which a experiment, different quantizers were compared within the quantized model is obtained. Fine-tuning, on the other hand, uniform noise injection scheme. is useful when a pre-trained full-precision model is already The bins of the uniform quantizer were allocated evenly in available; it can then be quantized with a short re-training. the range [−3σ, 3σ] with σ denoting the standard deviation Table 4 compares the accuracy achieved in the two regimes of the parameters. For both the k-quantile and the k-means on a narrow version of ResNet-18 trained on CIFAR-10 and quantizers, normal distribution of the weights was assumed 100. 5-bit quantization of weights only and 5-bit quantiza- and the normal cumulative distribution and quantile func- tion of both weights and activations were compared. We tions were used for the uniformization and deuniformization found that both regimes work equally well, reaching accu- of the quantized parameter. The k-means and uniform quan- racy close to the full-precision baseline. tizers used a pre-calculated set of thresholds translated to the uniformized domain. Since the resulting bins in the uniformized domain had different widths, the level of noise Table 4. Top-1 accuracy (in percent) on CIFAR-10 and 100 of a was different in each bin. This required an additional step narrow version on ResNet-18 trained with UNIQ from random of finding the bin index for each parameter approximately initialization vs. fine-tuning a full-precision model. Number of bits is reported as (weights,activations). doubling the training time. The three quantizers were evaluated in a ResNet-18 network Dataset Bits Full training Fine-tuning Baseline trained on the CIFAR-10 dataset with weights quantized to 3 bits (k = 8) and activations computed in full precision (32 5,32 93.8 90.9 CIFAR-10 92.0 5,5 91.56 91.21 bit). Table 3 reports the obtained top-1 accuracy. k-quantile quantization outperforms other quantization methods and 5,32 66.54 65.73 CIFAR-100 66.3 is only slightly inferior to the full-precision baseline. In 5,5 65.29 65.05 terms of training time, the k-quantile quantizer requires about 60% more time to train for k = 8; this is compared
UNIQ: Uniform Noise Injection for Quantization Accuracy vs. number of quantization stages. We found quantizer that was investigated in this paper. that injecting noise to all layers simultaneously does not To set a common basis for comparison, we have suggested perform well for deeper networks. As described in Section measuring the complexity of each method by BOPs. In this 3.3, we suggest splitting the training into N stages, such case, we achieved improved results for quantized neural that at each stage the noise is injected only into a subset of networks on ImageNet. Our solution outperforms any other layers. quantized network in terms of accuracy in the low com- To determine the optimal number of stages, we fine-tuned putational range (< 400 GBOPs). Our MobileNet based ResNet-18 on CIFAR-10 with a fixed 18 epoch budget. Bit solution achieves 66% accuracy with slightly more compu- width was set to 4 for both the weights and the activations. tational resources compared to the 50% achieved with ex- tremely quantized XNOR networks (Rastegari et al., 2016) Figure 2 reports the classification accuracy as a function of and QNN (Hubara et al., 2016b). the number of quantization stages. Based on these results, we conclude that the best strategy is injecting noise to a While this paper considered a setting in which all parameters single layer at each stage. We follow this strategy in all have the same bitwidth, more complicated bit allocations ResNet-18 experiments conducted in this paper. For Mo- will be explored in following studies. The proposed scheme bileNet, since it is deeper and includes 28 layers, we chose is not restricted to the k-quantile quantizer discussed in this to inject noise to 2 consecutive layers at every stage. paper, but rather applies to any quantizer. In the general case, the noise injected into each bin is uniform, but its variance changes with the bin width. 90.8 90.7 REFERENCES Anderson, A. G. and Berg, C. P. The high-dimensional 90.6 geometry of binary neural networks. International Con- Accuracy [%] ference on Learning Representations (ICLR), 2018. 90.5 Baskin, C., Liss, N., Chai, Y., Zheltonozhkii, E., Schwartz, 90.4 E., Giryes, R., Mendelson, A., and Bronstein, A. M. NICE: Noise Injection and Clamping Estimation for Neu- 90.3 ral Network Quantization. ArXiv e-prints, September 2018. 90.2 2 4 8 16 Blundell, C., Cornebise, J., Kavukcuoglu, K., and Wierstra, Number of stages D. Weight uncertainty in neural network. In International Conference on Machine Learning, pp. 1613–1622, 2015. Figure 2. Classification accuracy on CIFAR-10 of the ResNet-18 architecture quantized using UNIQ with different number of quan- Cai, Z., He, X., Sun, J., and Vasconcelos, N. Deep learning tization stages during training. with low precision by half-wave gaussian quantization. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2017. Weight distribution. We have verified that neural net- Chen, L., Papandreou, G., Kokkinos, I., Murphy, K., and works weights follow Gaussian distribution. We observed Yuille, A. L. Deeplab: Semantic image segmentation this is true for most of the layers with Shapiro-Wilk test with deep convolutional nets, atrous convolution, and statistic higher than 0.82 for all of the layers, as can be seen fully connected crfs. CoRR, abs/1606.00915, 2016. URL in Figure 3. http://arxiv.org/abs/1606.00915. Dong, Y., Ni, R., Li, J., Chen, Y., Zhu, J., and Su, H. Learn- 5 CONCLUSION ing accurate low-bit deep neural networks with stochas- We introduced UNIQ – a training scheme for quantized tic quantization. In British Machine Vision Conference neural networks. The scheme is based on the uniformization (BMVC’17), 2017. of the distribution of the quantized parameters, injection of Gray, R. M. and Neuhoff, D. L. Quantization. IEEE trans- additive uniform noise, followed by the de-uniformization actions on information theory, 44(6):2325–2383, 1998. of the obtained result. The scheme is amenable to efficient training by back propagation in full precision arithmetic, and Gupta, S., Agrawal, A., Gopalakrishnan, K., and Narayanan, achieves maximum efficiency with the k-quantile (balanced) P. Deep learning with limited numerical precision. In
UNIQ: Uniform Noise Injection for Quantization conv1 W=0.86 layer1.0.conv1 W=0.82 layer1.0.conv2 W=0.95 layer1.1.conv1 W=0.90 layer1.1.conv2 W=0.96 0.5 0.0 0.5 1.0 0.5 0.0 0.5 0.4 0.2 0.0 0.2 0.5 0.0 0.5 0.4 0.2 0.0 0.2 layer2.0.conv1 W=0.95 layer2.0.conv2 W=0.94 layer2.0.downsample.0 W=0.85 layer2.1.conv1 W=0.94 layer2.1.conv2 W=0.97 0.2 0.0 0.2 0.2 0.0 0.2 0.4 0.5 0.0 0.5 0.4 0.2 0.0 0.2 0.4 0.2 0.0 0.2 layer3.0.conv1 W=0.95 layer3.0.conv2 W=0.96 layer3.0.downsample.0 W=0.96 layer3.1.conv1 W=0.97 layer3.1.conv2 W=0.98 0.4 0.2 0.0 0.2 0.4 0.2 0.0 0.2 0.2 0.0 0.2 0.2 0.0 0.2 0.2 0.0 0.2 layer4.0.conv1 W=0.98 layer4.0.conv2 W=0.99 layer4.0.downsample.0 W=0.96 layer4.1.conv1 W=0.99 layer4.1.conv2 W=0.97 0.2 0.0 0.2 0.4 0.2 0.0 0.2 0.5 0.0 0.5 0.2 0.0 0.2 0.2 0.0 0.2 last_linear W=0.91 0.25 0.00 0.25 0.50 0.75 Figure 3. Distribution of weights for the layers of pretrained ResNet-18 network. For each layer Shapiro-Wilk test statistic W is mentioned. Proceedings of the 32nd International Conference on ing for image recognition. In Proceedings of the IEEE Machine Learning (ICML-15), pp. 1737–1746, 2015. conference on computer vision and pattern recognition, pp. 770–778, 2016. Han, S., Mao, H., and Dally, W. J. Deep compression: Compressing deep neural networks with pruning, trained Hinton, G., Deng, L., Yu, D., Dahl, G. E., Mohamed, A.-r., quantization and huffman coding. International Confer- Jaitly, N., Senior, A., Vanhoucke, V., Nguyen, P., Sainath, ence on Learning Representations (ICLR), 2016. T. N., and Kingsbury, B. Deep neural networks for acous- tic modeling in speech recognition: The shared views He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learn- of four research groups. IEEE Signal Processing Mag-
UNIQ: Uniform Noise Injection for Quantization azine, 29(6):82–97, Nov 2012. ISSN 1053-5888. doi: Molchanov, D., Ashukha, A., and Vetrov, D. Variational 10.1109/MSP.2012.2205597. dropout sparsifies deep neural networks. In International Conference on Machine Learning, 2017. Howard, A. G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., and Adam, H. Mobilenets: Park, E., Yoo, S., and Vajda, P. Value-aware quantization for Efficient convolutional neural networks for mobile vision training and inference of neural networks. arXiv preprint applications. arXiv preprint arXiv:1704.04861, 2017. arXiv:1804.07802, 2018. Hubara, I., Courbariaux, M., Soudry, D., El-Yaniv, R., and Polino, A., Pascanu, R., and Alistarh, D. Model Bengio, Y. Binarized neural networks. In Advances in compression via distillation and quantization. In- neural information processing systems, pp. 4107–4115, ternational Conference on Learning Representations, 2016a. 2018. URL https://openreview.net/forum? id=S1XolQbRW. Hubara, I., Courbariaux, M., Soudry, D., El-Yaniv, R., and Bengio, Y. Quantized neural networks: Training neu- Rastegari, M., Ordonez, V., Redmon, J., and Farhadi, A. ral networks with low precision weights and activations. Xnor-net: Imagenet classification using binary convo- arXiv preprint arXiv:1609.07061, 2016b. lutional neural networks. In European Conference on Computer Vision, pp. 525–542. Springer, 2016. Krizhevsky, A. Learning multiple layers of features from tiny images. Master’s thesis, Department of Computer Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Science, University of Toronto, 2009. Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., Berg, A. C., and Fei-Fei, L. ImageNet Large Scale Lai, S., Xu, L., Liu, K., and Zhao, J. Recurrent convolutional Visual Recognition Challenge. International Journal of neural networks for text classification. In Proceedings Computer Vision (IJCV), 115(3):211–252, 2015. doi: of the Twenty-Ninth AAAI Conference on Artificial Intel- 10.1007/s11263-015-0816-y. ligence, AAAI’15, pp. 2267–2273. AAAI Press, 2015. ISBN 0-262-51129-0. URL http://dl.acm.org/ Sheng, T., Feng, C., Zhuo, S., Zhang, X., Shen, L., and citation.cfm?id=2886521.2886636. Aleksic, M. A quantization-friendly separable convolu- tion for mobilenets. CoRR, abs/1803.08607, 2018. URL Lee, E. H., Miyashita, D., Chai, E., Murmann, B., and http://arxiv.org/abs/1803.08607. Wong, S. S. Lognet: Energy-efficient neural networks Ullrich, K., Meeds, E., and Welling, M. Soft weight-sharing using logarithmic computation. In Acoustics, Speech and for neural network compression. International Confer- Signal Processing (ICASSP), 2017 IEEE International ence on Learning Representations (ICLR), 2017. Conference on, pp. 5900–5904. IEEE, 2017. Xu, Y., Wang, Y., Zhou, A., Lin, W., and Xiong, Lloyd, S. Least squares quantization in pcm. IEEE transac- H. Deep neural network compression with sin- tions on information theory, 28(2):129–137, 1982. gle and multiple level quantization, 2018. URL Louizos, C., Ullrich, K., and Welling, M. Bayesian https://www.aaai.org/ocs/index.php/ compression for deep learning. In Guyon, I., AAAI/AAAI18/paper/view/16479. Luxburg, U. V., Bengio, S., Wallach, H., Fer- Zhou, A., Yao, A., Guo, Y., Xu, L., and Chen, Y. Incre- gus, R., Vishwanathan, S., and Garnett, R. (eds.), mental network quantization: Towards lossless cnns with Advances in Neural Information Processing Sys- low-precision weights. In International Conference on tems 30, pp. 3288–3298. Curran Associates, Inc., Learning Representations,ICLR2017, 2017a. 2017. URL http://papers.nips.cc/paper/ 6921-bayesian-compression-for-deep-learning. Zhou, S., Wu, Y., Ni, Z., Zhou, X., Wen, H., and Zou, Y. pdf. Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint Mishra, A. and Marr, D. Apprentice: Using knowledge dis- arXiv:1606.06160, 2016. tillation techniques to improve low-precision network ac- curacy. International Conference on Learning Represen- Zhou, S.-C., Wang, Y.-Z., Wen, H., He, Q.-Y., and Zou, tations, 2018. URL https://openreview.net/ Y.-H. Balanced quantization: An effective and efficient forum?id=B1ae1lZRb. approach to quantized neural networks. Journal of Com- puter Science and Technology, 32(4):667–682, 2017b. Mishra, A., Nurvitadhi, E., Cook, J. J., and Marr, D. Wrpn: Wide reduced-precision networks. International Confer- Zhu, C., Han, S., Mao, H., and Dally, W. J. Trained ternary ence on Learning Representations, 2018. URL https: quantization. International Conference on Learning Rep- //openreview.net/forum?id=B1ZvaaeAZ. resentations (ICLR), 2016.
You can also read