<div class="csl-bib-body">
<div class="csl-entry">Wess, M., Dinakarrao, S. M. P., & Jantsch, A. (2018). Weighted Quantization-Regularization in DNNs for Weight Memory Minimization Toward HW Implementation. <i>IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems</i>, <i>37</i>(11), 2929–2939. https://doi.org/10.1109/TCAD.2018.2857080</div>
</div>
-
dc.identifier.issn
0278-0070
-
dc.identifier.uri
http://hdl.handle.net/20.500.12708/191908
-
dc.description.abstract
Deployment of deep neural networks on hardware platforms is often constrained by limited on-chip memory and computational power. The proposed weight quantization offers the possibility of optimizing weight memory alongside transforming the weights to hardware friendly data types. We apply dynamic fixed point (DFP) and power-of-two (Po2) quantization in conjunction with layer-wise precision scaling to minimize the weight memory. To alleviate accuracy degradation due to precision scaling, we employ quantization-aware fine-tuning. For fine-tuning, quantization-regularization (QR) and weighted QR are introduced to force the trained quantization by adding the distance of the weights to the desired quantization levels as a regularization term to the loss-function. While DFP quantization performs better when allowing different bit-widths for each layer, Po2 quantization in combination with retraining allows higher compression rates for equal bit-width quantization. The techniques are verified on an all-convolutional network. With accuracy degradation of 0.10% points, for DFP with layer-wise precision scaling we achieve compression ratios of 7.34 for CIFAR-10, 4.7 for CIFAR-100, and 9.33 for SVHN dataset.
en
dc.description.sponsorship
Christian Doppler Forschungsgesells
-
dc.language.iso
en
-
dc.publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
-
dc.relation.ispartof
IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems
-
dc.subject
Convolutional neural networks
en
dc.subject
memory minimization
en
dc.subject
quantization
en
dc.subject
regularization
en
dc.title
Weighted Quantization-Regularization in DNNs for Weight Memory Minimization Toward HW Implementation