UNIQ: Uniform Noise Injection for Non-Uniform Quantization of neural networks
Baskin C, Liss N, Schwartz E, Zheltonozhskii E, Giryes R, Bronstein AM, Mendelson A. 2021. UNIQ: Uniform Noise Injection for Non-Uniform Quantization of neural networks. ACM Transactions on Computer Systems. 37(1–4), 1–15.
Download (ext.)
https://doi.org/10.48550/arXiv.1804.10969
[Preprint]
DOI
Journal Article
| Published
| English
Scopus indexed
Author
Baskin, Chaim;
Liss, Natan;
Schwartz, Eli;
Zheltonozhskii, Evgenii;
Giryes, Raja;
Bronstein, Alex M.ISTA ;
Mendelson, Avi
Abstract
We present a novel method for neural network quantization. Our method, named UNIQ, emulates a non-uniform k-quantile quantizer and adapts the model to perform well with quantized weights by injecting noise to the weights at training time. As a by-product of injecting noise to weights, we find that activations can also be quantized to as low as 8-bit with only a minor accuracy degradation. Our non-uniform quantization approach provides a novel alternative to the existing uniform quantization techniques for neural networks. We further propose a novel complexity metric of number of bit operations performed (BOPs), and we show that this metric has a linear relation with logic utilization and power. We suggest evaluating the trade-off of accuracy vs. complexity (BOPs). The proposed method, when evaluated on ResNet18/34/50 and MobileNet on ImageNet, outperforms the prior state of the art both in the low-complexity regime and the high accuracy regime. We demonstrate the practical applicability of this approach, by implementing our non-uniformly quantized CNN on FPGA.
Publishing Year
Date Published
2021-03-26
Journal Title
ACM Transactions on Computer Systems
Publisher
Association for Computing Machinery
Volume
37
Issue
1-4
Page
1-15
ISSN
eISSN
IST-REx-ID
Cite this
Baskin C, Liss N, Schwartz E, et al. UNIQ: Uniform Noise Injection for Non-Uniform Quantization of neural networks. ACM Transactions on Computer Systems. 2021;37(1-4):1-15. doi:10.1145/3444943
Baskin, C., Liss, N., Schwartz, E., Zheltonozhskii, E., Giryes, R., Bronstein, A. M., & Mendelson, A. (2021). UNIQ: Uniform Noise Injection for Non-Uniform Quantization of neural networks. ACM Transactions on Computer Systems. Association for Computing Machinery. https://doi.org/10.1145/3444943
Baskin, Chaim, Natan Liss, Eli Schwartz, Evgenii Zheltonozhskii, Raja Giryes, Alex M. Bronstein, and Avi Mendelson. “UNIQ: Uniform Noise Injection for Non-Uniform Quantization of Neural Networks.” ACM Transactions on Computer Systems. Association for Computing Machinery, 2021. https://doi.org/10.1145/3444943.
C. Baskin et al., “UNIQ: Uniform Noise Injection for Non-Uniform Quantization of neural networks,” ACM Transactions on Computer Systems, vol. 37, no. 1–4. Association for Computing Machinery, pp. 1–15, 2021.
Baskin C, Liss N, Schwartz E, Zheltonozhskii E, Giryes R, Bronstein AM, Mendelson A. 2021. UNIQ: Uniform Noise Injection for Non-Uniform Quantization of neural networks. ACM Transactions on Computer Systems. 37(1–4), 1–15.
Baskin, Chaim, et al. “UNIQ: Uniform Noise Injection for Non-Uniform Quantization of Neural Networks.” ACM Transactions on Computer Systems, vol. 37, no. 1–4, Association for Computing Machinery, 2021, pp. 1–15, doi:10.1145/3444943.
All files available under the following license(s):
Copyright Statement:
This Item is protected by copyright and/or related rights. [...]
Link(s) to Main File(s)
Access Level
Open Access
Export
Marked PublicationsOpen Data ISTA Research Explorer
Sources
arXiv 1804.10969