Communication-efficient distributed optimization with quantized preconditioners

Alimisis F, Davies P, Alistarh D-A. 2021. Communication-efficient distributed optimization with quantized preconditioners. Proceedings of the 38th International Conference on Machine Learning. International Conference on Machine Learning vol. 139, 196–206.

Download
OA 2021_PMLR_Alimisis.pdf 429.09 KB
Conference Paper | Published | English

Scopus indexed
Author
Department
Abstract
We investigate fast and communication-efficient algorithms for the classic problem of minimizing a sum of strongly convex and smooth functions that are distributed among n different nodes, which can communicate using a limited number of bits. Most previous communication-efficient approaches for this problem are limited to first-order optimization, and therefore have \emph{linear} dependence on the condition number in their communication complexity. We show that this dependence is not inherent: communication-efficient methods can in fact have sublinear dependence on the condition number. For this, we design and analyze the first communication-efficient distributed variants of preconditioned gradient descent for Generalized Linear Models, and for Newton’s method. Our results rely on a new technique for quantizing both the preconditioner and the descent direction at each step of the algorithms, while controlling their convergence rate. We also validate our findings experimentally, showing faster convergence and reduced communication relative to previous methods.
Publishing Year
Date Published
2021-07-01
Proceedings Title
Proceedings of the 38th International Conference on Machine Learning
Acknowledgement
The authors would like to thank Janne Korhonen, Aurelien Lucchi, Celestine MendlerDunner and Antonio Orvieto for helpful discussions. FA ¨and DA were supported during this work by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 805223 ScaleML). PD was supported by the European Union’s Horizon 2020 programme under the Marie Skłodowska-Curie grant agreement No. 754411.
Volume
139
Page
196-206
Conference
International Conference on Machine Learning
Conference Location
Virtual
Conference Date
2021-07-18 – 2021-07-24
eISSN
IST-REx-ID

Cite this

Alimisis F, Davies P, Alistarh D-A. Communication-efficient distributed optimization with quantized preconditioners. In: Proceedings of the 38th International Conference on Machine Learning. Vol 139. ML Research Press; 2021:196-206.
Alimisis, F., Davies, P., & Alistarh, D.-A. (2021). Communication-efficient distributed optimization with quantized preconditioners. In Proceedings of the 38th International Conference on Machine Learning (Vol. 139, pp. 196–206). Virtual: ML Research Press.
Alimisis, Foivos, Peter Davies, and Dan-Adrian Alistarh. “Communication-Efficient Distributed Optimization with Quantized Preconditioners.” In Proceedings of the 38th International Conference on Machine Learning, 139:196–206. ML Research Press, 2021.
F. Alimisis, P. Davies, and D.-A. Alistarh, “Communication-efficient distributed optimization with quantized preconditioners,” in Proceedings of the 38th International Conference on Machine Learning, Virtual, 2021, vol. 139, pp. 196–206.
Alimisis F, Davies P, Alistarh D-A. 2021. Communication-efficient distributed optimization with quantized preconditioners. Proceedings of the 38th International Conference on Machine Learning. International Conference on Machine Learning vol. 139, 196–206.
Alimisis, Foivos, et al. “Communication-Efficient Distributed Optimization with Quantized Preconditioners.” Proceedings of the 38th International Conference on Machine Learning, vol. 139, ML Research Press, 2021, pp. 196–206.
All files available under the following license(s):
Creative Commons Attribution 4.0 International Public License (CC-BY 4.0):
Main File(s)
File Name
Access Level
OA Open Access
Date Uploaded
2023-06-19
MD5 Checksum
7ec0d59bac268b49c76bf2e036dedd7a


Export

Marked Publications

Open Data ISTA Research Explorer

Sources

arXiv 2102.07214

Search this title in

Google Scholar
ISBN Search