Please note that ISTA Research Explorer no longer supports Internet Explorer versions 8 or 9 (or earlier).

We recommend upgrading to the latest Internet Explorer, Google Chrome, or Firefox.

121 Publications


2022 | Conference Paper | IST-REx-ID: 11180 | OA
Postnikova A, Koval N, Nadiradze G, Alistarh D-A. Multi-queues can be state-of-the-art priority schedulers. In: Proceedings of the 27th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming. Association for Computing Machinery; 2022:353-367. doi:10.1145/3503221.3508432
[Preprint] View | Files available | DOI | Download Preprint (ext.) | WoS | arXiv
 

2022 | Research Data Reference | IST-REx-ID: 13076 | OA
Postnikova A, Koval N, Nadiradze G, Alistarh D-A. Multi-queues can be state-of-the-art priority schedulers. 2022. doi:10.5281/ZENODO.5733408
[Published Version] View | Files available | DOI | Download Published Version (ext.)
 

2022 | Conference Paper | IST-REx-ID: 11707 | OA
Balliu A, Hirvonen J, Melnyk D, Olivetti D, Rybicki J, Suomela J. Local mending. In: Parter M, ed. International Colloquium on Structural Information and Communication Complexity. Vol 13298. LNCS. Springer Nature; 2022:1-20. doi:10.1007/978-3-031-09993-9_1
[Preprint] View | DOI | Download Preprint (ext.) | WoS | arXiv
 

2022 | Conference Paper | IST-REx-ID: 12299 | OA
Iofinova EB, Peste E-A, Kurtz M, Alistarh D-A. How well do sparse ImageNet models transfer? In: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Institute of Electrical and Electronics Engineers; 2022:12256-12266. doi:10.1109/cvpr52688.2022.01195
[Preprint] View | Files available | DOI | Download Preprint (ext.) | WoS | arXiv
 

2021 | Journal Article | IST-REx-ID: 10180 | OA
Hoefler T, Alistarh D-A, Ben-Nun T, Dryden N, Peste E-A. Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks. Journal of Machine Learning Research. 2021;22(241):1-124.
[Published Version] View | Files available | Download Published Version (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 10218 | OA
Alistarh D-A, Gelashvili R, Rybicki J. Brief announcement: Fast graphical population protocols. In: 35th International Symposium on Distributed Computing. Vol 209. Schloss Dagstuhl - Leibniz-Zentrum für Informatik; 2021. doi:10.4230/LIPIcs.DISC.2021.43
[Published Version] View | Files available | DOI | arXiv
 

2021 | Conference Paper | IST-REx-ID: 10217 | OA
Alistarh D-A, Gelashvili R, Nadiradze G. Lower bounds for shared-memory leader election under bounded write contention. In: 35th International Symposium on Distributed Computing. Vol 209. Schloss Dagstuhl - Leibniz Zentrum für Informatik; 2021. doi:10.4230/LIPIcs.DISC.2021.4
[Published Version] View | Files available | DOI
 

2021 | Conference Paper | IST-REx-ID: 10216 | OA
Chatterjee B, Peri S, Sa M. Brief announcement: Non-blocking dynamic unbounded graphs with worst-case amortized bounds. In: 35th International Symposium on Distributed Computing. Vol 209. Schloss Dagstuhl - Leibniz Zentrum für Informatik; 2021. doi:10.4230/LIPIcs.DISC.2021.52
[Published Version] View | Files available | DOI | arXiv
 

2021 | Conference Paper | IST-REx-ID: 10219 | OA
Korhonen J, Paz A, Rybicki J, Schmid S, Suomela J. Brief announcement: Sinkless orientation is hard also in the supported LOCAL model. In: 35th International Symposium on Distributed Computing. Vol 209. Schloss Dagstuhl - Leibniz Zentrum für Informatik; 2021. doi:10.4230/LIPIcs.DISC.2021.58
[Published Version] View | Files available | DOI | arXiv
 

2021 | Conference Paper | IST-REx-ID: 10853 | OA
Fedorov A, Koval N, Alistarh D-A. A scalable concurrent algorithm for dynamic connectivity. In: Proceedings of the 33rd ACM Symposium on Parallelism in Algorithms and Architectures. Association for Computing Machinery; 2021:208-220. doi:10.1145/3409964.3461810
[Preprint] View | DOI | Download Preprint (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 11436 | OA
Kungurtsev V, Egan M, Chatterjee B, Alistarh D-A. Asynchronous optimization methods for efficient training of deep neural networks with guarantees. In: 35th AAAI Conference on Artificial Intelligence, AAAI 2021. Vol 35. AAAI Press; 2021:8209-8216.
[Preprint] View | Download Preprint (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 11452 | OA
Alimisis F, Davies P, Vandereycken B, Alistarh D-A. Distributed principal component analysis with limited communication. In: Advances in Neural Information Processing Systems - 35th Conference on Neural Information Processing Systems. Vol 4. Neural Information Processing Systems Foundation; 2021:2823-2834.
[Published Version] View | Download Published Version (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 11463 | OA
Frantar E, Kurtic E, Alistarh D-A. M-FAC: Efficient matrix-free approximations of second-order information. In: 35th Conference on Neural Information Processing Systems. Vol 34. Curran Associates; 2021:14873-14886.
[Published Version] View | Download Published Version (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 11464 | OA
Alistarh D-A, Korhonen J. Towards tight communication lower bounds for distributed optimisation. In: 35th Conference on Neural Information Processing Systems. Vol 34. Curran Associates; 2021:7254-7266.
[Published Version] View | Download Published Version (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 9543 | OA
Davies P, Gurunanthan V, Moshrefi N, Ashkboos S, Alistarh D-A. New bounds for distributed mean estimation and variance reduction. In: 9th International Conference on Learning Representations. ; 2021.
[Published Version] View | Download Published Version (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 9620 | OA
Alistarh D-A, Davies P. Collecting coupons is faster with friends. In: Structural Information and Communication Complexity. Vol 12810. Springer Nature; 2021:3-12. doi:10.1007/978-3-030-79527-6_1
[Preprint] View | Files available | DOI
 

2021 | Conference Paper | IST-REx-ID: 9823 | OA
Alistarh D-A, Ellen F, Rybicki J. Wait-free approximate agreement on graphs. In: Structural Information and Communication Complexity. Vol 12810. Springer Nature; 2021:87-105. doi:10.1007/978-3-030-79527-6_6
[Preprint] View | DOI | Download Preprint (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 11458 | OA
Peste E-A, Iofinova EB, Vladu A, Alistarh D-A. AC/DC: Alternating Compressed/DeCompressed training of deep neural networks. In: 35th Conference on Neural Information Processing Systems. Vol 34. Curran Associates; 2021:8557-8570.
[Published Version] View | Files available | Download Published Version (ext.) | arXiv
 

2021 | Conference Paper | IST-REx-ID: 13147 | OA
Alimisis F, Davies P, Alistarh D-A. Communication-efficient distributed optimization with quantized preconditioners. In: Proceedings of the 38th International Conference on Machine Learning. Vol 139. ML Research Press; 2021:196-206.
[Published Version] View | Files available | arXiv
 

2021 | Journal Article | IST-REx-ID: 8723 | OA
Li S, Tal Ben-Nun TB-N, Nadiradze G, et al. Breaking (global) barriers in parallel stochastic optimization with wait-avoiding group averaging. IEEE Transactions on Parallel and Distributed Systems. 2021;32(7). doi:10.1109/TPDS.2020.3040606
[Preprint] View | DOI | Download Preprint (ext.) | WoS | arXiv
 

Filters and Search Terms

department=DaAl

Search

Filter Publications