Please note that ISTA Research Explorer no longer supports Internet Explorer versions 8 or 9 (or earlier).

We recommend upgrading to the latest Internet Explorer, Google Chrome, or Firefox.

121 Publications


2023 | Journal Article | IST-REx-ID: 12566 | OA
Alistarh, D.-A., Ellen, F., & Rybicki, J. (2023). Wait-free approximate agreement on graphs. Theoretical Computer Science. Elsevier. https://doi.org/10.1016/j.tcs.2023.113733
[Published Version] View | Files available | DOI | WoS
 

2023 | Thesis | IST-REx-ID: 13074 | OA
Peste, E.-A. (2023). Efficiency and generalization of sparse neural networks. Institute of Science and Technology Austria. https://doi.org/10.15479/at:ista:13074
[Published Version] View | Files available | DOI
 

2023 | Journal Article | IST-REx-ID: 12330 | OA
Aksenov, V., Alistarh, D.-A., Drozdova, A., & Mohtashami, A. (2023). The splay-list: A distribution-adaptive concurrent skip-list. Distributed Computing. Springer Nature. https://doi.org/10.1007/s00446-022-00441-x
[Preprint] View | DOI | Download Preprint (ext.) | WoS | arXiv
 

2023 | Conference Paper | IST-REx-ID: 14461 | OA
Markov, I., Vladu, A., Guo, Q., & Alistarh, D.-A. (2023). Quantized distributed training of large models with convergence guarantees. In Proceedings of the 40th International Conference on Machine Learning (Vol. 202, pp. 24020–24044). Honolulu, Hawaii, HI, United States: ML Research Press.
[Preprint] View | Download Preprint (ext.) | arXiv
 

2023 | Conference Paper | IST-REx-ID: 14459 | OA
Shevchenko, A., Kögler, K., Hassani, H., & Mondelli, M. (2023). Fundamental limits of two-layer autoencoders, and achieving them with gradient methods. In Proceedings of the 40th International Conference on Machine Learning (Vol. 202, pp. 31151–31209). Honolulu, Hawaii, HI, United States: ML Research Press.
[Preprint] View | Download Preprint (ext.) | arXiv
 

2023 | Conference Paper | IST-REx-ID: 14460 | OA
Nikdan, M., Pegolotti, T., Iofinova, E. B., Kurtic, E., & Alistarh, D.-A. (2023). SparseProp: Efficient sparse backpropagation for faster training of neural networks at the edge. In Proceedings of the 40th International Conference on Machine Learning (Vol. 202, pp. 26215–26227). Honolulu, Hawaii, HI, United States: ML Research Press.
[Preprint] View | Download Preprint (ext.) | arXiv
 

2023 | Conference Paper | IST-REx-ID: 14458 | OA
Frantar, E., & Alistarh, D.-A. (2023). SparseGPT: Massive language models can be accurately pruned in one-shot. In Proceedings of the 40th International Conference on Machine Learning (Vol. 202, pp. 10323–10337). Honolulu, Hawaii, HI, United States: ML Research Press.
[Preprint] View | Download Preprint (ext.) | arXiv
 

2023 | Journal Article | IST-REx-ID: 14364 | OA
Alistarh, D.-A., Aspnes, J., Ellen, F., Gelashvili, R., & Zhu, L. (2023). Why extension-based proofs fail. SIAM Journal on Computing. Society for Industrial and Applied Mathematics. https://doi.org/10.1137/20M1375851
[Preprint] View | Files available | DOI | Download Preprint (ext.) | WoS | arXiv
 

2023 | Conference Paper | IST-REx-ID: 14771 | OA
Iofinova, E. B., Peste, E.-A., & Alistarh, D.-A. (2023). Bias in pruned vision models: In-depth analysis and countermeasures. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 24364–24373). Vancouver, BC, Canada: IEEE. https://doi.org/10.1109/cvpr52729.2023.02334
[Preprint] View | Files available | DOI | Download Preprint (ext.) | WoS | arXiv
 

2023 | Journal Article | IST-REx-ID: 14815 | OA
Beznosikov, A., Horvath, S., Richtarik, P., & Safaryan, M. (2023). On biased compression for distributed learning. Journal of Machine Learning Research. Journal of Machine Learning Research.
[Published Version] View | Files available | WoS | arXiv
 

Filters and Search Terms

department=DaAl

Search

Filter Publications