Please note that LibreCat no longer supports Internet Explorer versions 8 or 9 (or earlier).

We recommend upgrading to the latest Internet Explorer, Google Chrome, or Firefox.




141 Publications

2024 | Published | Thesis | IST-REx-ID: 17485 | OA
Frantar, Elias. “Compressing Large Neural Networks : Algorithms, Systems and Scaling Laws.” Institute of Science and Technology Austria, 2024. https://doi.org/10.15479/at:ista:17485.
[Published Version] View | Files available | DOI
 
2024 | Published | Conference Paper | IST-REx-ID: 18061 | OA
Frantar, Elias, and Dan-Adrian Alistarh. “QMoE: Sub-1-Bit Compression of Trillion Parameter Models.” In Proceedings of Machine Learning and Systems, edited by P. Gibbons, G. Pekhimenko, and C. De Sa, Vol. 6, 2024.
[Published Version] View | Files available | Download Published Version (ext.)
 
2024 | Published | Conference Paper | IST-REx-ID: 18062 | OA
Frantar, Elias, Carlos Riquelme Ruiz, Neil Houlsby, Dan-Adrian Alistarh, and Utku Evci. “Scaling Laws for Sparsely-Connected Foundation Models.” In The Twelfth International Conference on Learning Representations, 2024.
[Published Version] View | Files available | Download Published Version (ext.) | arXiv
 
2024 | Published | Conference Paper | IST-REx-ID: 18070
Chatterjee, Bapi, Vyacheslav Kungurtsev, and Dan-Adrian Alistarh. “Federated SGD with Local Asynchrony.” In Proceedings of the 44th International Conference on Distributed Computing Systems, 857–68. IEEE, 2024. https://doi.org/10.1109/ICDCS60910.2024.00084.
View | DOI
 
2024 | Published | Conference Paper | IST-REx-ID: 18113 | OA
Egiazarian, Vage, Andrei Panferov, Denis Kuznedelev, Elias Frantar, Artem Babenko, and Dan-Adrian Alistarh. “Extreme Compression of Large Language Models via Additive Quantization.” In Proceedings of the 41st International Conference on Machine Learning, 235:12284–303. ML Research Press, 2024.
[Preprint] View | Download Preprint (ext.) | arXiv
 
2024 | Published | Conference Paper | IST-REx-ID: 18117 | OA
Nikdan, Mahdi, Soroush Tabesh, Elvir Crncevic, and Dan-Adrian Alistarh. “RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation.” In Proceedings of the 41st International Conference on Machine Learning, 235:38187–206. ML Research Press, 2024.
[Preprint] View | Files available | Download Preprint (ext.) | arXiv
 
2024 | Published | Conference Paper | IST-REx-ID: 18121 | OA
Moakhar, Arshia Soltani, Eugenia B Iofinova, Elias Frantar, and Dan-Adrian Alistarh. “SPADE: Sparsity-Guided Debugging for Deep Neural Networks.” In Proceedings of the 41st International Conference on Machine Learning, 235:45955–87. ML Research Press, 2024.
[Preprint] View | Files available | Download Preprint (ext.) | arXiv
 
2024 | Published | Conference Paper | IST-REx-ID: 15011 | OA
Kurtic, Eldar, Torsten Hoefler, and Dan-Adrian Alistarh. “How to Prune Your Language Model: Recovering Accuracy on the ‘Sparsity May Cry’ Benchmark.” In Proceedings of Machine Learning Research, 234:542–53. ML Research Press, 2024.
[Preprint] View | Download Preprint (ext.) | arXiv
 
2024 | Published | Conference Paper | IST-REx-ID: 17469 | OA
Kögler, Kevin, Alexander Shevchenko, Hamed Hassani, and Marco Mondelli. “Compression of Structured Data with Autoencoders: Provable Benefit of Nonlinearities and Depth.” In Proceedings of the 41st International Conference on Machine Learning, 235:24964–15. ML Research Press, 2024.
[Published Version] View | Files available | Download Published Version (ext.) | arXiv
 
2024 | Published | Thesis | IST-REx-ID: 17465 | OA
Shevchenko, Alexander. “High-Dimensional Limits in Artificial Neural Networks.” Institute of Science and Technology Austria, 2024. https://doi.org/10.15479/at:ista:17465.
[Published Version] View | Files available | DOI
 

Search

Filter Publications

Display / Sort

Citation Style: Chicago

Export / Embed