[{"ddc":["000"],"month":"11","quality_controlled":"1","user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","acknowledgement":"C P acknowledges funding from Astex through the Sustaining Innovation Program under the Milner Consortium. B C acknowledges resources provided by the Cambridge Tier-2 system operated by the University of Cambridge Research Computing Service funded by EPSRC Tier-2 capital Grant EP/P020259/1. F A F acknowledges funding from the Swiss National Science Foundation (Grant No. P2BSP2_191736). ","_id":"12128","volume":3,"title":"BenchML: An extensible pipelining framework for benchmarking representations of materials and molecules at scale","status":"public","tmp":{"name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","image":"/images/cc_by.png","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","short":"CC BY (4.0)"},"issue":"4","scopus_import":"1","publication_identifier":{"issn":["2632-2153"]},"keyword":["Artificial Intelligence","Human-Computer Interaction","Software"],"corr_author":"1","external_id":{"isi":["000886534000001"]},"has_accepted_license":"1","oa_version":"Published Version","year":"2022","author":[{"first_name":"Carl","full_name":"Poelking, Carl","last_name":"Poelking"},{"full_name":"Faber, Felix A","first_name":"Felix A","last_name":"Faber"},{"last_name":"Cheng","orcid":"0000-0002-3584-9632","full_name":"Cheng, Bingqing","first_name":"Bingqing","id":"cbe3cda4-d82c-11eb-8dc7-8ff94289fcc9"}],"article_processing_charge":"No","abstract":[{"text":"We introduce a machine-learning (ML) framework for high-throughput benchmarking of diverse representations of chemical systems against datasets of materials and molecules. The guiding principle underlying the benchmarking approach is to evaluate raw descriptor performance by limiting model complexity to simple regression schemes while enforcing best ML practices, allowing for unbiased hyperparameter optimization, and assessing learning progress through learning curves along series of synchronized train-test splits. The resulting models are intended as baselines that can inform future method development, in addition to indicating how easily a given dataset can be learnt. Through a comparative analysis of the training outcome across a diverse set of physicochemical, topological and geometric representations, we glean insight into the relative merits of these representations as well as their interrelatedness.","lang":"eng"}],"date_updated":"2024-10-09T21:03:32Z","license":"https://creativecommons.org/licenses/by/4.0/","article_type":"original","related_material":{"link":[{"relation":"software","url":"https://github.com/capoe/benchml"}]},"article_number":"040501","publication_status":"published","day":"17","oa":1,"date_published":"2022-11-17T00:00:00Z","doi":"10.1088/2632-2153/ac4d11","file_date_updated":"2023-01-23T10:42:04Z","publication":"Machine Learning: Science and Technology","language":[{"iso":"eng"}],"file":[{"checksum":"8930d4ad6ed9b47358c6f1a68666adb6","relation":"main_file","file_name":"2022_MachLearning_Poelking.pdf","content_type":"application/pdf","date_created":"2023-01-23T10:42:04Z","file_size":13814559,"date_updated":"2023-01-23T10:42:04Z","creator":"dernst","success":1,"access_level":"open_access","file_id":"12343"}],"date_created":"2023-01-12T12:02:21Z","isi":1,"intvolume":"         3","publisher":"IOP Publishing","type":"journal_article","department":[{"_id":"BiCh"}],"citation":{"ista":"Poelking C, Faber FA, Cheng B. 2022. BenchML: An extensible pipelining framework for benchmarking representations of materials and molecules at scale. Machine Learning: Science and Technology. 3(4), 040501.","mla":"Poelking, Carl, et al. “BenchML: An Extensible Pipelining Framework for Benchmarking Representations of Materials and Molecules at Scale.” <i>Machine Learning: Science and Technology</i>, vol. 3, no. 4, 040501, IOP Publishing, 2022, doi:<a href=\"https://doi.org/10.1088/2632-2153/ac4d11\">10.1088/2632-2153/ac4d11</a>.","ieee":"C. Poelking, F. A. Faber, and B. Cheng, “BenchML: An extensible pipelining framework for benchmarking representations of materials and molecules at scale,” <i>Machine Learning: Science and Technology</i>, vol. 3, no. 4. IOP Publishing, 2022.","chicago":"Poelking, Carl, Felix A Faber, and Bingqing Cheng. “BenchML: An Extensible Pipelining Framework for Benchmarking Representations of Materials and Molecules at Scale.” <i>Machine Learning: Science and Technology</i>. IOP Publishing, 2022. <a href=\"https://doi.org/10.1088/2632-2153/ac4d11\">https://doi.org/10.1088/2632-2153/ac4d11</a>.","short":"C. Poelking, F.A. Faber, B. Cheng, Machine Learning: Science and Technology 3 (2022).","ama":"Poelking C, Faber FA, Cheng B. BenchML: An extensible pipelining framework for benchmarking representations of materials and molecules at scale. <i>Machine Learning: Science and Technology</i>. 2022;3(4). doi:<a href=\"https://doi.org/10.1088/2632-2153/ac4d11\">10.1088/2632-2153/ac4d11</a>","apa":"Poelking, C., Faber, F. A., &#38; Cheng, B. (2022). BenchML: An extensible pipelining framework for benchmarking representations of materials and molecules at scale. <i>Machine Learning: Science and Technology</i>. IOP Publishing. <a href=\"https://doi.org/10.1088/2632-2153/ac4d11\">https://doi.org/10.1088/2632-2153/ac4d11</a>"}},{"quality_controlled":"1","month":"11","ddc":["000"],"status":"public","title":"Closed-form continuous-time neural networks","volume":4,"_id":"12147","acknowledgement":"This research was supported in part by the AI2050 program at Schmidt Futures (grant G-22-63172), the Boeing Company, and the United States Air Force Research Laboratory and the United States Air Force Artificial Intelligence Accelerator and was accomplished under cooperative agreement number FA8750-19-2-1000. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the United States Air Force or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes, notwithstanding any copyright notation herein. This work was further supported by The Boeing Company and Office of Naval Research grant N00014-18-1-2830. M.T. is supported by the Poul Due Jensen Foundation, grant 883901. M.L. was supported in part by the Austrian Science Fund under grant Z211-N23 (Wittgenstein Award). A.A. was supported by the National Science Foundation Graduate Research Fellowship Program. We thank T.-H. Wang, P. Kao, M. Chahine, W. Xiao, X. Li, L. Yin and Y. Ben for useful suggestions and for testing of CfC models to confirm the results across other domains.","user_id":"4359f0d1-fa6c-11eb-b949-802e58b17ae8","external_id":{"arxiv":["2106.13898"],"isi":["000884215600003"]},"publication_identifier":{"issn":["2522-5839"]},"keyword":["Artificial Intelligence","Computer Networks and Communications","Computer Vision and Pattern Recognition","Human-Computer Interaction","Software"],"issue":"11","scopus_import":"1","tmp":{"name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","image":"/images/cc_by.png","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode","short":"CC BY (4.0)"},"related_material":{"link":[{"relation":"erratum","url":"https://doi.org/10.1038/s42256-022-00597-y"}]},"page":"992-1003","article_type":"original","date_updated":"2025-04-15T06:26:02Z","abstract":[{"lang":"eng","text":"Continuous-time neural networks are a class of machine learning systems that can tackle representation learning on spatiotemporal decision-making tasks. These models are typically represented by continuous differential equations. However, their expressive power when they are deployed on computers is bottlenecked by numerical differential equation solvers. This limitation has notably slowed down the scaling and understanding of numerous natural physical phenomena such as the dynamics of nervous systems. Ideally, we would circumvent this bottleneck by solving the given dynamical system in closed form. This is known to be intractable in general. Here, we show that it is possible to closely approximate the interaction between neurons and synapses—the building blocks of natural and artificial neural networks—constructed by liquid time-constant networks efficiently in closed form. To this end, we compute a tightly bounded approximation of the solution of an integral appearing in liquid time-constant dynamics that has had no known closed-form solution so far. This closed-form solution impacts the design of continuous-time and continuous-depth neural models. For instance, since time appears explicitly in closed form, the formulation relaxes the need for complex numerical solvers. Consequently, we obtain models that are between one and five orders of magnitude faster in training and inference compared with differential equation-based counterparts. More importantly, in contrast to ordinary differential equation-based continuous networks, closed-form networks can scale remarkably well compared with other deep learning instances. Lastly, as these models are derived from liquid networks, they show good performance in time-series modelling compared with advanced recurrent neural network models."}],"year":"2022","article_processing_charge":"No","author":[{"first_name":"Ramin","full_name":"Hasani, Ramin","last_name":"Hasani"},{"full_name":"Lechner, Mathias","first_name":"Mathias","id":"3DC22916-F248-11E8-B48F-1D18A9856A87","last_name":"Lechner"},{"last_name":"Amini","full_name":"Amini, Alexander","first_name":"Alexander"},{"last_name":"Liebenwein","first_name":"Lucas","full_name":"Liebenwein, Lucas"},{"last_name":"Ray","full_name":"Ray, Aaron","first_name":"Aaron"},{"full_name":"Tschaikowski, Max","first_name":"Max","last_name":"Tschaikowski"},{"last_name":"Teschl","full_name":"Teschl, Gerald","first_name":"Gerald"},{"full_name":"Rus, Daniela","first_name":"Daniela","last_name":"Rus"}],"has_accepted_license":"1","arxiv":1,"oa_version":"Published Version","day":"15","publication_status":"published","file":[{"file_id":"12355","date_updated":"2023-01-24T09:49:44Z","success":1,"creator":"dernst","access_level":"open_access","file_size":3259553,"date_created":"2023-01-24T09:49:44Z","content_type":"application/pdf","checksum":"b4789122ce04bfb4ac042390f59aaa8b","relation":"main_file","file_name":"2022_NatureMachineIntelligence_Hasani.pdf"}],"language":[{"iso":"eng"}],"file_date_updated":"2023-01-24T09:49:44Z","project":[{"grant_number":"Z211","_id":"25F42A32-B435-11E9-9278-68D0E5697425","name":"Formal methods for the design and analysis of complex systems","call_identifier":"FWF"}],"publication":"Nature Machine Intelligence","doi":"10.1038/s42256-022-00556-7","oa":1,"date_published":"2022-11-15T00:00:00Z","date_created":"2023-01-12T12:07:21Z","citation":{"short":"R. Hasani, M. Lechner, A. Amini, L. Liebenwein, A. Ray, M. Tschaikowski, G. Teschl, D. Rus, Nature Machine Intelligence 4 (2022) 992–1003.","ama":"Hasani R, Lechner M, Amini A, et al. Closed-form continuous-time neural networks. <i>Nature Machine Intelligence</i>. 2022;4(11):992-1003. doi:<a href=\"https://doi.org/10.1038/s42256-022-00556-7\">10.1038/s42256-022-00556-7</a>","apa":"Hasani, R., Lechner, M., Amini, A., Liebenwein, L., Ray, A., Tschaikowski, M., … Rus, D. (2022). Closed-form continuous-time neural networks. <i>Nature Machine Intelligence</i>. Springer Nature. <a href=\"https://doi.org/10.1038/s42256-022-00556-7\">https://doi.org/10.1038/s42256-022-00556-7</a>","ista":"Hasani R, Lechner M, Amini A, Liebenwein L, Ray A, Tschaikowski M, Teschl G, Rus D. 2022. Closed-form continuous-time neural networks. Nature Machine Intelligence. 4(11), 992–1003.","mla":"Hasani, Ramin, et al. “Closed-Form Continuous-Time Neural Networks.” <i>Nature Machine Intelligence</i>, vol. 4, no. 11, Springer Nature, 2022, pp. 992–1003, doi:<a href=\"https://doi.org/10.1038/s42256-022-00556-7\">10.1038/s42256-022-00556-7</a>.","ieee":"R. Hasani <i>et al.</i>, “Closed-form continuous-time neural networks,” <i>Nature Machine Intelligence</i>, vol. 4, no. 11. Springer Nature, pp. 992–1003, 2022.","chicago":"Hasani, Ramin, Mathias Lechner, Alexander Amini, Lucas Liebenwein, Aaron Ray, Max Tschaikowski, Gerald Teschl, and Daniela Rus. “Closed-Form Continuous-Time Neural Networks.” <i>Nature Machine Intelligence</i>. Springer Nature, 2022. <a href=\"https://doi.org/10.1038/s42256-022-00556-7\">https://doi.org/10.1038/s42256-022-00556-7</a>."},"type":"journal_article","department":[{"_id":"ToHe"}],"publisher":"Springer Nature","isi":1,"intvolume":"         4"}]
