[{"article_type":"original","date_published":"2020-08-18T00:00:00Z","oa_version":"None","publication_status":"published","year":"2020","citation":{"chicago":"Gunasekaran, Suman, Douglas A. Reed, Daniel W. Paley, Amymarie K. Bartholomew, Latha Venkataraman, Michael L. Steigerwald, Xavier Roy, and Colin Nuckolls. “Single-Electron Currents in Designer Single-Cluster Devices.” <i>Journal of the American Chemical Society</i>. American Chemical Society, 2020. <a href=\"https://doi.org/10.1021/jacs.0c04970\">https://doi.org/10.1021/jacs.0c04970</a>.","short":"S. Gunasekaran, D.A. Reed, D.W. Paley, A.K. Bartholomew, L. Venkataraman, M.L. Steigerwald, X. Roy, C. Nuckolls, Journal of the American Chemical Society 142 (2020) 14924–14932.","ista":"Gunasekaran S, Reed DA, Paley DW, Bartholomew AK, Venkataraman L, Steigerwald ML, Roy X, Nuckolls C. 2020. Single-electron currents in designer single-cluster devices. Journal of the American Chemical Society. 142(35), 14924–14932.","apa":"Gunasekaran, S., Reed, D. A., Paley, D. W., Bartholomew, A. K., Venkataraman, L., Steigerwald, M. L., … Nuckolls, C. (2020). Single-electron currents in designer single-cluster devices. <i>Journal of the American Chemical Society</i>. American Chemical Society. <a href=\"https://doi.org/10.1021/jacs.0c04970\">https://doi.org/10.1021/jacs.0c04970</a>","ieee":"S. Gunasekaran <i>et al.</i>, “Single-electron currents in designer single-cluster devices,” <i>Journal of the American Chemical Society</i>, vol. 142, no. 35. American Chemical Society, pp. 14924–14932, 2020.","mla":"Gunasekaran, Suman, et al. “Single-Electron Currents in Designer Single-Cluster Devices.” <i>Journal of the American Chemical Society</i>, vol. 142, no. 35, American Chemical Society, 2020, pp. 14924–32, doi:<a href=\"https://doi.org/10.1021/jacs.0c04970\">10.1021/jacs.0c04970</a>.","ama":"Gunasekaran S, Reed DA, Paley DW, et al. Single-electron currents in designer single-cluster devices. <i>Journal of the American Chemical Society</i>. 2020;142(35):14924-14932. doi:<a href=\"https://doi.org/10.1021/jacs.0c04970\">10.1021/jacs.0c04970</a>"},"issue":"35","month":"08","author":[{"full_name":"Gunasekaran, Suman","last_name":"Gunasekaran","first_name":"Suman"},{"first_name":"Douglas A.","last_name":"Reed","full_name":"Reed, Douglas A."},{"full_name":"Paley, Daniel W.","first_name":"Daniel W.","last_name":"Paley"},{"full_name":"Bartholomew, Amymarie K.","last_name":"Bartholomew","first_name":"Amymarie K."},{"id":"9ebb78a5-cc0d-11ee-8322-fae086a32caf","orcid":"0000-0002-6957-6089","last_name":"Venkataraman","first_name":"Latha","full_name":"Venkataraman, Latha"},{"full_name":"Steigerwald, Michael L.","first_name":"Michael L.","last_name":"Steigerwald"},{"full_name":"Roy, Xavier","last_name":"Roy","first_name":"Xavier"},{"last_name":"Nuckolls","first_name":"Colin","full_name":"Nuckolls, Colin"}],"extern":"1","_id":"17909","title":"Single-electron currents in designer single-cluster devices","publication_identifier":{"eissn":["1520-5126"],"issn":["0002-7863"]},"intvolume":"       142","quality_controlled":"1","language":[{"iso":"eng"}],"OA_type":"closed access","scopus_import":"1","abstract":[{"lang":"eng","text":"Atomically precise clusters can be used to create single-electron devices wherein a single redox-active cluster is connected to two macroscopic electrodes via anchoring ligands. Unlike single-electron devices comprising nanocrystals, these cluster-based devices can be fabricated with atomic precision. This affords an unprecedented level of control over the device properties. Herein, we design a series of cobalt chalcogenide clusters with varying ligand geometries and core nuclearities to control their current–voltage (I–V) characteristics in a scanning tunneling microscope-based break junction (STM-BJ) device. First, the device geometry is modified by precisely positioning junction-anchoring ligands on the surface of the cluster. We show that the I–V characteristics are independent of ligand placement, confirming a sequential, single-electron tunneling mechanism. Next, we chemically fuse two clusters to realize a larger cluster dimer that behaves as a single electronic unit, possessing a smaller reorganization energy and more accessible redox states than the monomeric analogues. As a result, dimer-based devices exhibit significantly higher currents and can even be pushed to current saturation at high bias. Owing to these controllable properties, single-cluster junctions serve as an excellent platform for exploring incoherent charge transport processes at the nanoscale. With this understanding, as well as properties such as nonlinear I–V characteristics and rectification, these molecular clusters may function as conductive inorganic nodes in new devices and materials."}],"article_processing_charge":"No","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","external_id":{"pmid":["32809814"]},"doi":"10.1021/jacs.0c04970","volume":142,"pmid":1,"page":"14924-14932","publication":"Journal of the American Chemical Society","publisher":"American Chemical Society","date_updated":"2024-12-10T12:04:31Z","day":"18","type":"journal_article","date_created":"2024-09-09T07:19:56Z","status":"public"},{"quality_controlled":"1","intvolume":"        20","publication_identifier":{"issn":["1530-6984"],"eissn":["1530-6992"]},"title":"Using deep learning to identify molecular junction characteristics","_id":"17910","extern":"1","author":[{"last_name":"Fu","first_name":"Tianren","full_name":"Fu, Tianren"},{"full_name":"Zang, Yaping","first_name":"Yaping","last_name":"Zang"},{"first_name":"Qi","last_name":"Zou","full_name":"Zou, Qi"},{"full_name":"Nuckolls, Colin","first_name":"Colin","last_name":"Nuckolls"},{"orcid":"0000-0002-6957-6089","id":"9ebb78a5-cc0d-11ee-8322-fae086a32caf","full_name":"Venkataraman, Latha","first_name":"Latha","last_name":"Venkataraman"}],"month":"04","issue":"5","citation":{"ama":"Fu T, Zang Y, Zou Q, Nuckolls C, Venkataraman L. Using deep learning to identify molecular junction characteristics. <i>Nano Letters</i>. 2020;20(5):3320-3325. doi:<a href=\"https://doi.org/10.1021/acs.nanolett.0c00198\">10.1021/acs.nanolett.0c00198</a>","apa":"Fu, T., Zang, Y., Zou, Q., Nuckolls, C., &#38; Venkataraman, L. (2020). Using deep learning to identify molecular junction characteristics. <i>Nano Letters</i>. American Chemical Society. <a href=\"https://doi.org/10.1021/acs.nanolett.0c00198\">https://doi.org/10.1021/acs.nanolett.0c00198</a>","ieee":"T. Fu, Y. Zang, Q. Zou, C. Nuckolls, and L. Venkataraman, “Using deep learning to identify molecular junction characteristics,” <i>Nano Letters</i>, vol. 20, no. 5. American Chemical Society, pp. 3320–3325, 2020.","mla":"Fu, Tianren, et al. “Using Deep Learning to Identify Molecular Junction Characteristics.” <i>Nano Letters</i>, vol. 20, no. 5, American Chemical Society, 2020, pp. 3320–25, doi:<a href=\"https://doi.org/10.1021/acs.nanolett.0c00198\">10.1021/acs.nanolett.0c00198</a>.","short":"T. Fu, Y. Zang, Q. Zou, C. Nuckolls, L. Venkataraman, Nano Letters 20 (2020) 3320–3325.","chicago":"Fu, Tianren, Yaping Zang, Qi Zou, Colin Nuckolls, and Latha Venkataraman. “Using Deep Learning to Identify Molecular Junction Characteristics.” <i>Nano Letters</i>. American Chemical Society, 2020. <a href=\"https://doi.org/10.1021/acs.nanolett.0c00198\">https://doi.org/10.1021/acs.nanolett.0c00198</a>.","ista":"Fu T, Zang Y, Zou Q, Nuckolls C, Venkataraman L. 2020. Using deep learning to identify molecular junction characteristics. Nano Letters. 20(5), 3320–3325."},"year":"2020","publication_status":"published","oa_version":"None","date_published":"2020-04-03T00:00:00Z","article_type":"letter_note","status":"public","date_created":"2024-09-09T07:20:52Z","type":"journal_article","day":"03","date_updated":"2024-12-10T12:08:53Z","publisher":"American Chemical Society","publication":"Nano Letters","page":"3320-3325","pmid":1,"volume":20,"doi":"10.1021/acs.nanolett.0c00198","external_id":{"pmid":["32242671"]},"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_processing_charge":"No","abstract":[{"lang":"eng","text":"The scanning tunneling microscope-based break junction (STM-BJ) is used widely to create and characterize single metal-molecule-metal junctions. In this technique, conductance is continuously recorded as a metal point contact is broken in a solution of molecules. Conductance plateaus are seen when stable molecular junctions are formed. Typically, thousands of junctions are created and measured, yielding thousands of distinct conductance versus extension traces. However, such traces are rarely analyzed individually to recognize the types of junctions formed. Here, we present a deep learning-based method to identify molecular junctions and show that it performs better than several commonly used and recently reported techniques. We demonstrate molecular junction identification from mixed solution measurements with accuracies as high as 97%. We also apply this model to an in situ electric field-driven isomerization reaction of a [3]cumulene to follow the reaction over time. Furthermore, we demonstrate that our model can remain accurate even when a key parameter, the average junction conductance, is eliminated from the analysis, showing that our model goes beyond conventional analysis in existing methods."}],"scopus_import":"1","language":[{"iso":"eng"}]},{"extern":"1","main_file_link":[{"open_access":"1","url":"https://doi.org/10.1039/D0NR00467G"}],"oa":1,"_id":"17911","title":"Unsupervised feature recognition in single-molecule break junction data","publication_identifier":{"issn":["2040-3364"],"eissn":["2040-3372"]},"intvolume":"        12","quality_controlled":"1","article_type":"original","OA_place":"publisher","date_published":"2020-03-25T00:00:00Z","license":"https://creativecommons.org/licenses/by-nc/3.0/","publication_status":"published","oa_version":"Published Version","year":"2020","citation":{"ista":"Magyarkuti A, Balogh N, Balogh Z, Venkataraman L, Halbritter A. 2020. Unsupervised feature recognition in single-molecule break junction data. Nanoscale. 12(15), 8355–8363.","chicago":"Magyarkuti, András, Nóra Balogh, Zoltán Balogh, Latha Venkataraman, and András Halbritter. “Unsupervised Feature Recognition in Single-Molecule Break Junction Data.” <i>Nanoscale</i>. Royal Society of Chemistry, 2020. <a href=\"https://doi.org/10.1039/d0nr00467g\">https://doi.org/10.1039/d0nr00467g</a>.","short":"A. Magyarkuti, N. Balogh, Z. Balogh, L. Venkataraman, A. Halbritter, Nanoscale 12 (2020) 8355–8363.","mla":"Magyarkuti, András, et al. “Unsupervised Feature Recognition in Single-Molecule Break Junction Data.” <i>Nanoscale</i>, vol. 12, no. 15, Royal Society of Chemistry, 2020, pp. 8355–63, doi:<a href=\"https://doi.org/10.1039/d0nr00467g\">10.1039/d0nr00467g</a>.","ieee":"A. Magyarkuti, N. Balogh, Z. Balogh, L. Venkataraman, and A. Halbritter, “Unsupervised feature recognition in single-molecule break junction data,” <i>Nanoscale</i>, vol. 12, no. 15. Royal Society of Chemistry, pp. 8355–8363, 2020.","apa":"Magyarkuti, A., Balogh, N., Balogh, Z., Venkataraman, L., &#38; Halbritter, A. (2020). Unsupervised feature recognition in single-molecule break junction data. <i>Nanoscale</i>. Royal Society of Chemistry. <a href=\"https://doi.org/10.1039/d0nr00467g\">https://doi.org/10.1039/d0nr00467g</a>","ama":"Magyarkuti A, Balogh N, Balogh Z, Venkataraman L, Halbritter A. Unsupervised feature recognition in single-molecule break junction data. <i>Nanoscale</i>. 2020;12(15):8355-8363. doi:<a href=\"https://doi.org/10.1039/d0nr00467g\">10.1039/d0nr00467g</a>"},"issue":"15","month":"03","author":[{"full_name":"Magyarkuti, András","last_name":"Magyarkuti","first_name":"András"},{"full_name":"Balogh, Nóra","first_name":"Nóra","last_name":"Balogh"},{"full_name":"Balogh, Zoltán","first_name":"Zoltán","last_name":"Balogh"},{"full_name":"Venkataraman, Latha","last_name":"Venkataraman","first_name":"Latha","id":"9ebb78a5-cc0d-11ee-8322-fae086a32caf","orcid":"0000-0002-6957-6089"},{"last_name":"Halbritter","first_name":"András","full_name":"Halbritter, András"}],"page":"8355-8363","publication":"Nanoscale","publisher":"Royal Society of Chemistry","tmp":{"name":"Creative Commons Attribution-NonCommercial 3.0 Unported (CC BY-NC 3.0)","legal_code_url":"https://creativecommons.org/licenses/by-nc/3.0/legalcode","image":"/images/cc_by_nc.png","short":"CC BY-NC (3.0)"},"date_updated":"2024-12-10T12:13:16Z","arxiv":1,"day":"25","type":"journal_article","date_created":"2024-09-09T07:21:34Z","status":"public","OA_type":"hybrid","language":[{"iso":"eng"}],"scopus_import":"1","abstract":[{"lang":"eng","text":"Single-molecule break junction measurements deliver a huge number of conductance vs. electrode separation traces. During such measurements, the target molecules may bind to the electrodes in different geometries, and the evolution and rupture of the single-molecule junction may also follow distinct trajectories. The unraveling of the various typical trace classes is a prerequisite to the proper physical interpretation of the data. Here we exploit the efficient feature recognition properties of neural networks to automatically find the relevant trace classes. To eliminate the need for manually labeled training data we apply a combined method, which automatically selects training traces according to the extreme values of principal component projections or some auxiliary measured quantities. Then the network captures the features of these characteristic traces and generalizes its inference to the entire dataset. The use of a simple neural network structure also enables a direct insight into the decision-making mechanism. We demonstrate that this combined machine learning method is efficient in the unsupervised recognition of unobvious, but highly relevant trace classes within low and room temperature gold–4,4′ bipyridine–gold single-molecule break junction data."}],"article_processing_charge":"Yes","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","external_id":{"arxiv":["2001.03006"]},"doi":"10.1039/d0nr00467g","volume":12},{"year":"2020","oa_version":"None","publication_status":"published","date_published":"2020-03-26T00:00:00Z","article_type":"original","author":[{"first_name":"Rachel L.","last_name":"Starr","full_name":"Starr, Rachel L."},{"full_name":"Fu, Tianren","last_name":"Fu","first_name":"Tianren"},{"full_name":"Doud, Evan A.","first_name":"Evan A.","last_name":"Doud"},{"last_name":"Stone","first_name":"Ilana","full_name":"Stone, Ilana"},{"full_name":"Roy, Xavier","first_name":"Xavier","last_name":"Roy"},{"orcid":"0000-0002-6957-6089","id":"9ebb78a5-cc0d-11ee-8322-fae086a32caf","full_name":"Venkataraman, Latha","first_name":"Latha","last_name":"Venkataraman"}],"month":"03","issue":"15","citation":{"apa":"Starr, R. L., Fu, T., Doud, E. A., Stone, I., Roy, X., &#38; Venkataraman, L. (2020). Gold–carbon contacts from oxidative addition of aryl iodides. <i>Journal of the American Chemical Society</i>. American Chemical Society. <a href=\"https://doi.org/10.1021/jacs.0c01466\">https://doi.org/10.1021/jacs.0c01466</a>","ieee":"R. L. Starr, T. Fu, E. A. Doud, I. Stone, X. Roy, and L. Venkataraman, “Gold–carbon contacts from oxidative addition of aryl iodides,” <i>Journal of the American Chemical Society</i>, vol. 142, no. 15. American Chemical Society, pp. 7128–7133, 2020.","mla":"Starr, Rachel L., et al. “Gold–Carbon Contacts from Oxidative Addition of Aryl Iodides.” <i>Journal of the American Chemical Society</i>, vol. 142, no. 15, American Chemical Society, 2020, pp. 7128–33, doi:<a href=\"https://doi.org/10.1021/jacs.0c01466\">10.1021/jacs.0c01466</a>.","ama":"Starr RL, Fu T, Doud EA, Stone I, Roy X, Venkataraman L. Gold–carbon contacts from oxidative addition of aryl iodides. <i>Journal of the American Chemical Society</i>. 2020;142(15):7128-7133. doi:<a href=\"https://doi.org/10.1021/jacs.0c01466\">10.1021/jacs.0c01466</a>","short":"R.L. Starr, T. Fu, E.A. Doud, I. Stone, X. Roy, L. Venkataraman, Journal of the American Chemical Society 142 (2020) 7128–7133.","chicago":"Starr, Rachel L., Tianren Fu, Evan A. Doud, Ilana Stone, Xavier Roy, and Latha Venkataraman. “Gold–Carbon Contacts from Oxidative Addition of Aryl Iodides.” <i>Journal of the American Chemical Society</i>. American Chemical Society, 2020. <a href=\"https://doi.org/10.1021/jacs.0c01466\">https://doi.org/10.1021/jacs.0c01466</a>.","ista":"Starr RL, Fu T, Doud EA, Stone I, Roy X, Venkataraman L. 2020. Gold–carbon contacts from oxidative addition of aryl iodides. Journal of the American Chemical Society. 142(15), 7128–7133."},"extern":"1","quality_controlled":"1","intvolume":"       142","publication_identifier":{"issn":["0002-7863"],"eissn":["1520-5126"]},"title":"Gold–carbon contacts from oxidative addition of aryl iodides","_id":"17912","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_processing_charge":"No","abstract":[{"lang":"eng","text":"Aryl halides are ubiquitous functional groups in organic chemistry, yet despite their obvious appeal as surface-binding linkers and as precursors for controlled graphene nanoribbon synthesis, they have seldom been used as such in molecular electronics. The confusion regarding the bonding of aryl iodides to Au electrodes is a case in point, with ambiguous reports of both dative Au–I and covalent Au–C contacts. Here we form single-molecule junctions with a series of oligophenylene molecular wires terminated asymmetrically with iodine and thiomethyl to show that the dative Au–I contact has a lower conductance than the covalent Au–C interaction, which we propose occurs via an in situ oxidative addition reaction at the Au surface. Furthermore, we confirm the formation of the Au–C bond by measuring an analogous series of molecules prepared ex situ with the complex AuI(PPh3) in place of the iodide. Density functional theory-based transport calculations support our experimental observations that Au–C linkages have higher conductance than Au–I linkages. Finally, we demonstrate selective promotion of the Au–C bond formation by controlling the bias applied across the junction. In addition to establishing the different binding modes of aryl iodides, our results chart a path to actively controlling oxidative addition on an Au surface using an applied bias."}],"scopus_import":"1","language":[{"iso":"eng"}],"OA_type":"closed access","volume":142,"doi":"10.1021/jacs.0c01466","external_id":{"pmid":["32212683"]},"publisher":"American Chemical Society","publication":"Journal of the American Chemical Society","page":"7128-7133","pmid":1,"status":"public","date_created":"2024-09-09T07:22:26Z","type":"journal_article","day":"26","date_updated":"2024-12-10T12:20:47Z"},{"status":"public","date_created":"2024-09-09T07:36:41Z","date_updated":"2024-12-10T12:24:13Z","day":"06","type":"journal_article","publication":"Nano Letters","publisher":"American Chemical Society","pmid":1,"page":"2843-2848","volume":20,"external_id":{"pmid":["32142291"]},"doi":"10.1021/acs.nanolett.0c00605","abstract":[{"text":"Electron transport across a molecular junction is characterized by an energy-dependent transmission function. The transmission function accounts for electrons tunneling through multiple molecular orbitals (MOs) with different phases, which gives rise to quantum interference (QI) effects. Because the transmission function comprises both interfering and noninterfering effects, individual interferences between MOs cannot be deduced from the transmission function directly. Herein, we demonstrate how the transmission function can be deconstructed into its constituent interfering and noninterfering contributions for any model molecular junction. These contributions are arranged in a matrix and displayed pictorially as a QI map, which allows one to easily identify individual QI effects. Importantly, we show that exponential conductance decay with increasing oligomer length is primarily due to an increase in destructive QI. With an ability to “see” QI effects using the QI map, we find that QI is vital to all molecular-scale electron transport.","lang":"eng"}],"article_processing_charge":"No","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","OA_type":"closed access","language":[{"iso":"eng"}],"scopus_import":"1","intvolume":"        20","quality_controlled":"1","_id":"17913","title":"Visualizing quantum interference in molecular junctions","publication_identifier":{"eissn":["1530-6992"],"issn":["1530-6984"]},"extern":"1","author":[{"full_name":"Gunasekaran, Suman","first_name":"Suman","last_name":"Gunasekaran"},{"first_name":"Julia E.","last_name":"Greenwald","full_name":"Greenwald, Julia E."},{"full_name":"Venkataraman, Latha","last_name":"Venkataraman","first_name":"Latha","orcid":"0000-0002-6957-6089","id":"9ebb78a5-cc0d-11ee-8322-fae086a32caf"}],"citation":{"ista":"Gunasekaran S, Greenwald JE, Venkataraman L. 2020. Visualizing quantum interference in molecular junctions. Nano Letters. 20(4), 2843–2848.","chicago":"Gunasekaran, Suman, Julia E. Greenwald, and Latha Venkataraman. “Visualizing Quantum Interference in Molecular Junctions.” <i>Nano Letters</i>. American Chemical Society, 2020. <a href=\"https://doi.org/10.1021/acs.nanolett.0c00605\">https://doi.org/10.1021/acs.nanolett.0c00605</a>.","short":"S. Gunasekaran, J.E. Greenwald, L. Venkataraman, Nano Letters 20 (2020) 2843–2848.","ama":"Gunasekaran S, Greenwald JE, Venkataraman L. Visualizing quantum interference in molecular junctions. <i>Nano Letters</i>. 2020;20(4):2843-2848. doi:<a href=\"https://doi.org/10.1021/acs.nanolett.0c00605\">10.1021/acs.nanolett.0c00605</a>","mla":"Gunasekaran, Suman, et al. “Visualizing Quantum Interference in Molecular Junctions.” <i>Nano Letters</i>, vol. 20, no. 4, American Chemical Society, 2020, pp. 2843–48, doi:<a href=\"https://doi.org/10.1021/acs.nanolett.0c00605\">10.1021/acs.nanolett.0c00605</a>.","apa":"Gunasekaran, S., Greenwald, J. E., &#38; Venkataraman, L. (2020). Visualizing quantum interference in molecular junctions. <i>Nano Letters</i>. American Chemical Society. <a href=\"https://doi.org/10.1021/acs.nanolett.0c00605\">https://doi.org/10.1021/acs.nanolett.0c00605</a>","ieee":"S. Gunasekaran, J. E. Greenwald, and L. Venkataraman, “Visualizing quantum interference in molecular junctions,” <i>Nano Letters</i>, vol. 20, no. 4. American Chemical Society, pp. 2843–2848, 2020."},"issue":"4","month":"03","date_published":"2020-03-06T00:00:00Z","oa_version":"None","publication_status":"published","year":"2020","article_type":"letter_note"},{"author":[{"full_name":"Hernangómez-Pérez, Daniel","last_name":"Hernangómez-Pérez","first_name":"Daniel"},{"full_name":"Gunasekaran, Suman","first_name":"Suman","last_name":"Gunasekaran"},{"orcid":"0000-0002-6957-6089","id":"9ebb78a5-cc0d-11ee-8322-fae086a32caf","full_name":"Venkataraman, Latha","last_name":"Venkataraman","first_name":"Latha"},{"full_name":"Evers, Ferdinand","last_name":"Evers","first_name":"Ferdinand"}],"citation":{"mla":"Hernangómez-Pérez, Daniel, et al. “Solitonics with Polyacetylenes.” <i>Nano Letters</i>, vol. 20, no. 4, American Chemical Society, 2020, pp. 2615–19, doi:<a href=\"https://doi.org/10.1021/acs.nanolett.0c00136\">10.1021/acs.nanolett.0c00136</a>.","ieee":"D. Hernangómez-Pérez, S. Gunasekaran, L. Venkataraman, and F. Evers, “Solitonics with polyacetylenes,” <i>Nano Letters</i>, vol. 20, no. 4. American Chemical Society, pp. 2615–2619, 2020.","apa":"Hernangómez-Pérez, D., Gunasekaran, S., Venkataraman, L., &#38; Evers, F. (2020). Solitonics with polyacetylenes. <i>Nano Letters</i>. American Chemical Society. <a href=\"https://doi.org/10.1021/acs.nanolett.0c00136\">https://doi.org/10.1021/acs.nanolett.0c00136</a>","ama":"Hernangómez-Pérez D, Gunasekaran S, Venkataraman L, Evers F. Solitonics with polyacetylenes. <i>Nano Letters</i>. 2020;20(4):2615-2619. doi:<a href=\"https://doi.org/10.1021/acs.nanolett.0c00136\">10.1021/acs.nanolett.0c00136</a>","ista":"Hernangómez-Pérez D, Gunasekaran S, Venkataraman L, Evers F. 2020. Solitonics with polyacetylenes. Nano Letters. 20(4), 2615–2619.","chicago":"Hernangómez-Pérez, Daniel, Suman Gunasekaran, Latha Venkataraman, and Ferdinand Evers. “Solitonics with Polyacetylenes.” <i>Nano Letters</i>. American Chemical Society, 2020. <a href=\"https://doi.org/10.1021/acs.nanolett.0c00136\">https://doi.org/10.1021/acs.nanolett.0c00136</a>.","short":"D. Hernangómez-Pérez, S. Gunasekaran, L. Venkataraman, F. Evers, Nano Letters 20 (2020) 2615–2619."},"issue":"4","month":"03","publication_status":"published","oa_version":"None","year":"2020","date_published":"2020-03-03T00:00:00Z","article_type":"letter_note","quality_controlled":"1","intvolume":"        20","publication_identifier":{"eissn":["1530-6992"],"issn":["1530-6984"]},"_id":"17914","title":"Solitonics with polyacetylenes","extern":"1","volume":20,"external_id":{"pmid":["32125870"]},"doi":"10.1021/acs.nanolett.0c00136","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","abstract":[{"lang":"eng","text":"Polyacetylene molecular wires have attracted a long-standing interest for the past 40 years. From a fundamental perspective, there are two main reasons for the interest. First, polyacetylenes are a prime realization of a one-dimensional topological insulator. Second, long molecules support freely propagating topological domain-wall states, so-called “solitons,” which provide an early paradigm for spin-charge separation. Because of recent experimental developments, individual polyacetylene chains can now be synthesized on substrates. Motivated by this breakthrough, we here propose a novel way for chemically supported soliton design in these systems. We demonstrate how to control the soliton position and how to read it out via external means. Also, we show how extra soliton–antisoliton pairs arise when applying a moderate static electric field. We thus make a step toward functionality of electronic devices based on soliton manipulation, that is, “solitonics”."}],"article_processing_charge":"No","scopus_import":"1","OA_type":"closed access","language":[{"iso":"eng"}],"date_created":"2024-09-09T07:38:36Z","status":"public","day":"03","type":"journal_article","date_updated":"2024-12-10T12:26:43Z","publication":"Nano Letters","publisher":"American Chemical Society","page":"2615-2619","pmid":1},{"author":[{"last_name":"Repellin","first_name":"C.","full_name":"Repellin, C."},{"id":"b75b3f45-7995-11ef-9bfd-9a9cd02c3577","full_name":"Leonard, Julian","first_name":"Julian","last_name":"Leonard"},{"first_name":"N.","last_name":"Goldman","full_name":"Goldman, N."}],"month":"12","citation":{"ieee":"C. Repellin, J. Leonard, and N. Goldman, “Fractional Chern insulators of few bosons in a box: Hall plateaus from center-of-mass drifts and density profiles,” <i>Physical Review A</i>, vol. 102, no. 6. American Physical Society, 2020.","apa":"Repellin, C., Leonard, J., &#38; Goldman, N. (2020). Fractional Chern insulators of few bosons in a box: Hall plateaus from center-of-mass drifts and density profiles. <i>Physical Review A</i>. American Physical Society. <a href=\"https://doi.org/10.1103/physreva.102.063316\">https://doi.org/10.1103/physreva.102.063316</a>","mla":"Repellin, C., et al. “Fractional Chern Insulators of Few Bosons in a Box: Hall Plateaus from Center-of-Mass Drifts and Density Profiles.” <i>Physical Review A</i>, vol. 102, no. 6, 063316, American Physical Society, 2020, doi:<a href=\"https://doi.org/10.1103/physreva.102.063316\">10.1103/physreva.102.063316</a>.","ama":"Repellin C, Leonard J, Goldman N. Fractional Chern insulators of few bosons in a box: Hall plateaus from center-of-mass drifts and density profiles. <i>Physical Review A</i>. 2020;102(6). doi:<a href=\"https://doi.org/10.1103/physreva.102.063316\">10.1103/physreva.102.063316</a>","short":"C. Repellin, J. Leonard, N. Goldman, Physical Review A 102 (2020).","chicago":"Repellin, C., Julian Leonard, and N. Goldman. “Fractional Chern Insulators of Few Bosons in a Box: Hall Plateaus from Center-of-Mass Drifts and Density Profiles.” <i>Physical Review A</i>. American Physical Society, 2020. <a href=\"https://doi.org/10.1103/physreva.102.063316\">https://doi.org/10.1103/physreva.102.063316</a>.","ista":"Repellin C, Leonard J, Goldman N. 2020. Fractional Chern insulators of few bosons in a box: Hall plateaus from center-of-mass drifts and density profiles. Physical Review A. 102(6), 063316."},"issue":"6","date_published":"2020-12-14T00:00:00Z","year":"2020","publication_status":"published","license":"https://creativecommons.org/licenses/by/4.0/","oa_version":"Published Version","article_type":"original","intvolume":"       102","quality_controlled":"1","title":"Fractional Chern insulators of few bosons in a box: Hall plateaus from center-of-mass drifts and density profiles","_id":"18194","publication_identifier":{"eissn":["2469-9934"],"issn":["2469-9926"]},"oa":1,"has_accepted_license":"1","ddc":["530"],"extern":"1","main_file_link":[{"open_access":"1","url":"https://doi.org/10.1103/PhysRevA.102.063316"}],"volume":102,"doi":"10.1103/physreva.102.063316","external_id":{"arxiv":["2005.09689"]},"article_processing_charge":"Yes (in subscription journal)","abstract":[{"lang":"eng","text":"Realizing strongly correlated topological phases of ultracold gases is a central goal for ongoing experiments. While fractional quantum Hall states could soon be implemented in small atomic ensembles, detecting their signatures in few-particle settings remains a fundamental challenge. In this work, we numerically analyze the center-of-mass Hall drift of a small ensemble of hardcore bosons, initially prepared in the ground state of the Harper-Hofstadter-Hubbard model in a box potential. By monitoring the Hall drift upon release, for a wide range of magnetic flux values, we identify an emergent Hall plateau compatible with a fractional Chern insulator state: The extracted Hall conductivity approaches a fractional value determined by the many-body Chern number, while the width of the plateau agrees with the spectral and topological properties of the prepared ground state. Besides, a direct application of Streda's formula indicates that such Hall plateaus can also be directly obtained from static density-profile measurements. Our calculations suggest that fractional Chern insulators can be detected in cold-atom experiments, using available detection methods."}],"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","language":[{"iso":"eng"}],"article_number":"063316","scopus_import":"1","date_created":"2024-10-07T11:48:07Z","status":"public","arxiv":1,"date_updated":"2024-10-08T09:51:57Z","type":"journal_article","day":"14","tmp":{"short":"CC BY (4.0)","image":"/images/cc_by.png","name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode"},"publisher":"American Physical Society","publication":"Physical Review A"},{"scopus_import":"1","language":[{"iso":"eng"}],"OA_type":"green","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","abstract":[{"lang":"eng","text":"We introduce two constructions in geometric deep learning for 1) transporting orientation-dependent convolutional filters over a manifold in a continuous way and thereby defining a convolution operator that naturally incorporates the rotational effect of holonomy; and 2) allowing efficient evaluation of manifold convolution layers by sampling manifold valued random variables that center around a weighted diffusion mean. Both methods are inspired by stochastics on manifolds and geometric statistics, and provide examples of how stochastic methods – here horizontal frame bundle flows and non-linear bridge sampling schemes, can be used in geometric deep learning. We outline the theoretical foundation of the two methods, discuss their relation to Euclidean deep networks and existing methodology in geometric deep learning, and establish important properties of the proposed constructions."}],"article_processing_charge":"No","external_id":{"arxiv":["1909.06397"]},"doi":"10.1109/tpami.2020.2994507","volume":44,"page":"811-822","publication":"IEEE Transactions on Pattern Analysis and Machine Intelligence","publisher":"Institute of Electrical and Electronics Engineers","day":"01","type":"journal_article","date_updated":"2024-10-15T06:56:47Z","arxiv":1,"status":"public","date_created":"2024-10-08T12:55:23Z","article_type":"original","oa_version":"Preprint","publication_status":"published","year":"2020","OA_place":"repository","date_published":"2020-02-01T00:00:00Z","issue":"2","citation":{"ista":"Sommer S, Bronstein AM. 2020. Horizontal flows and manifold stochastics in geometric deep learning. IEEE Transactions on Pattern Analysis and Machine Intelligence. 44(2), 811–822.","chicago":"Sommer, Stefan, and Alex M. Bronstein. “Horizontal Flows and Manifold Stochastics in Geometric Deep Learning.” <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>. Institute of Electrical and Electronics Engineers, 2020. <a href=\"https://doi.org/10.1109/tpami.2020.2994507\">https://doi.org/10.1109/tpami.2020.2994507</a>.","short":"S. Sommer, A.M. Bronstein, IEEE Transactions on Pattern Analysis and Machine Intelligence 44 (2020) 811–822.","mla":"Sommer, Stefan, and Alex M. Bronstein. “Horizontal Flows and Manifold Stochastics in Geometric Deep Learning.” <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>, vol. 44, no. 2, Institute of Electrical and Electronics Engineers, 2020, pp. 811–22, doi:<a href=\"https://doi.org/10.1109/tpami.2020.2994507\">10.1109/tpami.2020.2994507</a>.","apa":"Sommer, S., &#38; Bronstein, A. M. (2020). Horizontal flows and manifold stochastics in geometric deep learning. <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>. Institute of Electrical and Electronics Engineers. <a href=\"https://doi.org/10.1109/tpami.2020.2994507\">https://doi.org/10.1109/tpami.2020.2994507</a>","ieee":"S. Sommer and A. M. Bronstein, “Horizontal flows and manifold stochastics in geometric deep learning,” <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>, vol. 44, no. 2. Institute of Electrical and Electronics Engineers, pp. 811–822, 2020.","ama":"Sommer S, Bronstein AM. Horizontal flows and manifold stochastics in geometric deep learning. <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>. 2020;44(2):811-822. doi:<a href=\"https://doi.org/10.1109/tpami.2020.2994507\">10.1109/tpami.2020.2994507</a>"},"month":"02","author":[{"last_name":"Sommer","first_name":"Stefan","full_name":"Sommer, Stefan"},{"orcid":"0000-0001-9699-8730","id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","last_name":"Bronstein","first_name":"Alexander","full_name":"Bronstein, Alexander"}],"main_file_link":[{"url":"https://doi.org/10.48550/arXiv.1909.06397","open_access":"1"}],"extern":"1","oa":1,"publication_identifier":{"eissn":["1939-3539"],"issn":["0162-8828"]},"_id":"18228","title":"Horizontal flows and manifold stochastics in geometric deep learning","quality_controlled":"1","intvolume":"        44"},{"abstract":[{"lang":"eng","text":"Intel® RealSense™ SR300 is a depth camera capable of providing a VGA-size depth map at 60 fps and 0.125mm depth resolution. In addition, it outputs an infrared VGA-resolution image and a 1080p color texture image at 30 fps. SR300 form-factor enables it to be integrated into small consumer products and as a front facing camera in laptops and Ultrabooks™. The SR300 depth camera is based on a coded-light technology where triangulation between projected patterns and images captured by a dedicated sensor is used to produce the depth map. Each projected line is coded by a special temporal optical code, that enables a dense depth map reconstruction from its reflection. The solid mechanical assembly of the camera allows it to stay calibrated throughout temperature and pressure changes, drops, and hits. In addition, active dynamic control maintains a calibrated depth output. An extended API LibRS released with the camera allows developers to integrate the camera in various applications. Algorithms for 3D scanning, facial analysis, hand gesture recognition, and tracking are within reach for applications using the SR300. In this paper, we describe the underlying technology, hardware, and algorithms of the SR300, as well as its calibration procedure, and outline some use cases. We believe that this paper will provide a full case study of a mass-produced depth sensing product and technology."}],"article_processing_charge":"No","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","language":[{"iso":"eng"}],"OA_type":"closed access","scopus_import":"1","volume":42,"external_id":{"pmid":["31094683"]},"doi":"10.1109/tpami.2019.2915841","publication":"IEEE Transactions on Pattern Analysis and Machine Intelligence","publisher":"Institute of Electrical and Electronics Engineers","pmid":1,"page":"2333-2345","status":"public","date_created":"2024-10-08T13:04:18Z","date_updated":"2024-10-15T09:40:01Z","day":"01","type":"journal_article","date_published":"2020-10-01T00:00:00Z","publication_status":"published","oa_version":"None","year":"2020","article_type":"original","author":[{"full_name":"Zabatani, Aviad","first_name":"Aviad","last_name":"Zabatani"},{"full_name":"Surazhsky, Vitaly","last_name":"Surazhsky","first_name":"Vitaly"},{"full_name":"Sperling, Erez","first_name":"Erez","last_name":"Sperling"},{"last_name":"Moshe","first_name":"Sagi Ben","full_name":"Moshe, Sagi Ben"},{"last_name":"Menashe","first_name":"Ohad","full_name":"Menashe, Ohad"},{"full_name":"Silver, David H.","first_name":"David H.","last_name":"Silver"},{"last_name":"Karni","first_name":"Zachi","full_name":"Karni, Zachi"},{"id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730","last_name":"Bronstein","first_name":"Alexander","full_name":"Bronstein, Alexander"},{"full_name":"Bronstein, Michael K","first_name":"Michael K","last_name":"Bronstein"},{"first_name":"Ron","last_name":"Kimmel","full_name":"Kimmel, Ron"}],"issue":"10","citation":{"ista":"Zabatani A, Surazhsky V, Sperling E, Moshe SB, Menashe O, Silver DH, Karni Z, Bronstein AM, Bronstein MK, Kimmel R. 2020. Intel® RealSense<sup>TM</sup> SR300 coded light depth camera. IEEE Transactions on Pattern Analysis and Machine Intelligence. 42(10), 2333–2345.","chicago":"Zabatani, Aviad, Vitaly Surazhsky, Erez Sperling, Sagi Ben Moshe, Ohad Menashe, David H. Silver, Zachi Karni, Alex M. Bronstein, Michael K Bronstein, and Ron Kimmel. “Intel® RealSense<sup>TM</sup> SR300 Coded Light Depth Camera.” <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>. Institute of Electrical and Electronics Engineers, 2020. <a href=\"https://doi.org/10.1109/tpami.2019.2915841\">https://doi.org/10.1109/tpami.2019.2915841</a>.","short":"A. Zabatani, V. Surazhsky, E. Sperling, S.B. Moshe, O. Menashe, D.H. Silver, Z. Karni, A.M. Bronstein, M.K. Bronstein, R. Kimmel, IEEE Transactions on Pattern Analysis and Machine Intelligence 42 (2020) 2333–2345.","mla":"Zabatani, Aviad, et al. “Intel® RealSense<sup>TM</sup> SR300 Coded Light Depth Camera.” <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>, vol. 42, no. 10, Institute of Electrical and Electronics Engineers, 2020, pp. 2333–45, doi:<a href=\"https://doi.org/10.1109/tpami.2019.2915841\">10.1109/tpami.2019.2915841</a>.","ieee":"A. Zabatani <i>et al.</i>, “Intel® RealSense<sup>TM</sup> SR300 coded light depth camera,” <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>, vol. 42, no. 10. Institute of Electrical and Electronics Engineers, pp. 2333–2345, 2020.","apa":"Zabatani, A., Surazhsky, V., Sperling, E., Moshe, S. B., Menashe, O., Silver, D. H., … Kimmel, R. (2020). Intel® RealSense<sup>TM</sup> SR300 coded light depth camera. <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>. Institute of Electrical and Electronics Engineers. <a href=\"https://doi.org/10.1109/tpami.2019.2915841\">https://doi.org/10.1109/tpami.2019.2915841</a>","ama":"Zabatani A, Surazhsky V, Sperling E, et al. Intel® RealSense<sup>TM</sup> SR300 coded light depth camera. <i>IEEE Transactions on Pattern Analysis and Machine Intelligence</i>. 2020;42(10):2333-2345. doi:<a href=\"https://doi.org/10.1109/tpami.2019.2915841\">10.1109/tpami.2019.2915841</a>"},"month":"10","extern":"1","intvolume":"        42","quality_controlled":"1","_id":"18245","title":"Intel® RealSense™ SR300 coded light depth camera","publication_identifier":{"eissn":["1939-3539"],"issn":["0162-8828"]}},{"volume":39,"doi":"10.1111/cgf.14083","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_processing_charge":"No","abstract":[{"lang":"eng","text":"Adversarial attacks have demonstrated remarkable efficacy in altering the output of a learning model by applying a minimal perturbation to the input data. While increasing attention has been placed on the image domain, however, the study of adversarial perturbations for geometric data has been notably lagging behind. In this paper, we show that effective adversarial attacks can be concocted for surfaces embedded in 3D, under weak smoothness assumptions on the perceptibility of the attack. We address the case of deformable 3D shapes in particular, and introduce a general model that is not tailored to any specific surface representation, nor does it assume access to a parametric description of the 3D object. In this context, we consider targeted and untargeted variants of the attack, demonstrating compelling results in either case. We further show how discovering adversarial examples, and then using them for adversarial training, leads to an increase in both robustness and accuracy. Our findings are confirmed empirically over multiple datasets spanning different semantic classes and deformations."}],"scopus_import":"1","OA_type":"closed access","language":[{"iso":"eng"}],"date_created":"2024-10-08T13:04:35Z","status":"public","type":"journal_article","day":"01","date_updated":"2024-10-15T09:36:46Z","publisher":"Wiley","publication":"Computer Graphics Forum","page":"253-264","author":[{"full_name":"Mariani, G.","last_name":"Mariani","first_name":"G."},{"full_name":"Cosmo, L.","first_name":"L.","last_name":"Cosmo"},{"id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730","full_name":"Bronstein, Alexander","first_name":"Alexander","last_name":"Bronstein"},{"full_name":"Rodolà, E.","last_name":"Rodolà","first_name":"E."}],"month":"08","issue":"5","citation":{"short":"G. Mariani, L. Cosmo, A.M. Bronstein, E. Rodolà, Computer Graphics Forum 39 (2020) 253–264.","chicago":"Mariani, G., L. Cosmo, Alex M. Bronstein, and E. Rodolà. “Generating Adversarial Surfaces via Band‐limited Perturbations.” <i>Computer Graphics Forum</i>. Wiley, 2020. <a href=\"https://doi.org/10.1111/cgf.14083\">https://doi.org/10.1111/cgf.14083</a>.","ista":"Mariani G, Cosmo L, Bronstein AM, Rodolà E. 2020. Generating adversarial surfaces via band‐limited perturbations. Computer Graphics Forum. 39(5), 253–264.","ama":"Mariani G, Cosmo L, Bronstein AM, Rodolà E. Generating adversarial surfaces via band‐limited perturbations. <i>Computer Graphics Forum</i>. 2020;39(5):253-264. doi:<a href=\"https://doi.org/10.1111/cgf.14083\">10.1111/cgf.14083</a>","ieee":"G. Mariani, L. Cosmo, A. M. Bronstein, and E. Rodolà, “Generating adversarial surfaces via band‐limited perturbations,” <i>Computer Graphics Forum</i>, vol. 39, no. 5. Wiley, pp. 253–264, 2020.","apa":"Mariani, G., Cosmo, L., Bronstein, A. M., &#38; Rodolà, E. (2020). Generating adversarial surfaces via band‐limited perturbations. <i>Computer Graphics Forum</i>. Wiley. <a href=\"https://doi.org/10.1111/cgf.14083\">https://doi.org/10.1111/cgf.14083</a>","mla":"Mariani, G., et al. “Generating Adversarial Surfaces via Band‐limited Perturbations.” <i>Computer Graphics Forum</i>, vol. 39, no. 5, Wiley, 2020, pp. 253–64, doi:<a href=\"https://doi.org/10.1111/cgf.14083\">10.1111/cgf.14083</a>."},"year":"2020","oa_version":"None","publication_status":"published","date_published":"2020-08-01T00:00:00Z","article_type":"original","quality_controlled":"1","intvolume":"        39","publication_identifier":{"issn":["0167-7055"],"eissn":["1467-8659"]},"title":"Generating adversarial surfaces via band‐limited perturbations","_id":"18246","extern":"1"},{"language":[{"iso":"eng"}],"article_number":"9206968","scopus_import":"1","article_processing_charge":"No","abstract":[{"text":"Convolutional neural networks (CNNs) achieve state-of-the-art accuracy in a variety of tasks in computer vision and beyond. One of the major obstacles hindering the ubiquitous use of CNNs for inference on low-power edge devices is their high computational complexity and memory bandwidth requirements. The latter often dominates the energy footprint on modern hardware. In this paper, we introduce a lossy transform coding approach, inspired by image and video compression, designed to reduce the memory bandwidth due to the storage of intermediate activation calculation results. Our method does not require fine-tuning the network weights and halves the data transfer volumes to the main memory by compressing feature maps, which are highly correlated, with variable length coding. Our method outperform previous approach in term of the number of bits per value with minor accuracy degradation on ResNet-34 and MobileNetV2. We analyze the performance of our approach on a variety of CNN architectures and demonstrate that FPGA implementation of ResNet-18 with our approach results in a reduction of around 40% in the memory energy footprint, compared to quantized network, with negligible impact on accuracy. When allowing accuracy degradation of up to 2%, the reduction of 60% is achieved. A reference implementation accompanies the paper.","lang":"eng"}],"user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","doi":"10.1109/ijcnn48605.2020.9206968","external_id":{"arxiv":["1905.10830"]},"publisher":"IEEE","publication":"2020 International Joint Conference on Neural Networks (IJCNN)","arxiv":1,"date_updated":"2024-12-12T10:04:54Z","type":"conference","day":"28","date_created":"2024-10-08T13:04:52Z","status":"public","date_published":"2020-09-28T00:00:00Z","year":"2020","oa_version":"Preprint","publication_status":"published","month":"09","citation":{"ama":"Chmiel B, Baskin C, Zheltonozhskii E, et al. Feature map transform coding for energy-efficient CNN inference. In: <i>2020 International Joint Conference on Neural Networks (IJCNN)</i>. IEEE; 2020. doi:<a href=\"https://doi.org/10.1109/ijcnn48605.2020.9206968\">10.1109/ijcnn48605.2020.9206968</a>","mla":"Chmiel, Brian, et al. “Feature Map Transform Coding for Energy-Efficient CNN Inference.” <i>2020 International Joint Conference on Neural Networks (IJCNN)</i>, 9206968, IEEE, 2020, doi:<a href=\"https://doi.org/10.1109/ijcnn48605.2020.9206968\">10.1109/ijcnn48605.2020.9206968</a>.","apa":"Chmiel, B., Baskin, C., Zheltonozhskii, E., Banner, R., Yermolin, Y., Karbachevsky, A., … Mendelson, A. (2020). Feature map transform coding for energy-efficient CNN inference. In <i>2020 International Joint Conference on Neural Networks (IJCNN)</i>. Glasgow, United Kingdom: IEEE. <a href=\"https://doi.org/10.1109/ijcnn48605.2020.9206968\">https://doi.org/10.1109/ijcnn48605.2020.9206968</a>","ieee":"B. Chmiel <i>et al.</i>, “Feature map transform coding for energy-efficient CNN inference,” in <i>2020 International Joint Conference on Neural Networks (IJCNN)</i>, Glasgow, United Kingdom, 2020.","ista":"Chmiel B, Baskin C, Zheltonozhskii E, Banner R, Yermolin Y, Karbachevsky A, Bronstein AM, Mendelson A. 2020. Feature map transform coding for energy-efficient CNN inference. 2020 International Joint Conference on Neural Networks (IJCNN). International Joint Conference on Neural Networks, 9206968.","chicago":"Chmiel, Brian, Chaim Baskin, Evgenii Zheltonozhskii, Ron Banner, Yevgeny Yermolin, Alex Karbachevsky, Alex M. Bronstein, and Avi Mendelson. “Feature Map Transform Coding for Energy-Efficient CNN Inference.” In <i>2020 International Joint Conference on Neural Networks (IJCNN)</i>. IEEE, 2020. <a href=\"https://doi.org/10.1109/ijcnn48605.2020.9206968\">https://doi.org/10.1109/ijcnn48605.2020.9206968</a>.","short":"B. Chmiel, C. Baskin, E. Zheltonozhskii, R. Banner, Y. Yermolin, A. Karbachevsky, A.M. Bronstein, A. Mendelson, in:, 2020 International Joint Conference on Neural Networks (IJCNN), IEEE, 2020."},"author":[{"last_name":"Chmiel","first_name":"Brian","full_name":"Chmiel, Brian"},{"full_name":"Baskin, Chaim","first_name":"Chaim","last_name":"Baskin"},{"full_name":"Zheltonozhskii, Evgenii","last_name":"Zheltonozhskii","first_name":"Evgenii"},{"last_name":"Banner","first_name":"Ron","full_name":"Banner, Ron"},{"first_name":"Yevgeny","last_name":"Yermolin","full_name":"Yermolin, Yevgeny"},{"full_name":"Karbachevsky, Alex","last_name":"Karbachevsky","first_name":"Alex"},{"first_name":"Alexander","last_name":"Bronstein","full_name":"Bronstein, Alexander","id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730"},{"first_name":"Avi","last_name":"Mendelson","full_name":"Mendelson, Avi"}],"main_file_link":[{"open_access":"1","url":"https://doi.org/10.48550/arXiv.1905.10830"}],"extern":"1","conference":{"start_date":"2020-07-19","end_date":"2020-07-24","name":"International Joint Conference on Neural Networks","location":"Glasgow, United Kingdom"},"oa":1,"title":"Feature map transform coding for energy-efficient CNN inference","_id":"18247","publication_identifier":{"eissn":["2161-4407"],"isbn":["9781728169279"]},"quality_controlled":"1"},{"publication_status":"published","oa_version":"None","user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","year":"2020","date_published":"2020-07-28T00:00:00Z","abstract":[{"text":"Learning an object detection or retrieval system requires a large data set with manual annotations. Such data are expensive and time-consuming to create and therefore difficult to obtain on a large scale. In this work, we propose using the natural correlation in narrations and the visual presence of objects in video to learn an object detector and retriever without any manual labeling involved. We pose the problem as weakly supervised learning with noisy labels, and propose a novel object detection and retrieval paradigm under these constraints. We handle the background rejection by using contrastive samples and confront the high level of label noise with a new clustering score. Our evaluation is based on a set of ten objects with manual ground truth annotation in almost 5000 frames extracted from instructional videos from the web. We demonstrate superior results compared to state-of-the-art weakly- supervised approaches and report a strongly-labeled upper bound as well. While the focus of the paper is object detection and retrieval, the proposed methodology can be applied to a broader range of noisy weakly-supervised problems.","lang":"eng"}],"article_processing_charge":"No","scopus_import":"1","article_number":"9150938","language":[{"iso":"eng"}],"author":[{"full_name":"Amrani, Elad","first_name":"Elad","last_name":"Amrani"},{"first_name":"Rami","last_name":"Ben-Ari","full_name":"Ben-Ari, Rami"},{"full_name":"Shapira, Inbar","last_name":"Shapira","first_name":"Inbar"},{"first_name":"Tal","last_name":"Hakim","full_name":"Hakim, Tal"},{"first_name":"Alexander","last_name":"Bronstein","full_name":"Bronstein, Alexander","id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730"}],"doi":"10.1109/cvprw50498.2020.00485","citation":{"ista":"Amrani E, Ben-Ari R, Shapira I, Hakim T, Bronstein AM. 2020. Self-supervised object detection and retrieval using unlabeled videos. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 9150938.","chicago":"Amrani, Elad, Rami Ben-Ari, Inbar Shapira, Tal Hakim, and Alex M. Bronstein. “Self-Supervised Object Detection and Retrieval Using Unlabeled Videos.” In <i>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</i>. IEEE, 2020. <a href=\"https://doi.org/10.1109/cvprw50498.2020.00485\">https://doi.org/10.1109/cvprw50498.2020.00485</a>.","short":"E. Amrani, R. Ben-Ari, I. Shapira, T. Hakim, A.M. Bronstein, in:, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), IEEE, 2020.","ama":"Amrani E, Ben-Ari R, Shapira I, Hakim T, Bronstein AM. Self-supervised object detection and retrieval using unlabeled videos. In: <i>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</i>. IEEE; 2020. doi:<a href=\"https://doi.org/10.1109/cvprw50498.2020.00485\">10.1109/cvprw50498.2020.00485</a>","mla":"Amrani, Elad, et al. “Self-Supervised Object Detection and Retrieval Using Unlabeled Videos.” <i>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</i>, 9150938, IEEE, 2020, doi:<a href=\"https://doi.org/10.1109/cvprw50498.2020.00485\">10.1109/cvprw50498.2020.00485</a>.","apa":"Amrani, E., Ben-Ari, R., Shapira, I., Hakim, T., &#38; Bronstein, A. M. (2020). Self-supervised object detection and retrieval using unlabeled videos. In <i>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</i>. Seattle, WA, United States: IEEE. <a href=\"https://doi.org/10.1109/cvprw50498.2020.00485\">https://doi.org/10.1109/cvprw50498.2020.00485</a>","ieee":"E. Amrani, R. Ben-Ari, I. Shapira, T. Hakim, and A. M. Bronstein, “Self-supervised object detection and retrieval using unlabeled videos,” in <i>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</i>, Seattle, WA, United States, 2020."},"month":"07","publication":"2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","publisher":"IEEE","conference":{"location":"Seattle, WA, United States","name":"IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops","start_date":"2020-06-14","end_date":"2020-06-19"},"extern":"1","quality_controlled":"1","date_created":"2024-10-08T13:05:08Z","status":"public","day":"28","type":"conference","publication_identifier":{"isbn":["9781728193618"],"eissn":["2160-7516"]},"_id":"18248","date_updated":"2024-12-12T09:59:41Z","title":"Self-supervised object detection and retrieval using unlabeled videos"},{"main_file_link":[{"url":"https://doi.org/10.48550/arXiv.1905.09324","open_access":"1"}],"extern":"1","conference":{"location":"Barcelona, Spain","start_date":"2020-05-04","end_date":"2020-05-08","name":"IEEE International Conference on Acoustics, Speech, and Signal Processing"},"oa":1,"title":"Joint learning of cartesian undersampling and reconstruction for accelerated MRI","_id":"18249","publication_identifier":{"isbn":["9781509066322"],"eissn":["2379-190X"]},"quality_controlled":"1","date_published":"2020-04-09T00:00:00Z","year":"2020","publication_status":"published","oa_version":"Preprint","month":"04","citation":{"ama":"Weiss T, Vedula S, Senouf O, Michailovich O, Zibulevsky M, Bronstein AM. Joint learning of cartesian undersampling and reconstruction for accelerated MRI. In: <i>ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</i>. IEEE; 2020. doi:<a href=\"https://doi.org/10.1109/icassp40776.2020.9054542\">10.1109/icassp40776.2020.9054542</a>","ieee":"T. Weiss, S. Vedula, O. Senouf, O. Michailovich, M. Zibulevsky, and A. M. Bronstein, “Joint learning of cartesian undersampling and reconstruction for accelerated MRI,” in <i>ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</i>, Barcelona, Spain, 2020.","apa":"Weiss, T., Vedula, S., Senouf, O., Michailovich, O., Zibulevsky, M., &#38; Bronstein, A. M. (2020). Joint learning of cartesian undersampling and reconstruction for accelerated MRI. In <i>ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</i>. Barcelona, Spain: IEEE. <a href=\"https://doi.org/10.1109/icassp40776.2020.9054542\">https://doi.org/10.1109/icassp40776.2020.9054542</a>","mla":"Weiss, Tomer, et al. “Joint Learning of Cartesian Undersampling and Reconstruction for Accelerated MRI.” <i>ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</i>, 9054542, IEEE, 2020, doi:<a href=\"https://doi.org/10.1109/icassp40776.2020.9054542\">10.1109/icassp40776.2020.9054542</a>.","chicago":"Weiss, Tomer, Sanketh Vedula, Ortal Senouf, Oleg Michailovich, Michael Zibulevsky, and Alex M. Bronstein. “Joint Learning of Cartesian Undersampling and Reconstruction for Accelerated MRI.” In <i>ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</i>. IEEE, 2020. <a href=\"https://doi.org/10.1109/icassp40776.2020.9054542\">https://doi.org/10.1109/icassp40776.2020.9054542</a>.","short":"T. Weiss, S. Vedula, O. Senouf, O. Michailovich, M. Zibulevsky, A.M. Bronstein, in:, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2020.","ista":"Weiss T, Vedula S, Senouf O, Michailovich O, Zibulevsky M, Bronstein AM. 2020. Joint learning of cartesian undersampling and reconstruction for accelerated MRI. ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE International Conference on Acoustics, Speech, and Signal Processing, 9054542."},"author":[{"last_name":"Weiss","first_name":"Tomer","full_name":"Weiss, Tomer"},{"last_name":"Vedula","first_name":"Sanketh","full_name":"Vedula, Sanketh"},{"full_name":"Senouf, Ortal","last_name":"Senouf","first_name":"Ortal"},{"full_name":"Michailovich, Oleg","last_name":"Michailovich","first_name":"Oleg"},{"full_name":"Zibulevsky, Michael","first_name":"Michael","last_name":"Zibulevsky"},{"id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730","first_name":"Alexander","last_name":"Bronstein","full_name":"Bronstein, Alexander"}],"publisher":"IEEE","publication":"ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","arxiv":1,"date_updated":"2024-12-11T16:06:20Z","type":"conference","day":"09","status":"public","date_created":"2024-10-08T13:05:24Z","language":[{"iso":"eng"}],"article_number":"9054542","scopus_import":"1","article_processing_charge":"No","abstract":[{"text":"Magnetic Resonance Imaging (MRI) is considered today the golden-standard modality for soft tissues. The long acquisition times, however, make it more prone to motion artifacts as well as contribute to the relative high costs of this examination. Over the years, multiple studies concentrated on designing reduced measurement schemes and image reconstruction schemes for MRI, however these problems have been so far addressed separately. On the other hand, recent works in optical computational imaging have demonstrated growing success of simultaneous learning-based design of the acquisition and reconstruction schemes manifesting significant improvement in the reconstruction quality with a constrained time budget. Inspired by these successes, in this work, we propose to learn accelerated MR acquisition schemes (in the form of Cartesian trajectories) jointly with the image reconstruction operator. To this end, we propose an algorithm for training the combined acquisition-reconstruction pipeline end-to-end in a differentiable way. We demonstrate the significance of using the learned Cartesian trajectories at different speed up rates. Code available at https://github.com/tomer196/fastMRI-Cartesian.","lang":"eng"}],"user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","doi":"10.1109/icassp40776.2020.9054542","external_id":{"arxiv":["1905.09324"]}},{"publication_identifier":{"eissn":["2160-9306"],"issn":["1077-2626"]},"title":"Hamiltonian operator for spectral shape analysis","_id":"18250","quality_controlled":"1","intvolume":"        26","extern":"1","main_file_link":[{"open_access":"1","url":"https://doi.org/10.48550/arXiv.1611.01990 "}],"oa":1,"month":"02","issue":"2","citation":{"ama":"Choukroun Y, Shtern A, Bronstein AM, Kimmel R. Hamiltonian operator for spectral shape analysis. <i>IEEE Transactions on Visualization and Computer Graphics</i>. 2020;26(2):1320-1331. doi:<a href=\"https://doi.org/10.1109/tvcg.2018.2867513\">10.1109/tvcg.2018.2867513</a>","mla":"Choukroun, Yoni, et al. “Hamiltonian Operator for Spectral Shape Analysis.” <i>IEEE Transactions on Visualization and Computer Graphics</i>, vol. 26, no. 2, Institute of Electrical and Electronics Engineers, 2020, pp. 1320–31, doi:<a href=\"https://doi.org/10.1109/tvcg.2018.2867513\">10.1109/tvcg.2018.2867513</a>.","ieee":"Y. Choukroun, A. Shtern, A. M. Bronstein, and R. Kimmel, “Hamiltonian operator for spectral shape analysis,” <i>IEEE Transactions on Visualization and Computer Graphics</i>, vol. 26, no. 2. Institute of Electrical and Electronics Engineers, pp. 1320–1331, 2020.","apa":"Choukroun, Y., Shtern, A., Bronstein, A. M., &#38; Kimmel, R. (2020). Hamiltonian operator for spectral shape analysis. <i>IEEE Transactions on Visualization and Computer Graphics</i>. Institute of Electrical and Electronics Engineers. <a href=\"https://doi.org/10.1109/tvcg.2018.2867513\">https://doi.org/10.1109/tvcg.2018.2867513</a>","ista":"Choukroun Y, Shtern A, Bronstein AM, Kimmel R. 2020. Hamiltonian operator for spectral shape analysis. IEEE Transactions on Visualization and Computer Graphics. 26(2), 1320–1331.","short":"Y. Choukroun, A. Shtern, A.M. Bronstein, R. Kimmel, IEEE Transactions on Visualization and Computer Graphics 26 (2020) 1320–1331.","chicago":"Choukroun, Yoni, Alon Shtern, Alex M. Bronstein, and Ron Kimmel. “Hamiltonian Operator for Spectral Shape Analysis.” <i>IEEE Transactions on Visualization and Computer Graphics</i>. Institute of Electrical and Electronics Engineers, 2020. <a href=\"https://doi.org/10.1109/tvcg.2018.2867513\">https://doi.org/10.1109/tvcg.2018.2867513</a>."},"author":[{"last_name":"Choukroun","first_name":"Yoni","full_name":"Choukroun, Yoni"},{"full_name":"Shtern, Alon","last_name":"Shtern","first_name":"Alon"},{"orcid":"0000-0001-9699-8730","id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","full_name":"Bronstein, Alexander","first_name":"Alexander","last_name":"Bronstein"},{"last_name":"Kimmel","first_name":"Ron","full_name":"Kimmel, Ron"}],"article_type":"original","year":"2020","publication_status":"published","oa_version":"Preprint","date_published":"2020-02-01T00:00:00Z","OA_place":"repository","type":"journal_article","day":"01","arxiv":1,"date_updated":"2024-10-15T09:43:31Z","status":"public","date_created":"2024-10-08T13:05:41Z","page":"1320-1331","pmid":1,"publisher":"Institute of Electrical and Electronics Engineers","publication":"IEEE Transactions on Visualization and Computer Graphics","doi":"10.1109/tvcg.2018.2867513","external_id":{"arxiv":["1611.01990"],"pmid":["30176599"]},"volume":26,"scopus_import":"1","language":[{"iso":"eng"}],"OA_type":"green","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_processing_charge":"No","abstract":[{"text":"Many shape analysis methods treat the geometry of an object as a metric space that can be captured by the Laplace-Beltrami operator. In this paper, we propose to adapt the classical Hamiltonian operator from quantum mechanics to the field of shape analysis. To this end, we study the addition of a potential function to the Laplacian as a generator for dual spaces in which shape processing is performed. We present general optimization approaches for solving variational problems involving the basis defined by the Hamiltonian using perturbation theory for its eigenvectors. The suggested operator is shown to produce better functional spaces to operate with, as demonstrated on different shape analysis tasks.","lang":"eng"}]},{"author":[{"full_name":"Alush-Aben, Jonathan","first_name":"Jonathan","last_name":"Alush-Aben"},{"first_name":"Linor","last_name":"Ackerman-Schraier","full_name":"Ackerman-Schraier, Linor"},{"first_name":"Tomer","last_name":"Weiss","full_name":"Weiss, Tomer"},{"first_name":"Sanketh","last_name":"Vedula","full_name":"Vedula, Sanketh"},{"full_name":"Senouf, Ortal","first_name":"Ortal","last_name":"Senouf"},{"full_name":"Bronstein, Alexander","first_name":"Alexander","last_name":"Bronstein","id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730"}],"volume":12450,"month":"10","citation":{"ista":"Alush-Aben J, Ackerman-Schraier L, Weiss T, Vedula S, Senouf O, Bronstein AM. 2020. 3D FLAT: Feasible learned acquisition trajectories for accelerated MRI. International Workshop on Machine Learning for Medical Image Reconstruction. MLMIR: Workshop on Machine Learning for Medical Image Reconstruction, LNCS, vol. 12450, 3–16.","short":"J. Alush-Aben, L. Ackerman-Schraier, T. Weiss, S. Vedula, O. Senouf, A.M. Bronstein, in:, International Workshop on Machine Learning for Medical Image Reconstruction, Springer Nature, 2020, pp. 3–16.","chicago":"Alush-Aben, Jonathan, Linor Ackerman-Schraier, Tomer Weiss, Sanketh Vedula, Ortal Senouf, and Alex M. Bronstein. “3D FLAT: Feasible Learned Acquisition Trajectories for Accelerated MRI.” In <i>International Workshop on Machine Learning for Medical Image Reconstruction</i>, 12450:3–16. Springer Nature, 2020. <a href=\"https://doi.org/10.1007/978-3-030-61598-7_1\">https://doi.org/10.1007/978-3-030-61598-7_1</a>.","ama":"Alush-Aben J, Ackerman-Schraier L, Weiss T, Vedula S, Senouf O, Bronstein AM. 3D FLAT: Feasible learned acquisition trajectories for accelerated MRI. In: <i>International Workshop on Machine Learning for Medical Image Reconstruction</i>. Vol 12450. Springer Nature; 2020:3-16. doi:<a href=\"https://doi.org/10.1007/978-3-030-61598-7_1\">10.1007/978-3-030-61598-7_1</a>","mla":"Alush-Aben, Jonathan, et al. “3D FLAT: Feasible Learned Acquisition Trajectories for Accelerated MRI.” <i>International Workshop on Machine Learning for Medical Image Reconstruction</i>, vol. 12450, Springer Nature, 2020, pp. 3–16, doi:<a href=\"https://doi.org/10.1007/978-3-030-61598-7_1\">10.1007/978-3-030-61598-7_1</a>.","apa":"Alush-Aben, J., Ackerman-Schraier, L., Weiss, T., Vedula, S., Senouf, O., &#38; Bronstein, A. M. (2020). 3D FLAT: Feasible learned acquisition trajectories for accelerated MRI. In <i>International Workshop on Machine Learning for Medical Image Reconstruction</i> (Vol. 12450, pp. 3–16). Lima, Peru: Springer Nature. <a href=\"https://doi.org/10.1007/978-3-030-61598-7_1\">https://doi.org/10.1007/978-3-030-61598-7_1</a>","ieee":"J. Alush-Aben, L. Ackerman-Schraier, T. Weiss, S. Vedula, O. Senouf, and A. M. Bronstein, “3D FLAT: Feasible learned acquisition trajectories for accelerated MRI,” in <i>International Workshop on Machine Learning for Medical Image Reconstruction</i>, Lima, Peru, 2020, vol. 12450, pp. 3–16."},"doi":"10.1007/978-3-030-61598-7_1","article_processing_charge":"No","abstract":[{"lang":"eng","text":"Magnetic Resonance Imaging (MRI) has long been considered to be among the gold standards of today’s diagnostic imaging. The most significant drawback of MRI is long acquisition times, prohibiting its use in standard practice for some applications. Compressed sensing (CS) proposes to subsample the k-space (the Fourier domain dual to the physical space of spatial coordinates) leading to significantly accelerated acquisition. However, the benefit of compressed sensing has not been fully exploited; most of the sampling densities obtained through CS do not produce a trajectory that obeys the stringent constraints of the MRI machine imposed in practice. Inspired by recent success of deep learning-based approaches for image reconstruction and ideas from computational imaging on learning-based design of imaging systems, we introduce 3D FLAT, a novel protocol for data-driven design of 3D non-Cartesian accelerated trajectories in MRI. Our proposal leverages the entire 3D k-space to simultaneously learn a physically feasible acquisition trajectory with a reconstruction method. Experimental results, performed as a proof-of-concept, suggest that 3D FLAT achieves higher image quality for a given readout time compared to standard trajectories such as radial, stack-of-stars, or 2D learned trajectories (trajectories that evolve only in the 2D plane while fully sampling along the third dimension). Furthermore, we demonstrate evidence supporting the significant benefit of performing MRI acquisitions using non-Cartesian 3D trajectories over 2D non-Cartesian trajectories acquired slice-wise."}],"date_published":"2020-10-20T00:00:00Z","year":"2020","oa_version":"None","user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","publication_status":"published","language":[{"iso":"eng"}],"scopus_import":"1","status":"public","date_created":"2024-10-08T13:06:03Z","intvolume":"     12450","alternative_title":["LNCS"],"quality_controlled":"1","title":"3D FLAT: Feasible learned acquisition trajectories for accelerated MRI","date_updated":"2025-01-23T15:13:44Z","_id":"18251","type":"conference","publication_identifier":{"issn":["0302-9743"],"eisbn":["9783030615987"],"eissn":["1611-3349"],"isbn":["9783030615970"]},"day":"20","publisher":"Springer Nature","publication":"International Workshop on Machine Learning for Medical Image Reconstruction","extern":"1","conference":{"location":"Lima, Peru","name":"MLMIR: Workshop on Machine Learning for Medical Image Reconstruction","start_date":"2020-10-08","end_date":"2020-10-08"},"page":"3 - 16"},{"publication_status":"published","oa_version":"None","year":"2020","date_published":"2020-01-01T00:00:00Z","article_type":"original","author":[{"last_name":"Rotker","first_name":"Keren","full_name":"Rotker, Keren"},{"first_name":"Dafna Ben","last_name":"Bashat","full_name":"Bashat, Dafna Ben"},{"id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730","last_name":"Bronstein","first_name":"Alexander","full_name":"Bronstein, Alexander"}],"issue":"3","citation":{"ista":"Rotker K, Bashat DB, Bronstein AM. 2020. Overparameterized models for vector fields. SIAM Journal on Imaging Sciences. 13(3), 1386–1414.","chicago":"Rotker, Keren, Dafna Ben Bashat, and Alex M. Bronstein. “Overparameterized Models for Vector Fields.” <i>SIAM Journal on Imaging Sciences</i>. Society for Industrial &#38; Applied Mathematics, 2020. <a href=\"https://doi.org/10.1137/19m1280697\">https://doi.org/10.1137/19m1280697</a>.","short":"K. Rotker, D.B. Bashat, A.M. Bronstein, SIAM Journal on Imaging Sciences 13 (2020) 1386–1414.","mla":"Rotker, Keren, et al. “Overparameterized Models for Vector Fields.” <i>SIAM Journal on Imaging Sciences</i>, vol. 13, no. 3, Society for Industrial &#38; Applied Mathematics, 2020, pp. 1386–414, doi:<a href=\"https://doi.org/10.1137/19m1280697\">10.1137/19m1280697</a>.","apa":"Rotker, K., Bashat, D. B., &#38; Bronstein, A. M. (2020). Overparameterized models for vector fields. <i>SIAM Journal on Imaging Sciences</i>. Society for Industrial &#38; Applied Mathematics. <a href=\"https://doi.org/10.1137/19m1280697\">https://doi.org/10.1137/19m1280697</a>","ieee":"K. Rotker, D. B. Bashat, and A. M. Bronstein, “Overparameterized models for vector fields,” <i>SIAM Journal on Imaging Sciences</i>, vol. 13, no. 3. Society for Industrial &#38; Applied Mathematics, pp. 1386–1414, 2020.","ama":"Rotker K, Bashat DB, Bronstein AM. Overparameterized models for vector fields. <i>SIAM Journal on Imaging Sciences</i>. 2020;13(3):1386-1414. doi:<a href=\"https://doi.org/10.1137/19m1280697\">10.1137/19m1280697</a>"},"month":"01","extern":"1","quality_controlled":"1","intvolume":"        13","publication_identifier":{"eissn":["1936-4954"]},"_id":"18252","title":"Overparameterized models for vector fields","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","abstract":[{"text":"Vector fields arise in a variety of quantity measure and visualization techniques, such as fluid flow imaging, motion estimation, deformation measures, and color imaging, leading to a better understanding of physical phenomena. Recent progress in vector field imaging technologies has emphasized the need for efficient noise removal and reconstruction algorithms. A key ingredient in the successful extraction of signals from noisy measurements is prior information, which can often be represented as a parameterized model. In this work, we extend the overparameterization variational framework in order to perform model-based reconstruction of vector fields. The overparameterization methodology combines local modeling of the data with global model parameter regularization. By considering the vector field as a linear combination of basis vector fields and appropriate scale and rotation coefficients, we can reduce the denoising problem to a simpler form of coefficient recovery. We introduce two versions of the overparameterization framework: a total variation-based method and a sparsity-based method, which relies on the cosparse analysis model. We demonstrate the efficiency of the proposed frameworks for two- and three-dimensional vector fields with linear and quadratic overparameterization models.","lang":"eng"}],"article_processing_charge":"No","scopus_import":"1","OA_type":"closed access","language":[{"iso":"eng"}],"volume":13,"doi":"10.1137/19m1280697","publication":"SIAM Journal on Imaging Sciences","publisher":"Society for Industrial & Applied Mathematics","page":"1386-1414","status":"public","date_created":"2024-10-08T13:06:25Z","day":"01","type":"journal_article","date_updated":"2024-10-15T10:43:38Z"},{"status":"public","date_created":"2024-10-08T13:06:43Z","date_updated":"2024-10-15T10:50:42Z","day":"01","type":"journal_article","publication":"mBio","publisher":"American Society for Microbiology","pmid":1,"volume":11,"external_id":{"pmid":["32371600"]},"doi":"10.1128/mbio.00705-20","abstract":[{"text":"PCNA, the ring that encircles DNA maintaining the processivity of DNA polymerases, is modified by ubiquitin and SUMO. Whereas ubiquitin is required for bypassing lesions through the DNA damage tolerance (DDT) pathways, we show here that SUMOylation represses another pathway, salvage recombination. The Srs2 helicase is recruited to SUMOylated PCNA and prevents the salvage pathway from acting. The pathway can be induced by overexpressing the PCNA unloader Elg1, or the homologous recombination protein Rad52. Our results underscore the role of PCNA modifications in controlling the various bypass and DNA repair mechanisms.","lang":"eng"}],"article_processing_charge":"Yes","user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","article_number":"00705-20","language":[{"iso":"eng"}],"OA_type":"gold","scopus_import":"1","intvolume":"        11","quality_controlled":"1","_id":"18253","title":"Access to PCNA by Srs2 and Elg1 controls the choice between alternative repair pathways in Saccharomyces cerevisiae","publication_identifier":{"issn":["2161-2129"],"eissn":["2150-7511"]},"oa":1,"extern":"1","main_file_link":[{"url":"https://doi.org/10.1128/mbio.00705-20","open_access":"1"}],"author":[{"first_name":"Matan","last_name":"Arbel","full_name":"Arbel, Matan"},{"full_name":"Bronstein, Alexander","last_name":"Bronstein","first_name":"Alexander","id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730"},{"last_name":"Sau","first_name":"Soumitra","full_name":"Sau, Soumitra"},{"full_name":"Liefshitz, Batia","last_name":"Liefshitz","first_name":"Batia"},{"full_name":"Kupiec, Martin","first_name":"Martin","last_name":"Kupiec"}],"citation":{"ieee":"M. Arbel, A. M. Bronstein, S. Sau, B. Liefshitz, and M. Kupiec, “Access to PCNA by Srs2 and Elg1 controls the choice between alternative repair pathways in Saccharomyces cerevisiae,” <i>mBio</i>, vol. 11, no. 3. American Society for Microbiology, 2020.","apa":"Arbel, M., Bronstein, A. M., Sau, S., Liefshitz, B., &#38; Kupiec, M. (2020). Access to PCNA by Srs2 and Elg1 controls the choice between alternative repair pathways in Saccharomyces cerevisiae. <i>MBio</i>. American Society for Microbiology. <a href=\"https://doi.org/10.1128/mbio.00705-20\">https://doi.org/10.1128/mbio.00705-20</a>","mla":"Arbel, Matan, et al. “Access to PCNA by Srs2 and Elg1 Controls the Choice between Alternative Repair Pathways in Saccharomyces Cerevisiae.” <i>MBio</i>, vol. 11, no. 3, 00705-20, American Society for Microbiology, 2020, doi:<a href=\"https://doi.org/10.1128/mbio.00705-20\">10.1128/mbio.00705-20</a>.","ama":"Arbel M, Bronstein AM, Sau S, Liefshitz B, Kupiec M. Access to PCNA by Srs2 and Elg1 controls the choice between alternative repair pathways in Saccharomyces cerevisiae. <i>mBio</i>. 2020;11(3). doi:<a href=\"https://doi.org/10.1128/mbio.00705-20\">10.1128/mbio.00705-20</a>","short":"M. Arbel, A.M. Bronstein, S. Sau, B. Liefshitz, M. Kupiec, MBio 11 (2020).","chicago":"Arbel, Matan, Alex M. Bronstein, Soumitra Sau, Batia Liefshitz, and Martin Kupiec. “Access to PCNA by Srs2 and Elg1 Controls the Choice between Alternative Repair Pathways in Saccharomyces Cerevisiae.” <i>MBio</i>. American Society for Microbiology, 2020. <a href=\"https://doi.org/10.1128/mbio.00705-20\">https://doi.org/10.1128/mbio.00705-20</a>.","ista":"Arbel M, Bronstein AM, Sau S, Liefshitz B, Kupiec M. 2020. Access to PCNA by Srs2 and Elg1 controls the choice between alternative repair pathways in Saccharomyces cerevisiae. mBio. 11(3), 00705-20."},"issue":"3","month":"06","OA_place":"publisher","date_published":"2020-06-01T00:00:00Z","DOAJ_listed":"1","oa_version":"Published Version","publication_status":"published","year":"2020","article_type":"original"},{"status":"public","date_created":"2024-10-08T13:07:16Z","date_updated":"2024-12-05T16:04:03Z","arxiv":1,"type":"conference","day":"05","publisher":"IEEE","publication":"2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)","doi":"10.1109/iccvw.2019.00567","external_id":{"arxiv":["1905.11137"]},"article_processing_charge":"No","abstract":[{"text":"Learning an object detection or retrieval system requires a large data set with manual annotations. Such data sets are expensive and time consuming to create and therefore difficult to obtain on a large scale. In this work, we propose to exploit the natural correlation in narrations and the visual presence of objects in video, to learn an object detector and retrieval without any manual labeling involved. We pose the problem as weakly supervised learning with noisy labels, and propose a novel object detection paradigm under these constraints. We handle the background rejection by using contrastive samples and confront the high level of label noise with a new clustering score. Our evaluation is based on a set of 11 manually annotated objects in over 5000 frames. We show comparison to a weakly-supervised approach as baseline and provide a strongly labeled upper bound.","lang":"eng"}],"user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","language":[{"iso":"eng"}],"article_number":"9022341","scopus_import":"1","quality_controlled":"1","title":"Learning to detect and retrieve objects from unlabeled videos","_id":"18255","publication_identifier":{"isbn":["9781728150246"],"eissn":["2473-9944"]},"oa":1,"main_file_link":[{"open_access":"1","url":"https://doi.org/10.48550/arXiv.1905.11137"}],"extern":"1","conference":{"name":"17th IEEE/CVF International Conference on Computer Vision Workshop","start_date":"2019-10-27","end_date":"2019-10-28","location":"Seoul, Korea (South)"},"author":[{"full_name":"Amrani, Elad","last_name":"Amrani","first_name":"Elad"},{"first_name":"Rami","last_name":"Ben-Ari","full_name":"Ben-Ari, Rami"},{"first_name":"Tal","last_name":"Hakim","full_name":"Hakim, Tal"},{"id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730","full_name":"Bronstein, Alexander","first_name":"Alexander","last_name":"Bronstein"}],"month":"03","citation":{"short":"E. Amrani, R. Ben-Ari, T. Hakim, A.M. Bronstein, in:, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), IEEE, 2020.","chicago":"Amrani, Elad, Rami Ben-Ari, Tal Hakim, and Alex M. Bronstein. “Learning to Detect and Retrieve Objects from Unlabeled Videos.” In <i>2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)</i>. IEEE, 2020. <a href=\"https://doi.org/10.1109/iccvw.2019.00567\">https://doi.org/10.1109/iccvw.2019.00567</a>.","ista":"Amrani E, Ben-Ari R, Hakim T, Bronstein AM. 2020. Learning to detect and retrieve objects from unlabeled videos. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW). 17th IEEE/CVF International Conference on Computer Vision Workshop, 9022341.","ama":"Amrani E, Ben-Ari R, Hakim T, Bronstein AM. Learning to detect and retrieve objects from unlabeled videos. In: <i>2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)</i>. IEEE; 2020. doi:<a href=\"https://doi.org/10.1109/iccvw.2019.00567\">10.1109/iccvw.2019.00567</a>","apa":"Amrani, E., Ben-Ari, R., Hakim, T., &#38; Bronstein, A. M. (2020). Learning to detect and retrieve objects from unlabeled videos. In <i>2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)</i>. Seoul, Korea (South): IEEE. <a href=\"https://doi.org/10.1109/iccvw.2019.00567\">https://doi.org/10.1109/iccvw.2019.00567</a>","ieee":"E. Amrani, R. Ben-Ari, T. Hakim, and A. M. Bronstein, “Learning to detect and retrieve objects from unlabeled videos,” in <i>2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)</i>, Seoul, Korea (South), 2020.","mla":"Amrani, Elad, et al. “Learning to Detect and Retrieve Objects from Unlabeled Videos.” <i>2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)</i>, 9022341, IEEE, 2020, doi:<a href=\"https://doi.org/10.1109/iccvw.2019.00567\">10.1109/iccvw.2019.00567</a>."},"date_published":"2020-03-05T00:00:00Z","year":"2020","oa_version":"Preprint","publication_status":"published"},{"article_processing_charge":"No","abstract":[{"text":"Distance metric learning (DML) has been successfully applied to object classification, both in the standard regime of rich training data and in the few-shot scenario, where each category is represented by only a few examples. In this work, we propose a new method for DML that simultaneously learns the backbone network parameters, the embedding space, and the multi-modal distribution of each of the training categories in that space, in a single end-to-end training process. Our approach outperforms state-of-the-art methods for DML-based object classification on a variety of standard fine-grained datasets. Furthermore, we demonstrate the effectiveness of our approach on the problem of few-shot object detection, by incorporating the proposed DML architecture as a classification head into a standard object detection model. We achieve the best results on the ImageNet-LOC dataset compared to strong baselines, when only a few training examples are available. We also offer the community a new episodic benchmark based on the ImageNet dataset for the few-shot object detection task.","lang":"eng"}],"date_published":"2020-01-09T00:00:00Z","year":"2020","user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","oa_version":"None","publication_status":"published","language":[{"iso":"eng"}],"article_number":"8953439","scopus_import":"1","author":[{"full_name":"Karlinsky, Leonid","last_name":"Karlinsky","first_name":"Leonid"},{"full_name":"Shtok, Joseph","last_name":"Shtok","first_name":"Joseph"},{"full_name":"Harary, Sivan","first_name":"Sivan","last_name":"Harary"},{"first_name":"Eli","last_name":"Schwartz","full_name":"Schwartz, Eli"},{"first_name":"Amit","last_name":"Aides","full_name":"Aides, Amit"},{"last_name":"Feris","first_name":"Rogerio","full_name":"Feris, Rogerio"},{"full_name":"Giryes, Raja","last_name":"Giryes","first_name":"Raja"},{"full_name":"Bronstein, Alexander","last_name":"Bronstein","first_name":"Alexander","orcid":"0000-0001-9699-8730","id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6"}],"month":"01","citation":{"ista":"Karlinsky L, Shtok J, Harary S, Schwartz E, Aides A, Feris R, Giryes R, Bronstein AM. 2020. Repmet: Representative-based metric learning for classification and few-shot object detection. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition, 8953439.","short":"L. Karlinsky, J. Shtok, S. Harary, E. Schwartz, A. Aides, R. Feris, R. Giryes, A.M. Bronstein, in:, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, 2020.","chicago":"Karlinsky, Leonid, Joseph Shtok, Sivan Harary, Eli Schwartz, Amit Aides, Rogerio Feris, Raja Giryes, and Alex M. Bronstein. “Repmet: Representative-Based Metric Learning for Classification and Few-Shot Object Detection.” In <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>. IEEE, 2020. <a href=\"https://doi.org/10.1109/cvpr.2019.00534\">https://doi.org/10.1109/cvpr.2019.00534</a>.","ama":"Karlinsky L, Shtok J, Harary S, et al. Repmet: Representative-based metric learning for classification and few-shot object detection. In: <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>. IEEE; 2020. doi:<a href=\"https://doi.org/10.1109/cvpr.2019.00534\">10.1109/cvpr.2019.00534</a>","mla":"Karlinsky, Leonid, et al. “Repmet: Representative-Based Metric Learning for Classification and Few-Shot Object Detection.” <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>, 8953439, IEEE, 2020, doi:<a href=\"https://doi.org/10.1109/cvpr.2019.00534\">10.1109/cvpr.2019.00534</a>.","apa":"Karlinsky, L., Shtok, J., Harary, S., Schwartz, E., Aides, A., Feris, R., … Bronstein, A. M. (2020). Repmet: Representative-based metric learning for classification and few-shot object detection. In <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>. Long Beach, CA, United States: IEEE. <a href=\"https://doi.org/10.1109/cvpr.2019.00534\">https://doi.org/10.1109/cvpr.2019.00534</a>","ieee":"L. Karlinsky <i>et al.</i>, “Repmet: Representative-based metric learning for classification and few-shot object detection,” in <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>, Long Beach, CA, United States, 2020."},"doi":"10.1109/cvpr.2019.00534","publisher":"IEEE","publication":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","extern":"1","conference":{"start_date":"2019-06-15","end_date":"2019-06-20","name":"32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition","location":"Long Beach, CA, United States"},"date_created":"2024-10-08T13:08:09Z","status":"public","quality_controlled":"1","date_updated":"2024-12-05T15:38:16Z","title":"Repmet: Representative-based metric learning for classification and few-shot object detection","_id":"18258","publication_identifier":{"isbn":["9781728132945"],"eissn":["2575-7075"]},"type":"conference","day":"09"},{"external_id":{"arxiv":["1902.09811"]},"doi":"10.1109/cvpr.2019.00671","article_number":"8954088","language":[{"iso":"eng"}],"scopus_import":"1","abstract":[{"lang":"eng","text":"Example synthesis is one of the leading methods to tackle the problem of few-shot learning, where only a small number of samples per class are available. However, current synthesis approaches only address the scenario of a single category label per image. In this work, we propose a novel technique for synthesizing samples with multiple labels for the (yet unhandled) multi-label few-shot classification scenario. We propose to combine pairs of given examples in feature space, so that the resulting synthesized feature vectors will correspond to examples whose label sets are obtained through certain set operations on the label sets of the corresponding input pairs. Thus, our method is capable of producing a sample containing the intersection, union or set-difference of labels present in two input samples. As we show, these set operations generalize to labels unseen during training. This enables performing augmentation on examples of novel categories, thus, facilitating multi-label few-shot classifier learning. We conduct numerous experiments showing promising results for the label-set manipulation capabilities of the proposed approach, both directly (using the classification and retrieval metrics), and in the context of performing data augmentation for multi-label few-shot learning. We propose a benchmark for this new and challenging task and show that our method compares favorably to all the common baselines."}],"article_processing_charge":"No","user_id":"3E5EF7F0-F248-11E8-B48F-1D18A9856A87","arxiv":1,"date_updated":"2024-12-05T15:33:21Z","day":"09","type":"conference","date_created":"2024-10-08T13:08:26Z","status":"public","publication":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","publisher":"IEEE","citation":{"mla":"Alfassy, Amit, et al. “Laso: Label-Set Operations Networks for Multi-Label Few-Shot Learning.” <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>, 8954088, IEEE, 2020, doi:<a href=\"https://doi.org/10.1109/cvpr.2019.00671\">10.1109/cvpr.2019.00671</a>.","ieee":"A. Alfassy <i>et al.</i>, “Laso: Label-set operations networks for multi-label few-shot learning,” in <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>, Long Beach, CA, United States, 2020.","apa":"Alfassy, A., Karlinsky, L., Aides, A., Shtok, J., Harary, S., Feris, R., … Bronstein, A. M. (2020). Laso: Label-set operations networks for multi-label few-shot learning. In <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>. Long Beach, CA, United States: IEEE. <a href=\"https://doi.org/10.1109/cvpr.2019.00671\">https://doi.org/10.1109/cvpr.2019.00671</a>","ama":"Alfassy A, Karlinsky L, Aides A, et al. Laso: Label-set operations networks for multi-label few-shot learning. In: <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>. IEEE; 2020. doi:<a href=\"https://doi.org/10.1109/cvpr.2019.00671\">10.1109/cvpr.2019.00671</a>","ista":"Alfassy A, Karlinsky L, Aides A, Shtok J, Harary S, Feris R, Giryes R, Bronstein AM. 2020. Laso: Label-set operations networks for multi-label few-shot learning. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition, 8954088.","short":"A. Alfassy, L. Karlinsky, A. Aides, J. Shtok, S. Harary, R. Feris, R. Giryes, A.M. Bronstein, in:, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, 2020.","chicago":"Alfassy, Amit, Leonid Karlinsky, Amit Aides, Joseph Shtok, Sivan Harary, Rogerio Feris, Raja Giryes, and Alex M. Bronstein. “Laso: Label-Set Operations Networks for Multi-Label Few-Shot Learning.” In <i>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</i>. IEEE, 2020. <a href=\"https://doi.org/10.1109/cvpr.2019.00671\">https://doi.org/10.1109/cvpr.2019.00671</a>."},"month":"01","author":[{"full_name":"Alfassy, Amit","first_name":"Amit","last_name":"Alfassy"},{"first_name":"Leonid","last_name":"Karlinsky","full_name":"Karlinsky, Leonid"},{"last_name":"Aides","first_name":"Amit","full_name":"Aides, Amit"},{"full_name":"Shtok, Joseph","first_name":"Joseph","last_name":"Shtok"},{"full_name":"Harary, Sivan","first_name":"Sivan","last_name":"Harary"},{"full_name":"Feris, Rogerio","last_name":"Feris","first_name":"Rogerio"},{"last_name":"Giryes","first_name":"Raja","full_name":"Giryes, Raja"},{"id":"58f3726e-7cba-11ef-ad8b-e6e8cb3904e6","orcid":"0000-0001-9699-8730","last_name":"Bronstein","first_name":"Alexander","full_name":"Bronstein, Alexander"}],"date_published":"2020-01-09T00:00:00Z","oa_version":"Preprint","publication_status":"published","year":"2020","_id":"18259","title":"Laso: Label-set operations networks for multi-label few-shot learning","publication_identifier":{"eissn":["2575-7075"],"isbn":["9781728132945"]},"quality_controlled":"1","extern":"1","main_file_link":[{"url":"https://doi.org/10.48550/arXiv.1902.09811","open_access":"1"}],"conference":{"end_date":"2019-06-20","start_date":"2019-06-15","name":"32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition","location":"Long Beach, CA, United States"},"oa":1}]
