Abstract
In this community review report, we discuss applications and techniques for fast machine learning (ML) in science—the concept of integrating powerful ML methods into the real-time experimental data processing loop to accelerate scientific discovery. The material for the report builds on two workshops held by the Fast ML for Science community and covers three main areas: applications for fast ML across a number of scientific domains; techniques for training and implementing performant and resource-efficient ML algorithms; and computing architectures, platforms, and technologies for deploying these algorithms. We also present overlapping challenges across the multiple scientific domains where common solutions can be found. This community report is intended to give plenty of examples and inspiration for scientific discovery through integrated and accelerated ML solutions. This is followed by a high-level overview and organization of technical advances, including an abundance of pointers to source material, which can enable these breakthroughs.
Original language | English (US) |
---|---|
Article number | 787421 |
Journal | Frontiers in Big Data |
Volume | 5 |
DOIs | |
State | Published - Apr 12 2022 |
Keywords
- big data
- codesign
- coprocessors
- fast machine learning
- heterogeneous computing
- machine learning for science
- particle physics
ASJC Scopus subject areas
- Computer Science (miscellaneous)
- Information Systems
- Artificial Intelligence
Fingerprint
Dive into the research topics of 'Applications and Techniques for Fast Machine Learning in Science'. Together they form a unique fingerprint.Cite this
- APA
- Standard
- Harvard
- Vancouver
- Author
- BIBTEX
- RIS
In: Frontiers in Big Data, Vol. 5, 787421, 12.04.2022.
Research output: Contribution to journal › Review article › peer-review
}
TY - JOUR
T1 - Applications and Techniques for Fast Machine Learning in Science
AU - Deiana, Allison Mc Carn
AU - Tran, Nhan
AU - Agar, Joshua
AU - Blott, Michaela
AU - Di Guglielmo, Giuseppe
AU - Duarte, Javier
AU - Harris, Philip
AU - Hauck, Scott
AU - Liu, Mia
AU - Neubauer, Mark S.
AU - Ngadiuba, Jennifer
AU - Ogrenci-Memik, Seda
AU - Pierini, Maurizio
AU - Aarrestad, Thea
AU - Bähr, Steffen
AU - Becker, Jürgen
AU - Berthold, Anne Sophie
AU - Bonventre, Richard J.
AU - Müller Bravo, Tomás E.
AU - Diefenthaler, Markus
AU - Dong, Zhen
AU - Fritzsche, Nick
AU - Gholami, Amir
AU - Govorkova, Ekaterina
AU - Guo, Dongning
AU - Hazelwood, Kyle J.
AU - Herwig, Christian
AU - Khan, Babar
AU - Kim, Sehoon
AU - Klijnsma, Thomas
AU - Liu, Yaling
AU - Lo, Kin Ho
AU - Nguyen, Tri
AU - Pezzullo, Gianantonio
AU - Rasoulinezhad, Seyedramin
AU - Rivera, Ryan A.
AU - Scholberg, Kate
AU - Selig, Justin
AU - Sen, Sougata
AU - Strukov, Dmitri
AU - Tang, William
AU - Thais, Savannah
AU - Unger, Kai Lukas
AU - Vilalta, Ricardo
AU - von Krosigk, Belina
AU - Wang, Shen
AU - Warburton, Thomas K.
N1 - Funding Information: We acknowledge the Fast Machine Learning collective as an open community of multi-domain experts and collaborators. This community was important for the development of this project. The work by AD was supported by the U.S. Department of Energy (DOE), Office of Science, Office of High Energy Physics, under Award No. DE-SC0010129, and the Fast Machine Learning in Science Workshop was financially supported by Southern Methodist University. The work by NT was supported by Fermi Research Alliance, LLC under Contract No. DE-AC02-07CH11359 with the DOE, Office of Science, Office of High Energy Physics and the DOE Early Career Research program under Award No. DE-0000247070. The work by DS was supported by NSF E2CDA grant #1740352. JA acknowledges primary support from our DOE program and secondary support from National Science Foundation under grant TRIPODS + X: RES-1839234Y. The work of DG was supported in part by the National Science Foundation under Grant No. CNS-2003098 and by a gift from Intel Incorporation. YL acknowledges the support of this work from the National Institutes of Health grant of R01HL131750, and National Science Foundation grant of CBET 2039310. The work by MN was supported by the U.S. National Science Foundation under Cooperative Agreement OAC-1836650 and Award No. OAC-1934757. KS is supported by the U.S. Department of Energy and the National Science Foundation. The work by BK is funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under the Emmy Noether Grant No. 420484612. The work by MD was supported by Jefferson Science Associates, LLC under Contract No. DE-AC05-06OR23177 with the DOE, Office of Science, Office of Nuclear Physics. We would like to acknowledge community members who have explicitly supported this work: Maria Acosta Flechas (Fermilab), Anthony Aportela (UC San Diego), Thomas Calvet (CPP Marseille), Leonardo Cristella (CERN), Daniel Diaz (UC San Diego), Caterina Doglioni (Lund), Maria Domenica Galati (University of Groningen), Elham E Khoda (University of Washington), Farah Fahim (Fermilab), Davide Giri (Columbia University), Benjamin Hawks (Fermilab), Duc Hoang (MIT), Burt Holzman (Fermilab), Shih-Chieh Hsu (University of Washington), Sergo Jindariani (Fermilab), Iris Johnson (Fermilab), Raghav Kansal (UC San Diego), Ryan Kastner (UC San Diego), Erik Katsavounidis (MIT), Jeffrey Krupa (MIT), Pan Li (Purdue University), Vladimir Loncar (CERN, Institute of Physics Belgrad), Sandeep Madireddy (ANL), Ethan Marx (MIT), Patrick McCormack (MIT) Andres Meza (UC San Diego), Jovan Mitrevski (Fermilab), Mohammed Attia Mohammed (CHEP-FU), Farouk Mokhtar (UC San Diego), Eric Moreno (MIT), Srishti Nagu (Lucknow University), Rohin Narayan (SMU), Noah Paladino (MIT), Adrian Alan Pol (CERN), Zhiqiang Que (Imperial College), Sang Eon Park (MIT), Subramanian Ramamoorthy 28, Dylan Rankin (MIT), Simon Rothman (MIT), Ashish Sharma (IIT Madras), Sioni Summers (CERN), Pietro Vischia (UC Louvain), Jean-Roch Vlimant (Caltech), Olivia Weng (UC San Diego). Funding Information: We acknowledge the Fast Machine Learning collective as an open community of multi-domain experts and collaborators. This community was important for the development of this project. The work by AD was supported by the U.S. Department of Energy (DOE), Office of Science, Office of High Energy Physics, under Award No. DE-SC0010129, and the Fast Machine Learning in Science Workshop was financially supported by Southern Methodist University. The work by NT was supported by Fermi Research Alliance, LLC under Contract No. DE-AC02-07CH11359 with the DOE, Office of Science, Office of High Energy Physics and the DOE Early Career Research program under Award No. DE-0000247070. The work by DS was supported by NSF E2CDA grant #1740352. JA acknowledges primary support from our DOE program and secondary support from National Science Foundation under grant TRIPODS + X: RES-1839234Y. The work of DG was supported in part by the National Science Foundation under Grant No. CNS-2003098 and by a gift from Intel Incorporation. YL acknowledges the support of this work from the National Institutes of Health grant of R01HL131750, and National Science Foundation grant of CBET 2039310. The work by MN was supported by the U.S. National Science Foundation under Cooperative Agreement OAC-1836650 and Award No. OAC-1934757. KS is supported by the U.S. Department of Energy and the National Science Foundation. The work by BK is funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under the Emmy Noether Grant No. 420484612. The work by MD was supported by Jefferson Science Associates, LLC under Contract No. DE-AC05-06OR23177 with the DOE, Office of Science, Office of Nuclear Physics. We would like to acknowledge community members who have explicitly supported this work: Maria Acosta Flechas (Fermilab), Anthony Aportela (UC San Diego), Thomas Calvet (CPP Marseille), Leonardo Cristella (CERN), Daniel Diaz (UC San Diego), Caterina Doglioni (Lund), Maria Domenica Galati (University of Groningen), Elham E Khoda (University of Washington), Farah Fahim (Fermilab), Davide Giri (Columbia University), Benjamin Hawks (Fermilab), Duc Hoang (MIT), Burt Holzman (Fermilab), Shih-Chieh Hsu (University of Washington), Sergo Jindariani (Fermilab), Iris Johnson (Fermilab), Raghav Kansal (UC San Diego), Ryan Kastner (UC San Diego), Erik Katsavounidis (MIT), Jeffrey Krupa (MIT), Pan Li (Purdue University), Vladimir Loncar (CERN, Institute of Physics Belgrad), Sandeep Madireddy (ANL), Ethan Marx (MIT), Patrick McCormack (MIT) Andres Meza (UC San Diego), Jovan Mitrevski (Fermilab), Mohammed Attia Mohammed (CHEP-FU), Farouk Mokhtar (UC San Diego), Eric Moreno (MIT), Srishti Nagu (Lucknow University), Rohin Narayan (SMU), Noah Paladino (MIT), Adrian Alan Pol (CERN), Zhiqiang Que (Imperial College), Sang Eon Park (MIT), Subramanian Ramamoorthy , Dylan Rankin (MIT), Simon Rothman (MIT), Ashish Sharma (IIT Madras), Sioni Summers (CERN), Pietro Vischia (UC Louvain), Jean-Roch Vlimant (Caltech), Olivia Weng (UC San Diego). 28 Publisher Copyright: Copyright © 2022 Deiana, Tran, Agar, Blott, Di Guglielmo, Duarte, Harris, Hauck, Liu, Neubauer, Ngadiuba, Ogrenci-Memik, Pierini, Aarrestad, Bähr, Becker, Berthold, Bonventre, Müller Bravo, Diefenthaler, Dong, Fritzsche, Gholami, Govorkova, Guo, Hazelwood, Herwig, Khan, Kim, Klijnsma, Liu, Lo, Nguyen, Pezzullo, Rasoulinezhad, Rivera, Scholberg, Selig, Sen, Strukov, Tang, Thais, Unger, Vilalta, von Krosigk, Wang and Warburton.
PY - 2022/4/12
Y1 - 2022/4/12
N2 - In this community review report, we discuss applications and techniques for fast machine learning (ML) in science—the concept of integrating powerful ML methods into the real-time experimental data processing loop to accelerate scientific discovery. The material for the report builds on two workshops held by the Fast ML for Science community and covers three main areas: applications for fast ML across a number of scientific domains; techniques for training and implementing performant and resource-efficient ML algorithms; and computing architectures, platforms, and technologies for deploying these algorithms. We also present overlapping challenges across the multiple scientific domains where common solutions can be found. This community report is intended to give plenty of examples and inspiration for scientific discovery through integrated and accelerated ML solutions. This is followed by a high-level overview and organization of technical advances, including an abundance of pointers to source material, which can enable these breakthroughs.
AB - In this community review report, we discuss applications and techniques for fast machine learning (ML) in science—the concept of integrating powerful ML methods into the real-time experimental data processing loop to accelerate scientific discovery. The material for the report builds on two workshops held by the Fast ML for Science community and covers three main areas: applications for fast ML across a number of scientific domains; techniques for training and implementing performant and resource-efficient ML algorithms; and computing architectures, platforms, and technologies for deploying these algorithms. We also present overlapping challenges across the multiple scientific domains where common solutions can be found. This community report is intended to give plenty of examples and inspiration for scientific discovery through integrated and accelerated ML solutions. This is followed by a high-level overview and organization of technical advances, including an abundance of pointers to source material, which can enable these breakthroughs.
KW - big data
KW - codesign
KW - coprocessors
KW - fast machine learning
KW - heterogeneous computing
KW - machine learning for science
KW - particle physics
UR - http://www.scopus.com/inward/record.url?scp=85128972935&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85128972935&partnerID=8YFLogxK
U2 - 10.3389/fdata.2022.787421
DO - 10.3389/fdata.2022.787421
M3 - Review article
C2 - 35496379
AN - SCOPUS:85128972935
SN - 2624-909X
VL - 5
JO - Frontiers in Big Data
JF - Frontiers in Big Data
M1 - 787421
ER -