TY - JOUR
T1 - JARVIS-Leaderboard
T2 - a large scale benchmark of materials design methods
AU - Choudhary, Kamal
AU - Wines, Daniel
AU - Li, Kangming
AU - Garrity, Kevin F.
AU - Gupta, Vishu
AU - Romero, Aldo H.
AU - Krogel, Jaron T.
AU - Saritas, Kayahan
AU - Fuhr, Addis
AU - Ganesh, Panchapakesan
AU - Kent, Paul R.C.
AU - Yan, Keqiang
AU - Lin, Yuchao
AU - Ji, Shuiwang
AU - Blaiszik, Ben
AU - Reiser, Patrick
AU - Friederich, Pascal
AU - Agrawal, Ankit
AU - Tiwary, Pratyush
AU - Beyerle, Eric
AU - Minch, Peter
AU - Rhone, Trevor David
AU - Takeuchi, Ichiro
AU - Wexler, Robert B.
AU - Mannodi-Kanakkithodi, Arun
AU - Ertekin, Elif
AU - Mishra, Avanish
AU - Mathew, Nithin
AU - Wood, Mitchell
AU - Rohskopf, Andrew Dale
AU - Hattrick-Simpers, Jason
AU - Wang, Shih Han
AU - Achenie, Luke E.K.
AU - Xin, Hongliang
AU - Williams, Maureen
AU - Biacchi, Adam J.
AU - Tavazza, Francesca
N1 - Publisher Copyright:
© This is a U.S. Government work and not under copyright protection in the US; foreign copyright protection may apply 2024.
PY - 2024/12
Y1 - 2024/12
N2 - Lack of rigorous reproducibility and validation are significant hurdles for scientific development across many fields. Materials science, in particular, encompasses a variety of experimental and theoretical approaches that require careful benchmarking. Leaderboard efforts have been developed previously to mitigate these issues. However, a comprehensive comparison and benchmarking on an integrated platform with multiple data modalities with perfect and defect materials data is still lacking. This work introduces JARVIS-Leaderboard, an open-source and community-driven platform that facilitates benchmarking and enhances reproducibility. The platform allows users to set up benchmarks with custom tasks and enables contributions in the form of dataset, code, and meta-data submissions. We cover the following materials design categories: Artificial Intelligence (AI), Electronic Structure (ES), Force-fields (FF), Quantum Computation (QC), and Experiments (EXP). For AI, we cover several types of input data, including atomic structures, atomistic images, spectra, and text. For ES, we consider multiple ES approaches, software packages, pseudopotentials, materials, and properties, comparing results to experiment. For FF, we compare multiple approaches for material property predictions. For QC, we benchmark Hamiltonian simulations using various quantum algorithms and circuits. Finally, for experiments, we use the inter-laboratory approach to establish benchmarks. There are 1281 contributions to 274 benchmarks using 152 methods with more than 8 million data points, and the leaderboard is continuously expanding. The JARVIS-Leaderboard is available at the website: https://pages.nist.gov/jarvis_leaderboard/.
AB - Lack of rigorous reproducibility and validation are significant hurdles for scientific development across many fields. Materials science, in particular, encompasses a variety of experimental and theoretical approaches that require careful benchmarking. Leaderboard efforts have been developed previously to mitigate these issues. However, a comprehensive comparison and benchmarking on an integrated platform with multiple data modalities with perfect and defect materials data is still lacking. This work introduces JARVIS-Leaderboard, an open-source and community-driven platform that facilitates benchmarking and enhances reproducibility. The platform allows users to set up benchmarks with custom tasks and enables contributions in the form of dataset, code, and meta-data submissions. We cover the following materials design categories: Artificial Intelligence (AI), Electronic Structure (ES), Force-fields (FF), Quantum Computation (QC), and Experiments (EXP). For AI, we cover several types of input data, including atomic structures, atomistic images, spectra, and text. For ES, we consider multiple ES approaches, software packages, pseudopotentials, materials, and properties, comparing results to experiment. For FF, we compare multiple approaches for material property predictions. For QC, we benchmark Hamiltonian simulations using various quantum algorithms and circuits. Finally, for experiments, we use the inter-laboratory approach to establish benchmarks. There are 1281 contributions to 274 benchmarks using 152 methods with more than 8 million data points, and the leaderboard is continuously expanding. The JARVIS-Leaderboard is available at the website: https://pages.nist.gov/jarvis_leaderboard/.
UR - http://www.scopus.com/inward/record.url?scp=85192277941&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85192277941&partnerID=8YFLogxK
U2 - 10.1038/s41524-024-01259-w
DO - 10.1038/s41524-024-01259-w
M3 - Article
AN - SCOPUS:85192277941
SN - 2057-3960
VL - 10
JO - npj Computational Materials
JF - npj Computational Materials
IS - 1
M1 - 93
ER -