TY - JOUR
T1 - Factuality challenges in the era of large language models and opportunities for fact-checking
AU - Augenstein, Isabelle
AU - Baldwin, Timothy
AU - Cha, Meeyoung
AU - Chakraborty, Tanmoy
AU - Ciampaglia, Giovanni Luca
AU - Corney, David
AU - DiResta, Renee
AU - Ferrara, Emilio
AU - Hale, Scott
AU - Halevy, Alon
AU - Hovy, Eduard
AU - Ji, Heng
AU - Menczer, Filippo
AU - Miguez, Ruben
AU - Nakov, Preslav
AU - Scheufele, Dietram
AU - Sharma, Shivam
AU - Zagni, Giovanni
N1 - M.C. is supported by the Institute for Basic Science (grant number IBS-R029-C2) and the National Research Foundation of Korea (grant number RS-2022-00165347). T.C. acknowledges the financial support of Wipro AI. I.A. is supported in part by the European Union (ERC, ExplainYourself, grant number 101077481). G.L.C. is supported by the National Science Foundation (grant numbers 2239194 and 2229885). E.F. and F.M. are partly supported by DARPA (award number HR001121C0169). F.M. is also partly supported by the Knight Foundation and Craig Newmark Philanthropies. G.Z.\u2019s fact-checking project receives funding from the European Union through multiple grants and is part of Meta\u2019s 3PFC Program. H.J. is partially supported by US DARPA SemaFor programme number HR001120C0123. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the funders.
PY - 2024/8
Y1 - 2024/8
N2 - The emergence of tools based on large language models (LLMs), such as OpenAI’s ChatGPT and Google’s Gemini, has garnered immense public attention owing to their advanced natural language generation capabilities. These remarkably natural-sounding tools have the potential to be highly useful for various tasks. However, they also tend to produce false, erroneous or misleading content—commonly referred to as hallucinations. Moreover, LLMs can be misused to generate convincing, yet false, content and profiles on a large scale, posing a substantial societal challenge by potentially deceiving users and spreading inaccurate information. This makes fact-checking increasingly important. Despite their issues with factual accuracy, LLMs have shown proficiency in various subtasks that support fact-checking, which is essential to ensure factually accurate responses. In light of these concerns, we explore issues related to factuality in LLMs and their impact on fact-checking. We identify key challenges, imminent threats and possible solutions to these factuality issues. We also thoroughly examine these challenges, existing solutions and potential prospects for fact-checking. By analysing the factuality constraints within LLMs and their impact on fact-checking, we aim to contribute to a path towards maintaining accuracy at a time of confluence of generative artificial intelligence and misinformation.
AB - The emergence of tools based on large language models (LLMs), such as OpenAI’s ChatGPT and Google’s Gemini, has garnered immense public attention owing to their advanced natural language generation capabilities. These remarkably natural-sounding tools have the potential to be highly useful for various tasks. However, they also tend to produce false, erroneous or misleading content—commonly referred to as hallucinations. Moreover, LLMs can be misused to generate convincing, yet false, content and profiles on a large scale, posing a substantial societal challenge by potentially deceiving users and spreading inaccurate information. This makes fact-checking increasingly important. Despite their issues with factual accuracy, LLMs have shown proficiency in various subtasks that support fact-checking, which is essential to ensure factually accurate responses. In light of these concerns, we explore issues related to factuality in LLMs and their impact on fact-checking. We identify key challenges, imminent threats and possible solutions to these factuality issues. We also thoroughly examine these challenges, existing solutions and potential prospects for fact-checking. By analysing the factuality constraints within LLMs and their impact on fact-checking, we aim to contribute to a path towards maintaining accuracy at a time of confluence of generative artificial intelligence and misinformation.
UR - http://www.scopus.com/inward/record.url?scp=85201793136&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85201793136&partnerID=8YFLogxK
U2 - 10.1038/s42256-024-00881-z
DO - 10.1038/s42256-024-00881-z
M3 - Article
AN - SCOPUS:85201793136
SN - 2522-5839
VL - 6
SP - 852
EP - 863
JO - Nature Machine Intelligence
JF - Nature Machine Intelligence
IS - 8
ER -