On the importance of severely testing deep learning models of cognition

Jeffrey S. Bowers, Gaurav Malhotra, Federico Adolfi, Marin Dujmović, Milton L. Montero, Valerio Biscione, Guillermo Puebla, John H. Hummel, Rachel F. Heaton

Research output: Contribution to journalArticlepeer-review

Abstract

Researchers studying the correspondences between Deep Neural Networks (DNNs) and humans often give little consideration to severe testing when drawing conclusions from empirical findings, and this is impeding progress in building better models of minds. We first detail what we mean by severe testing and highlight how this is especially important when working with opaque models with many free parameters that may solve a given task in multiple different ways. Second, we provide multiple examples of researchers making strong claims regarding DNN-human similarities without engaging in severe testing of their hypotheses. Third, we consider why severe testing is undervalued. We provide evidence that part of the fault lies with the review process. There is now a widespread appreciation in many areas of science that a bias for publishing positive results (among other practices) is leading to a credibility crisis, but there seems less awareness of the problem here.

Original languageEnglish (US)
Article number101158
JournalCognitive Systems Research
Volume82
DOIs
StatePublished - Dec 2023

Keywords

  • Memory
  • Neural networks
  • Perception
  • Psychology
  • Severe testing
  • Vision

ASJC Scopus subject areas

  • Software
  • Experimental and Cognitive Psychology
  • Cognitive Neuroscience
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'On the importance of severely testing deep learning models of cognition'. Together they form a unique fingerprint.

Cite this