Online Calibration of a Joint Model of Item Responses and Response Times in Computerized Adaptive Testing

Hyeon Ah Kang, Yi Zheng, Hua Hua Chang

Research output: Contribution to journalArticlepeer-review


With the widespread use of computers in modern assessment, online calibration has become increasingly popular as a way of replenishing an item pool. The present study discusses online calibration strategies for a joint model of responses and response times. The study proposes likelihood inference methods for item paramter estimation and evaluates their performance along with optimal sampling procedures. An extensive simulation study indicates that the proposed online calibration strategies perform well with relatively small samples (e.g., 500∼800 examinees). The analysis of estimated parameters suggests that response time information can be used to improve the recovery of the response model parameters. Among a number of sampling methods investigated, A-optimal sampling was found most advantageous when the item parameters were weakly correlated. When the parameters were strongly correlated, D-optimal sampling tended to achieve the most accurate parameter recovery. The study provides guidelines for deciding sampling design under a specific goal of online calibration given the characteristics of field-testing items.

Original languageEnglish (US)
Pages (from-to)175-208
Number of pages34
JournalJournal of Educational and Behavioral Statistics
Issue number2
StatePublished - Apr 1 2020
Externally publishedYes


  • computerized adaptive testing
  • item response theory
  • online calibration
  • optimal sampling
  • response time

ASJC Scopus subject areas

  • Education
  • Social Sciences (miscellaneous)


Dive into the research topics of 'Online Calibration of a Joint Model of Item Responses and Response Times in Computerized Adaptive Testing'. Together they form a unique fingerprint.

Cite this