Comparing methods of assessing differential item functioning in a computerized adaptive testing environment

Pui Wa Lei, Shu Ying Chen, Lan Yu

Research output: Contribution to journalArticlepeer-review

17 Scopus citations

Abstract

Mantel-Haenszel and SIBTEST, which have known difficulty in detecting non-unidirectional differential item functioning (DIF), have been adapted with some success for computerized adaptive testing (CAT). This study adapts logistic regression (LR) and the item-response-theory-likelihood-ratio test (IRT-LRT), capable of detecting both unidirectional and non-unidirectional DIP, to the CAT environment in which pretest items are assumed to be seeded in CATs but not used for trait estimation. The proposed adaptation methods were evaluated with simulated data under different sample size ratios and impact conditions in terms of Type I error, power, and specificity in identifying the form of DIF. The adapted LR and IRT-LRT procedures are more powerful than the CAT version of SIBTEST for non-unidirectional DIF detection. The good Type I error control provided by IRT-LRT under extremely unequal sample sizes and large impact is encouraging. Implications of these and other findings are discussed.

Original languageEnglish (US)
Pages (from-to)245-264
Number of pages20
JournalJournal of Educational Measurement
Volume43
Issue number3
DOIs
StatePublished - Sep 2006

All Science Journal Classification (ASJC) codes

  • Education
  • Developmental and Educational Psychology
  • Applied Psychology
  • Psychology (miscellaneous)

Fingerprint

Dive into the research topics of 'Comparing methods of assessing differential item functioning in a computerized adaptive testing environment'. Together they form a unique fingerprint.

Cite this