A Robust Method for Detecting Item Misfit in Large-Scale Assessments

Viable methods for the identification of item misfit or Differential Item Functioning (DIF) are central to scale construction and sound measurement. Many approaches rely on the derivation of a limiting distribution under the assumption that a certain model fits the data perfectly. Typical DIF assump...

Full description

Saved in:
Bibliographic Details
Published inEducational and psychological measurement Vol. 83; no. 4; pp. 740 - 765
Main Authors von Davier, Matthias, Bezirhan, Ummugul
Format Journal Article
LanguageEnglish
Published Los Angeles, CA SAGE Publications 01.08.2023
SAGE PUBLICATIONS, INC
Subjects
Online AccessGet full text
ISSN0013-1644
1552-3888
1552-3888
DOI10.1177/00131644221105819

Cover

More Information
Summary:Viable methods for the identification of item misfit or Differential Item Functioning (DIF) are central to scale construction and sound measurement. Many approaches rely on the derivation of a limiting distribution under the assumption that a certain model fits the data perfectly. Typical DIF assumptions such as the monotonicity and population independence of item functions are present even in classical test theory but are more explicitly stated when using item response theory or other latent variable models for the assessment of item fit. The work presented here provides a robust approach for DIF detection that does not assume perfect model data fit, but rather uses Tukey’s concept of contaminated distributions. The approach uses robust outlier detection to flag items for which adequate model data fit cannot be established.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ISSN:0013-1644
1552-3888
1552-3888
DOI:10.1177/00131644221105819