A Robust Method for Detecting Item Misfit in Large-Scale Assessments
Viable methods for the identification of item misfit or Differential Item Functioning (DIF) are central to scale construction and sound measurement. Many approaches rely on the derivation of a limiting distribution under the assumption that a certain model fits the data perfectly. Typical DIF assump...
Saved in:
Published in | Educational and psychological measurement Vol. 83; no. 4; pp. 740 - 765 |
---|---|
Main Authors | , |
Format | Journal Article |
Language | English |
Published |
Los Angeles, CA
SAGE Publications
01.08.2023
SAGE PUBLICATIONS, INC |
Subjects | |
Online Access | Get full text |
ISSN | 0013-1644 1552-3888 1552-3888 |
DOI | 10.1177/00131644221105819 |
Cover
Summary: | Viable methods for the identification of item misfit or Differential Item Functioning (DIF) are central to scale construction and sound measurement. Many approaches rely on the derivation of a limiting distribution under the assumption that a certain model fits the data perfectly. Typical DIF assumptions such as the monotonicity and population independence of item functions are present even in classical test theory but are more explicitly stated when using item response theory or other latent variable models for the assessment of item fit. The work presented here provides a robust approach for DIF detection that does not assume perfect model data fit, but rather uses Tukey’s concept of contaminated distributions. The approach uses robust outlier detection to flag items for which adequate model data fit cannot be established. |
---|---|
Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 content type line 23 |
ISSN: | 0013-1644 1552-3888 1552-3888 |
DOI: | 10.1177/00131644221105819 |