Detection of Gender-Related Differential Item Functioning in a Mathematics Performance Assessment
- 1 April 1996
- journal article
- conference paper
- Published by Taylor & Francis in Applied Measurement in Education
- Vol. 9 (2) , 175-199
- https://doi.org/10.1207/s15324818ame0902_6
Abstract
This study used three different differential item functioning (DIF) detection proce- dures to examine the extent to which items in a mathematics performance assessment functioned differently for matched gender groups. In addition to examining the appropriateness of individual items in terms of DIF with respect to gender, an attempt was made to identify factors (e.g., content, cognitive processes, differences in ability distributions, etc.) that may be related to DIF. The QUASAR (Quantitative Under- standing: Amplifying Student Achievement and Reasoning) Cognitive Assessment Instrument (QCAI) is designed to measure students' mathematical thinking and reasoning skills and consists of open-ended items that require students to show their solution processes and provide explanations for their answers. In this study, 33 polytomously scored items, which were distributed within four test forms, were evaluated with respect to gender-related DIF. The data source was sixth- and seventh- grade student responses to each of the four test forms administrated in the spring of 1992 at all six school sites participatingin the QUASARproject. The sample consisted of 1,782 students with approximately equal numbers of female and male students. The results indicated that DIF may not be serious for 3 1 of the 33 items (94%) in the QCAI. For the two items that were detected as functioning differently for male and female students, several plausible factors for DIF were discussed. The results from the secondary analyses, which removed the mutual influence of the two items, indicated that DIF in one item, PPPl, which favored female students rather than their matched male students, was of particular concern. These secondary analyses suggest that the detection of DIF in the other item in the original analysis may have been due to the influence of Item PPPl because they were both in the same test form.This publication has 17 references indexed in Scilit:
- Generalizability and Validity of a Mathematics Performance AssessmentJournal of Educational Measurement, 1996
- The Conceptual Framework for the Development of a Mathematics Performance Assessment InstrumentEducational Measurement: Issues and Practice, 1993
- Logistic Discriminant Function Analysis for DIF Identification of Polytomously Scored ItemsJournal of Educational Measurement, 1993
- Patterns of Gender Differences on Mathematics Items on the Scholastic Aptitude TestApplied Measurement in Education, 1993
- Applied Logistic RegressionTechnometrics, 1992
- Complex, Performance-Based Assessment: Expectations and Validation CriteriaEducational Researcher, 1991
- Detecting Differential Item Functioning Using Logistic Regression ProceduresJournal of Educational Measurement, 1990
- When Do Item Response Function and Mantel-Haenszel Definitions of Differential Item Functioning Coincide?Journal of Educational Statistics, 1990
- Gender‐Based Differential Item Performance in Mathematics Achievement ItemsJournal of Educational Measurement, 1987
- Contingency Table Models for Assessing Item BiasJournal of Educational Statistics, 1982