RELIABILITY OF CRITERION‐REFERENCED TESTS: A DECISION‐THEORETIC FORMULATION
- 1 December 1974
- journal article
- Published by Wiley in Journal of Educational Measurement
- Vol. 11 (4) , 263-267
- https://doi.org/10.1111/j.1745-3984.1974.tb00998.x
Abstract
It has been suggested that the primary purpose for criterion‐referenced testing in objective‐based instructional programs is to classify examinees into mastery states or categories on the objectives included in the test. We have proposed that the reliability of the criterion‐referenced test scores be defined in terms of the consistency of the decision‐making process across repeated administrations of the test. Specifically, reliability is defined as a measure of agreement over and above that which can be expected by chance between the decisions made about examinee mastery states in repeated test administrations for each objective measured by the criterion‐referenced test.Keywords
This publication has 8 references indexed in Scilit:
- TOWARD AN INTEGRATION OF THEORY AND METHOD FOR CRITERION‐REFERENCED TESTS1,2Journal of Educational Measurement, 1973
- Overview of innovations in organization for learningInterchange, 1972
- Weighted Chi Square: an Extension of the Kappa MethodEducational and Psychological Measurement, 1972
- What is individualized instruction?Interchange, 1970
- Large sample standard errors of kappa and weighted kappa.Psychological Bulletin, 1969
- MOMENTS OF THE STATISTICS KAPPA AND WEIGHTED KAPPABritish Journal of Mathematical and Statistical Psychology, 1968
- Weighted kappa: Nominal scale agreement provision for scaled disagreement or partial credit.Psychological Bulletin, 1968
- A Coefficient of Agreement for Nominal ScalesEducational and Psychological Measurement, 1960