Estimating the Errors Remaining in a Data Set: Techniques for Quality Control
- 1 February 1990
- journal article
- general
- Published by Taylor & Francis in The American Statistician
- Vol. 44 (1) , 14-18
- https://doi.org/10.1080/00031305.1990.10475684
Abstract
This article presents two methods of quantifying the adequacy with which research data have been checked in the process of quality control. In the duplicate performance method, the data operation is carried out twice, independently, and the results are compared; the remaining errors in the data set can be estimated thereby and a confidence limit can be obtained. In the known errors method, the supervisor purposely introduces into a data set known errors similar in form to suspected unknown errors. Then a staff member checks the file; the results yield the number of known errors found and the number of unknown errors found. The method, like the duplicate performance method, allows the accuracy of both workers to be quantified and allows an estimate, with a confidence limit, of the number of as-yet-unfound errors still lurking in the data set.Keywords
This publication has 6 references indexed in Scilit:
- Guidelines for clean data: Detection of common mistakes.Journal of Applied Psychology, 1986
- The Design of the Epidemiologic Catchment Area SurveysArchives of General Psychiatry, 1984
- Statistics in Society: Problems Unsolved and UnformulatedJournal of the American Statistical Association, 1981
- How often are our numbers wrong?American Psychologist, 1978
- Responsibility for Raw Data.American Psychologist, 1962
- The Case of the Indians and the Teen-Age WidowsJournal of the American Statistical Association, 1962