A pilot study of the relationship between expertsʼ ratings and scores generated by the NBMEʼs Computer-Based Examination System
- 1 February 1992
- journal article
- Published by Wolters Kluwer Health in Academic Medicine
- Vol. 67 (2) , 130-2
- https://doi.org/10.1097/00001888-199202000-00020
Abstract
This pilot study evaluates the consistency of experts' ratings of students' performances on the National Board of Medical Examiners' Computer Based Examination (CBX) cases and the relationship of those ratings to the CBX's scoring algorithm. The authors were investigating whether an automated scoring algorithm can adequately assess an examinee's management of a computer-simulated patient. In 1989–90, at the Michigan State University College of Human Medicine, eight students, completing a surgery clerkship, each managed eight CBX cases and took a computer-administered, multiple-choice examination. Six clerkship coordinators rated the students' performances in terms of overall management, efficiency, and dangerous actions. The ratings correlated highly with scores produced by the CBX's scoring system.Keywords
This publication has 0 references indexed in Scilit: