Does it fit? Is it good? Assessment of scoring systems
- 1 June 2000
- journal article
- severity scoring-in-the-critically-ill-patient
- Published by Wolters Kluwer Health in Current Opinion in Critical Care
- Vol. 6 (3) , 176-180
- https://doi.org/10.1097/00075198-200006000-00006
Abstract
The goal of intensive care is to provide the highest quality care to achieve the best outcomes for patients. Although the random allocation of patients to receive intensive care as part of a randomized controlled trial might in theory be the best method to evaluate its effectiveness, the practical application of such a study design is deemed unethical. The alternative in such a situation is to use observational methods in which the outcome of care patients receive as part of their “natural” treatment is studied. Before drawing inferences from the outcomes of treatment for such groups of patients, the characteristics of the patients admitted to intensive care must be taken into account. The purpose of scoring systems in intensive care, as for many other areas of health care, is to take into account the characteristics of patients that could affect their risk of a particular outcome, irrespective of the effect of the care they receive. Clearly, accounting for such factors, which are outside the control of those providing the care, is essential before any comparison of the outcome of care is possible. For each scoring system, the association between the independent variables (patient characteristics) and the dependent variable (death before discharge from hospital after intensive care) is described in the form of a mathematical model, known as a multiple logistic regression model. To assess any scoring system and model, four steps must be undertaken. First, the goodness-of-fit of the model should be assessed on the development dataset. Second, the goodness-of-fit of the model should be assessed on a subset of the dataset different from the subset used to develop the model. Third, the goodness-of-fit of the model should be assessed on an entirely new dataset. Fourth, the given model should be compared with other existing models on the same dataset. This article briefly describes the concept of scoring systems and their roles, and outlines the requirements for assessment in terms of model validation and goodness-of-fit. It concludes with the outstanding methodological issues in this area.Keywords
This publication has 5 references indexed in Scilit:
- What do we mean by validating a prognostic model?Statistics in Medicine, 2000
- A New Simplified Acute Physiology Score (SAPS II) Based on a European/North American Multicenter StudyJAMA, 1993
- The APACHE III Prognostic SystemChest, 1991
- APACHE IICritical Care Medicine, 1985
- Statistical Approaches to the Analysis of Receiver Operating Characteristic (ROC) CurvesMedical Decision Making, 1984