Quantification of Errors in Ordinal Outcome Scales Using Shannon Entropy: Effect on Sample Size Calculations. e67754
Objective Clinical trial outcomes often involve an ordinal scale of subjective functional assessments but the optimal way to quantify results is not clear. In stroke, the most commonly used scale, the modified Rankin Score (mRS), a range of scores ("Shift") is proposed as superior to dicho...
Saved in:
Published in: | PloS one Vol. 8; no. 7 |
---|---|
Main Authors: | , , , , |
Format: | Journal Article |
Language: | English |
Published: |
01-07-2013
|
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Objective Clinical trial outcomes often involve an ordinal scale of subjective functional assessments but the optimal way to quantify results is not clear. In stroke, the most commonly used scale, the modified Rankin Score (mRS), a range of scores ("Shift") is proposed as superior to dichotomization because of greater information transfer. The influence of known uncertainties in mRS assessment has not been quantified. We hypothesized that errors caused by uncertainties could be quantified by applying information theory. Using Shannon's model, we quantified errors of the "Shift" compared to dichotomized outcomes using published distributions of mRS uncertainties and applied this model to clinical trials. Methods We identified 35 randomized stroke trials that met inclusion criteria. Each trial's mRS distribution was multiplied with the noise distribution from published mRS inter-rater variability to generate an error percentage for "shift" and dichotomized cut-points. For the SAINT I neuroprotectant trial, considered positive by "shift" mRS while the larger follow-up SAINT II trial was negative, we recalculated sample size required if classification uncertainty was taken into account. Results Considering the full mRS range, error rate was 26.1% plus or minus 5.31 (Mean plus or minus SD). Error rates were lower for all dichotomizations tested using cut-points (e.g. mRS 1; 6.8% plus or minus 2.89; overall p<0.001). Taking errors into account, SAINT I would have required 24% more subjects than were randomized. Conclusion We show when uncertainty in assessments is considered, the lowest error rates are with dichotomization. While using the full range of mRS is conceptually appealing, a gain of information is counter-balanced by a decrease in reliability. The resultant errors need to be considered since sample size may otherwise be underestimated. In principle, we have outlined an approach to error estimation for any condition in which there are uncertainties in outcome assessment. We provide the user with programs to calculate and incorporate errors into sample size estimation. |
---|---|
Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 content type line 23 ObjectType-Feature-2 |
ISSN: | 1932-6203 |
DOI: | 10.1371/journal.pone.0067754 |