Reproducibility of Vertebral Fracture Assessment Readings From Dual-energy X-ray Absorptiometry in Both a Population-based and Clinical Cohort: Cohen's and Uniform Kappa.

TitreReproducibility of Vertebral Fracture Assessment Readings From Dual-energy X-ray Absorptiometry in Both a Population-based and Clinical Cohort: Cohen's and Uniform Kappa.
Publication TypeJournal Article
Year of Publication2015
AuthorsAubry-Rozier, B, Chapurlat, R, Duboeuf, F, Iglesias, K, Krieg, MA, Lamy, O, Burnand, B, Hans, D
JournalJournal of Clinical Densitometry
Volume18
Issue2
Pagination233-238
Date Published04-06/2015
DOI10.1016/j.jocd.2014.09.001
ISSN1094-6950
Mots-clésAbsorptiometry, Photon, Aged, Aged, 80 and over, Cohort Studies, Female, Humans, Lumbar Vertebrae, Middle Aged, Osteoporotic Fractures, Reproducibility of Results, Spinal Fractures, Spine, Switzerland, Thoracic Vertebrae
Abstract

Vertebral fracture assessments (VFAs) using dual-energy X-ray absorptiometry increase vertebral fracture detection in clinical practice and are highly reproducible. Measures of reproducibility are dependent on the frequency and distribution of the event. The aim of this study was to compare 2 reproducibility measures, reliability and agreement, in VFA readings in both a population-based and a clinical cohort. We measured agreement and reliability by uniform kappa and Cohen's kappa for vertebral reading and fracture identification: 360 VFAs from a population-based cohort and 85 from a clinical cohort. In the population-based cohort, 12% of vertebrae were unreadable. Vertebral fracture prevalence ranged from 3% to 4%. Inter-reader and intrareader reliability with Cohen's kappa was fair to good (0.35-0.71 and 0.36-0.74, respectively), with good inter-reader and intrareader agreement by uniform kappa (0.74-0.98 and 0.76-0.99, respectively). In the clinical cohort, 15% of vertebrae were unreadable, and vertebral fracture prevalence ranged from 7.6% to 8.1%. Inter-reader reliability was moderate to good (0.43-0.71), and the agreement was good (0.68-0.91). In clinical situations, the levels of reproducibility measured by the 2 kappa statistics are concordant, so that either could be used to measure agreement and reliability. However, if events are rare, as in a population-based cohort, we recommend evaluating reproducibility using the uniform kappa, as Cohen's kappa may be less accurate.

Notes

Publication types: Journal Article Publication Status: ppublish

Alternate URL

http://www.ncbi.nlm.nih.gov/pubmed/25439454?dopt=Abstract

First publication date (online)

11/2014

WOS ID (UT)

000354667800017

Alternate JournalJ Clin Densitom
Citation Key / SERVAL IDserval:BIB_A39DE0EB404F
Peer reviewRefereed
PubMed ID25439454

                         

IUMSP | www.iumsp.ch
Institut universitaire de médecine sociale et préventive
Route de la Corniche 10, 1010 Lausanne - Switzerland
+41 21 314 72 72 | dess.info@unisante.ch

Go to top