TY - JOUR
T1 - Multirater agreement of arthroscopic meniscal lesions
AU - Dunn, Warren R.
AU - Wolf, Brian R.
AU - Amendola, Annunziato
AU - Andrish, Jack T.
AU - Kaeding, Christopher
AU - Marx, Robert G.
AU - McCarty, Eric C.
AU - Parker, Richard D.
AU - Wright, Rick W.
AU - Spindler, Kurt P.
PY - 2004/12
Y1 - 2004/12
N2 - Background: Establishing the validity of classification schemes is a crucial preparatory step that should precede multicenter studies. There are no studies investigating the reproducibility of arthroscopic classification of meniscal pathology among multiple surgeons at different institutions. Hypothesis: Arthroscopic classification of meniscal pathology is reliable and reproducible and suitable for multicenter studies that involve multiple surgeons. Study Design: Multirater agreement study. Methods: Seven surgeons reviewed a video of 18 meniscal tears and completed a meniscal classification questionnaire. Multirater agreement was calculated based on the proportion of agreement, the kappa coefficient, and the intraclass correlation coefficient. Results: There was a 46% agreement on the central/peripheral location of tears (κ = 0.30), an 80% agreement on the depth of tears (κ = 0.46), a 72% agreement on the presence of a degenerative component (κ = 0.44), a 71% agreement on whether lateral tears were central to the popliteal hiatus (κ = 0.42), a 73% agreement on the type of tear (κ = 0.63), an 87% agreement on the location of the tear (κ = 0.61), and an 84% agreement on the treatment of tears (κ = 0.66). There was considerable agreement among surgeons on length, with an intraclass correlation coefficient of 0.78, 95% confidence interval of 0.57 to 0.92, and P < .001. Conclusions: Arthroscopic grading of meniscal pathology is reliable and reproducible. Clinical Relevance: Surgeons can reliably classify meniscal pathology and agree on treatment, which is important for multicenter trials.
AB - Background: Establishing the validity of classification schemes is a crucial preparatory step that should precede multicenter studies. There are no studies investigating the reproducibility of arthroscopic classification of meniscal pathology among multiple surgeons at different institutions. Hypothesis: Arthroscopic classification of meniscal pathology is reliable and reproducible and suitable for multicenter studies that involve multiple surgeons. Study Design: Multirater agreement study. Methods: Seven surgeons reviewed a video of 18 meniscal tears and completed a meniscal classification questionnaire. Multirater agreement was calculated based on the proportion of agreement, the kappa coefficient, and the intraclass correlation coefficient. Results: There was a 46% agreement on the central/peripheral location of tears (κ = 0.30), an 80% agreement on the depth of tears (κ = 0.46), a 72% agreement on the presence of a degenerative component (κ = 0.44), a 71% agreement on whether lateral tears were central to the popliteal hiatus (κ = 0.42), a 73% agreement on the type of tear (κ = 0.63), an 87% agreement on the location of the tear (κ = 0.61), and an 84% agreement on the treatment of tears (κ = 0.66). There was considerable agreement among surgeons on length, with an intraclass correlation coefficient of 0.78, 95% confidence interval of 0.57 to 0.92, and P < .001. Conclusions: Arthroscopic grading of meniscal pathology is reliable and reproducible. Clinical Relevance: Surgeons can reliably classify meniscal pathology and agree on treatment, which is important for multicenter trials.
KW - Meniscus
KW - Multicenter
KW - Multicenter Orthpaedic Outcomes Network (MOON)
KW - Multirater agreement
UR - https://www.scopus.com/pages/publications/9244243143
U2 - 10.1177/0363546504264586
DO - 10.1177/0363546504264586
M3 - Review article
C2 - 15572324
AN - SCOPUS:9244243143
SN - 0363-5465
VL - 32
SP - 1937
EP - 1940
JO - American Journal of Sports Medicine
JF - American Journal of Sports Medicine
IS - 8
ER -