Skip to main content

Table 2 Fleiss’ Kappa for inter-rater Reliability in detecting a modifier in principle and in detecting the separate modifiers M1, M2 and M3 in the first and second surveys (Total: complete rater cohort, DR: developing raters, UR: user raters). For each survey, the absolute agreement of the raters´ choices with the “gold standard” is given

From: OF-Pelvis classification of osteoporotic sacral and pelvic ring fractures

  

1st survey

 

2st survey

 
  

0.95 CI

absolute agreement

 

0.95 CI

absolute agreement

Modifier

kappa

(lower-upper limit)

kappa

(lower-upper limit)

total

in principle

0.646

(0.625–0.667)

89%

0.629

(0.608–0.650)

90%

 

M1

0.756

(0.736–0.776)

95%

0.777

(0.757–0.797)

95%

 

M2

0.662

(0.642–0.682)

89%

0.569

(0.549–0.590)

86%

 

M3

0.549

(0.529–0.569)

87%

0.602

(0.581–0.622)

88%

DR

in principle

0.667

(0.623–0.712)

89%

0.702

(0.657–0.746)

91%

 

M1

0.810

(0.766–0.854)

96%

0.748

(0.703–0.792)

95%

 

M2

0.718

(0.673–0.762)

90%

0.614

(0.570–0.659)

88%

 

M3

0.568

(0.524–0.613)

88%

0.603

(0.559–0.648)

88%

UR

in principle

0.617

(0.576–0.658)

88%

0.565

(0.524–0.606)

89%

 

M1

0.712

(0.674–0.750)

94%

0.803

(0.765–0.841)

96%

 

M2

0.618

(0.579–0.656)

87%

0.532

(0.493–0.570)

85%

 

M3

0.526

(0.487–0.564)

86%

0.608

(0.569–0.646)

88%