They also experience other important limitations such as poor
interobserver variability (mRS), (6) floor and ceiling effect (BI), (7) and a focus on measuring physical impairment but not functional deficit (National Institutes of Health Stroke Scale).
Interobserver variability was assessed with the [kappa] statistic.
Interobserver variability was analysed with the Statistical Package for Social Sciences (SPSS, Chicago, IL; version 17.0).
Intraclass correlation coefficients were calculated for the intra- and
interobserver agreement and values [greater than equal to]0.75, 0.60.0.74, 0.40.0.59, and [less than equal to]0.40 were featured as high, good, fair, and poor agreement, respectively.
Finally,
interobserver reliability was evaluated between two examiners in an effort to evaluate the repeatability of the method in an approach similar to Fojas et al.
(34) independently presented 40 cases of CMC arthritis to five muscuoloskeletal radiologists and eight hand surgeons to evaluate
interobserver reliability of classification and treatment.
Repeatability of goniometric methods: Intraclass correlation coefficients also were calculated to evaluate intraobserver and
interobserver reliability of goniometry measurements obtained using the modified universal plastic goniometer and electro-goniometer.
To control for
interobserver variability, the centers submitted a total of 40 slides for central dermatopathology review, which found agreement in 35 cases (87.8%).
For all variables, we evaluated intra-observer agreement (i.e., agreements on interpretations when interpreting the same patient images using the two reading systems, the MEDICAL-IMPAX and TABLET-XERO), and
interobserver agreement (i.e., agreements between radiologists when interpreting the same patient within a single reading system).
Moreover the STONE nephrolithometry has excellent
interobserver reliability.
Our study revealed a moderate level of
interobserver agreement among these readers, indicating that different experience levels may affect the interpretation of images, even under the guidance of the PI-RADS v2.
On the other hand, Thomson's study showed that the
interobserver agreement for staining intensity for each antibody was good for 0+ and 3+ groups but poor for 1+ and 2+ groups (12).