enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Cohen's kappa - Wikipedia

    en.wikipedia.org/wiki/Cohen's_kappa

    Cohen's kappa coefficient (κ, lowercase Greek kappa) is a statistic that is used to measure inter-rater reliability (and also intra-rater reliability) for qualitative (categorical) items. [1] It is generally thought to be a more robust measure than simple percent agreement calculation, as κ takes into account the possibility of the agreement ...

  3. Fleiss' kappa - Wikipedia

    en.wikipedia.org/wiki/Fleiss'_kappa

    Fleiss' kappa is a generalisation of Scott's pi statistic, [2] a statistical measure of inter-rater reliability. [3] It is also related to Cohen's kappa statistic and Youden's J statistic which may be more appropriate in certain instances. [4]

  4. Inter-rater reliability - Wikipedia

    en.wikipedia.org/wiki/Inter-rater_reliability

    In statistics, inter-rater reliability (also called by various similar names, such as inter-rater agreement, inter-rater concordance, inter-observer reliability, inter-coder reliability, and so on) is the degree of agreement among independent observers who rate, code, or assess the same phenomenon.

  5. Scott's Pi - Wikipedia

    en.wikipedia.org/wiki/Scott's_Pi

    Scott's pi (named after William A Scott) is a statistic for measuring inter-rater reliability for nominal data in communication studies.Textual entities are annotated with categories by different annotators, and various measures are used to assess the extent of agreement between the annotators, one of which is Scott's pi.

  6. Youden's J statistic - Wikipedia

    en.wikipedia.org/wiki/Youden's_J_statistic

    Youden's J statistic is = + = + with the two right-hand quantities being sensitivity and specificity.Thus the expanded formula is: = + + + = (+) (+) In this equation, TP is the number of true positives, TN the number of true negatives, FP the number of false positives and FN the number of false negatives.

  7. Krippendorff's alpha - Wikipedia

    en.wikipedia.org/wiki/Krippendorff's_alpha

    Krippendorff's alpha coefficient, [1] named after academic Klaus Krippendorff, is a statistical measure of the agreement achieved when coding a set of units of analysis.. Since the 1970s, alpha has been used in content analysis where textual units are categorized by trained readers, in counseling and survey research where experts code open-ended interview data into analyzable terms, in ...

  8. Bangdiwala's B - Wikipedia

    en.wikipedia.org/wiki/Bangdiwala's_B

    Bangdiwala's B statistic was created by Shrikant Bangdiwala in 1985 and is a measure of inter-rater agreement. [1] [2] While not as commonly used as the kappa statistic the B test has been used by various workers.

  9. Concordance correlation coefficient - Wikipedia

    en.wikipedia.org/wiki/Concordance_correlation...

    The concordance correlation coefficient is nearly identical to some of the measures called intra-class correlations.Comparisons of the concordance correlation coefficient with an "ordinary" intraclass correlation on different data sets found only small differences between the two correlations, in one case on the third decimal. [2]