enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Repeatability - Wikipedia

    en.wikipedia.org/wiki/Repeatability

    If the correlation between separate administrations of the test is high (e.g. 0.7 or higher as in this Cronbach's alpha-internal consistency-table [6]), then it has good testretest reliability. The repeatability coefficient is a precision measure which represents the value below which the absolute difference between two repeated test results ...

  3. Reliability (statistics) - Wikipedia

    en.wikipedia.org/wiki/Reliability_(statistics)

    Re-administering the same test to the same group at some later time; Correlating the first set of scores with the second; The correlation between scores on the first test and the scores on the retest is used to estimate the reliability of the test using the Pearson product-moment correlation coefficient: see also item-total correlation.

  4. Inter-rater reliability - Wikipedia

    en.wikipedia.org/wiki/Inter-rater_reliability

    A useful inter-rater reliability coefficient is expected (a) to be close to 0 when there is no "intrinsic" agreement and (b) to increase as the "intrinsic" agreement rate improves. Most chance-corrected agreement coefficients achieve the first objective. However, the second objective is not achieved by many known chance-corrected measures. [4]

  5. Cohen's kappa - Wikipedia

    en.wikipedia.org/wiki/Cohen's_kappa

    Cohen's kappa coefficient (κ, lowercase Greek kappa) is a statistic that is used to measure inter-rater reliability (and also intra-rater reliability) for qualitative (categorical) items. [1] It is generally thought to be a more robust measure than simple percent agreement calculation, as κ takes into account the possibility of the agreement ...

  6. Balanced Inventory of Desirable Responding - Wikipedia

    en.wikipedia.org/wiki/Balanced_Inventory_of...

    The correlation between the BIDR results and the MC-SDS was r=0.71, and the correlation between the BIDR and the Social Desirability Inventory [44] was r=0.80. The test-retest reliability of BIDR over a five-week follow-up period was 0.65-0.69. Numerous studies have demonstrated that BIDR has a good reliability. [13]

  7. Pearson correlation coefficient - Wikipedia

    en.wikipedia.org/.../Pearson_correlation_coefficient

    Pearson's correlation coefficient is the covariance of the two variables divided by the product of their standard deviations. The form of the definition involves a "product moment", that is, the mean (the first moment about the origin) of the product of the mean-adjusted random variables; hence the modifier product-moment in the name.

  8. Psychological statistics - Wikipedia

    en.wikipedia.org/wiki/Psychological_statistics

    This is most useful for subjective responses. Cohen's Kappa, Krippendorff's Alpha, Intra-Class correlation coefficients, Correlation coefficients, Kendal's concordance coefficient, etc. are useful statistical tools. (B) Test-Retest Reliability: Test-Retest Procedure is estimation of temporal consistency of the test.

  9. Kuder–Richardson formulas - Wikipedia

    en.wikipedia.org/wiki/Kuder–Richardson_formulas

    The name of this formula stems from the fact that is the twentieth formula discussed in Kuder and Richardson's seminal paper on test reliability. [1] It is a special case of Cronbach's α, computed for dichotomous scores. [2] [3] It is often claimed that a high KR-20 coefficient (e.g., > 0.90) indicates a homogeneous test. However, like ...