enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Multiple comparisons problem - Wikipedia

    en.wikipedia.org/wiki/Multiple_comparisons_problem

    Multiple comparisons arise when a statistical analysis involves multiple simultaneous statistical tests, each of which has a potential to produce a "discovery". A stated confidence level generally applies only to each test considered individually, but often it is desirable to have a confidence level for the whole family of simultaneous tests. [4]

  3. Reliability (statistics) - Wikipedia

    en.wikipedia.org/wiki/Reliability_(statistics)

    The correlation between scores on the two alternate forms is used to estimate the reliability of the test. This method provides a partial solution to many of the problems inherent in the test-retest reliability method. For example, since the two forms of the test are different, carryover effect is less of a problem. Reactivity effects are also ...

  4. Repeatability - Wikipedia

    en.wikipedia.org/wiki/Repeatability

    If the correlation between separate administrations of the test is high (e.g. 0.7 or higher as in this Cronbach's alpha-internal consistency-table [6]), then it has good testretest reliability. The repeatability coefficient is a precision measure which represents the value below which the absolute difference between two repeated test results ...

  5. Cohen's kappa - Wikipedia

    en.wikipedia.org/wiki/Cohen's_kappa

    Cohen's kappa measures the agreement between two raters who each classify N items into C mutually exclusive categories. The definition of is =, where p o is the relative observed agreement among raters, and p e is the hypothetical probability of chance agreement, using the observed data to calculate the probabilities of each observer randomly selecting each category.

  6. List of statistical tests - Wikipedia

    en.wikipedia.org/wiki/List_of_statistical_tests

    Statistical tests are used to test the fit between a hypothesis and the data. [ 1 ] [ 2 ] Choosing the right statistical test is not a trivial task. [ 1 ] The choice of the test depends on many properties of the research question.

  7. Inter-rater reliability - Wikipedia

    en.wikipedia.org/wiki/Inter-rater_reliability

    In statistics, inter-rater reliability (also called by various similar names, such as inter-rater agreement, inter-rater concordance, inter-observer reliability, inter-coder reliability, and so on) is the degree of agreement among independent observers who rate, code, or assess the same phenomenon.

  8. Page's trend test - Wikipedia

    en.wikipedia.org/wiki/Page's_trend_test

    In statistics, the Page test for multiple comparisons between ordered correlated variables is the counterpart of Spearman's rank correlation coefficient which summarizes the association of continuous variables. It is also known as Page's trend test or Page's L test. It is a repeated measure trend test. The Page test is useful where:

  9. Psychological statistics - Wikipedia

    en.wikipedia.org/wiki/Psychological_statistics

    This is most useful for subjective responses. Cohen's Kappa, Krippendorff's Alpha, Intra-Class correlation coefficients, Correlation coefficients, Kendal's concordance coefficient, etc. are useful statistical tools. (B) Test-Retest Reliability: Test-Retest Procedure is estimation of temporal consistency of the test.