enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Kolmogorov–Smirnov test - Wikipedia

    en.wikipedia.org/wiki/KolmogorovSmirnov_test

    Illustration of the KolmogorovSmirnov statistic. The red line is a model CDF, the blue line is an empirical CDF, and the black arrow is the KS statistic.. In statistics, the KolmogorovSmirnov test (also K–S test or KS test) is a nonparametric test of the equality of continuous (or discontinuous, see Section 2.2), one-dimensional probability distributions.

  3. Two-sample hypothesis testing - Wikipedia

    en.wikipedia.org/wiki/Two-sample_hypothesis_testing

    In statistical hypothesis testing, a two-sample test is a test performed on the data of two random samples, each independently obtained from a different given population. The purpose of the test is to determine whether the difference between these two populations is statistically significant .

  4. List of statistical tests - Wikipedia

    en.wikipedia.org/wiki/List_of_statistical_tests

    Unpaired samples are also called independent samples. Paired samples are also called dependent. Finally, there are some statistical tests that perform analysis of relationship between multiple variables like regression. [1] Number of samples: The number of samples of data. Exactness: A test can be exact or be asymptotic delivering approximate ...

  5. Nonparametric statistics - Wikipedia

    en.wikipedia.org/wiki/Nonparametric_statistics

    KolmogorovSmirnov test: tests whether a sample is drawn from a given distribution, or whether two samples are drawn from the same distribution. Kruskal–Wallis one-way analysis of variance by ranks: tests whether > 2 independent samples are drawn from the same distribution.

  6. Goodness of fit - Wikipedia

    en.wikipedia.org/wiki/Goodness_of_fit

    Such measures can be used in statistical hypothesis testing, e.g. to test for normality of residuals, to test whether two samples are drawn from identical distributions (see KolmogorovSmirnov test), or whether outcome frequencies follow a specified distribution (see Pearson's chi-square test).

  7. Normality test - Wikipedia

    en.wikipedia.org/wiki/Normality_test

    KolmogorovSmirnov test: this test only works if the mean and the variance of the normal distribution are assumed known under the null hypothesis, Lilliefors test: based on the KolmogorovSmirnov test, adjusted for when also estimating the mean and variance from the data, Shapiro–Wilk test, and; Pearson's chi-squared test.

  8. Cumulative distribution function - Wikipedia

    en.wikipedia.org/wiki/Cumulative_distribution...

    The KolmogorovSmirnov test is based on cumulative distribution functions and can be used to test to see whether two empirical distributions are different or whether an empirical distribution is different from an ideal distribution. The closely related Kuiper's test is useful if the domain of the distribution is cyclic as in day of the week ...

  9. Lilliefors test - Wikipedia

    en.wikipedia.org/wiki/Lilliefors_test

    Lilliefors test is a normality test based on the KolmogorovSmirnov test.It is used to test the null hypothesis that data come from a normally distributed population, when the null hypothesis does not specify which normal distribution; i.e., it does not specify the expected value and variance of the distribution. [1]