enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Naive Bayes classifier - Wikipedia

    en.wikipedia.org/wiki/Naive_Bayes_classifier

    Example of a naive Bayes classifier depicted as a Bayesian Network. In statistics, naive Bayes classifiers are a family of linear "probabilistic classifiers" which assumes that the features are conditionally independent, given the target class. The strength (naivety) of this assumption is what gives the classifier its name.

  3. Bayes classifier - Wikipedia

    en.wikipedia.org/wiki/Bayes_classifier

    In statistical classification, the Bayes classifier is the classifier having the smallest probability of misclassification of all classifiers using the same set of features. [ 1 ] Definition

  4. Bayes error rate - Wikipedia

    en.wikipedia.org/wiki/Bayes_error_rate

    This statistics -related article is a stub. You can help Wikipedia by expanding it.

  5. Bayes' theorem - Wikipedia

    en.wikipedia.org/wiki/Bayes'_theorem

    Price edited [3] Bayes's major work "An Essay Towards Solving a Problem in the Doctrine of Chances" (1763), which appeared in Philosophical Transactions, [4] and contains Bayes' theorem. Price wrote an introduction to the paper that provides some of the philosophical basis of Bayesian statistics and chose one of the two solutions Bayes offered.

  6. Naive Bayes spam filtering - Wikipedia

    en.wikipedia.org/wiki/Naive_Bayes_spam_filtering

    Naive Bayes spam filtering is a baseline technique for dealing with spam that can tailor itself to the email needs of individual users and give low false positive spam detection rates that are generally acceptable to users. It is one of the oldest ways of doing spam filtering, with roots in the 1990s.

  7. Probabilistic classification - Wikipedia

    en.wikipedia.org/wiki/Probabilistic_classification

    Binary probabilistic classifiers are also called binary regression models in statistics. In econometrics, probabilistic classification in general is called discrete choice. Some classification models, such as naive Bayes, logistic regression and multilayer perceptrons (when trained under an appropriate loss function) are

  8. Generative model - Wikipedia

    en.wikipedia.org/wiki/Generative_model

    Standard examples of each, all of which are linear classifiers, are: generative classifiers: naive Bayes classifier and; linear discriminant analysis; discriminative model: logistic regression; In application to classification, one wishes to go from an observation x to a label y (or probability distribution on labels

  9. Discriminative model - Wikipedia

    en.wikipedia.org/wiki/Discriminative_model

    In the repeated experiments, logistic regression and naive Bayes are applied here for different models on binary classification task, discriminative learning results in lower asymptotic errors, while generative one results in higher asymptotic errors faster. [3]