You are on page 1of 10

BY:POOJA.

R
160615733108
4-2-CSE-B
STANLEY COLLEGE OF ENGINEERING AND
TECHNOLOGY
SEMINAR INCHARGE:MRS.D.RADHIKA
INTRODUCTION
 All of Machine Learning Algorithms need to be
trained for supervised learning tasks like classification,
prediction etc .
 By training it means to train them on particular inputs
so that later on we may test them for unknown inputs
for which they may classify or predict etc based on
their learning
 In general , input set is divided into train set and test
set.
 Naive Bayes is one of the method of supervised
learning.
 Support Vector Machines, linear regression, logistic
regression are some other examples.
 Naive Bayes classification is based on , the idea of
Conditional Probability and Bayes rule.
 In Conditional probability we find the probability of
an event given that some event has already occurred .
 In Bayes theorem, we find just the opposite, we find the
cause of some event that has already occurred.
 Conditional Probability Formula:

If A and B are two dependent events such that


given event B has happened and we have to find the
probability of event A, then probability of A given B,
denoted by P(A/B) is given by
P(A/B) = P(A∩B)/P(B).

 Formula for Bayes’ Theorem:


If A and B are two independent events such that probability of event
A is denoted by,

P(A/B) = P(B/A).P(A)/P(B).
 In reality, we have to predict an outcome given
multiple evidences. In that case, the math gets very
complicated.
 So we have to 'uncouple' multiple pieces of evidence,
and treat each piece of evidence as independent.
 This approach is called Naive Bayes classification.
 When trying to classify, each outcome is called a class
and it has class label.
 The class that has the highest probability is declared
the "winner" and that class label gets assigned to that
combination of evidences.
NAIVE BAYES FORMULA:
 P(Outcome|MultipleEvidence)=P(Evidence1|Outcome
) x P(Evidence2|outcome) x ... x
P(EvidenceN|outcome) x P(Outcome) scaled by
P(Multiple Evidence).
 P(outcome|evidence) = P(Likelihood of Evidence) x
Prior probability of outcome /P(Evidence)
ADVANTAGES
 Very simple, easy to implement and fast.
 Need less training data.
 Highly scalable. It scales linearly with the any number
of events.
 Can make probabilistic predictions.
 Handles continuous and discrete data.
 Not sensitive to irrelevant features.
APPLICATIONS
 Categorizing news.
 Email spam detection.
 Face recognition.
 Sentiment analysis.
 Medical diagnosis.
 Digit recognition .
 Weather prediction .
CONCLUSION
The naive Bayes model is tremendously appealing
because of its simplicity, elegance, and robustness. It is one
of the oldest formal classification algorithms, and yet even
in its simplest form it is often surprisingly effective. It is
widely used in areas such as text classification and spam
filtering. A large number of modifications have been
introduced, by the statistical, data mining, machine
learning, and pattern recognition communities, in an
attempt to make it more flexible, but one has to recognize
that such modifications are necessarily complications,
which detract from its basic simplicity.
REFERENCES
 C.D. Manning, P. Raghavan and H. Schütze (2008).
Introduction to Information Retrieval. Cambridge
University Press, pp. 234-265.
 A. McCallum and K. Nigam (1998). A comparision of
event models for Naïve Bayes text classification. Proc.
AAAI/ICML-98 Workshop on Learning for Text
Categorization, pp. 41-48.
 V. Metsis, I. Androutsopoulos and G. Paliouras
(2006). Spam filtering with Naive Bayes – Which Naive
Bayes? 3rd Conf. on Email and Anti-Spam (CEAS).

You might also like