Naive Bayes Classification
Naive Bayes Classification
Naive Bayes Classification
R
160615733108
4-2-CSE-B
STANLEY COLLEGE OF ENGINEERING AND
TECHNOLOGY
SEMINAR INCHARGE:MRS.D.RADHIKA
INTRODUCTION
All of Machine Learning Algorithms need to be
trained for supervised learning tasks like classification,
prediction etc .
By training it means to train them on particular inputs
so that later on we may test them for unknown inputs
for which they may classify or predict etc based on
their learning
In general , input set is divided into train set and test
set.
Naive Bayes is one of the method of supervised
learning.
Support Vector Machines, linear regression, logistic
regression are some other examples.
Naive Bayes classification is based on , the idea of
Conditional Probability and Bayes rule.
In Conditional probability we find the probability of
an event given that some event has already occurred .
In Bayes theorem, we find just the opposite, we find the
cause of some event that has already occurred.
Conditional Probability Formula:
P(A/B) = P(B/A).P(A)/P(B).
In reality, we have to predict an outcome given
multiple evidences. In that case, the math gets very
complicated.
So we have to 'uncouple' multiple pieces of evidence,
and treat each piece of evidence as independent.
This approach is called Naive Bayes classification.
When trying to classify, each outcome is called a class
and it has class label.
The class that has the highest probability is declared
the "winner" and that class label gets assigned to that
combination of evidences.
NAIVE BAYES FORMULA:
P(Outcome|MultipleEvidence)=P(Evidence1|Outcome
) x P(Evidence2|outcome) x ... x
P(EvidenceN|outcome) x P(Outcome) scaled by
P(Multiple Evidence).
P(outcome|evidence) = P(Likelihood of Evidence) x
Prior probability of outcome /P(Evidence)
ADVANTAGES
Very simple, easy to implement and fast.
Need less training data.
Highly scalable. It scales linearly with the any number
of events.
Can make probabilistic predictions.
Handles continuous and discrete data.
Not sensitive to irrelevant features.
APPLICATIONS
Categorizing news.
Email spam detection.
Face recognition.
Sentiment analysis.
Medical diagnosis.
Digit recognition .
Weather prediction .
CONCLUSION
The naive Bayes model is tremendously appealing
because of its simplicity, elegance, and robustness. It is one
of the oldest formal classification algorithms, and yet even
in its simplest form it is often surprisingly effective. It is
widely used in areas such as text classification and spam
filtering. A large number of modifications have been
introduced, by the statistical, data mining, machine
learning, and pattern recognition communities, in an
attempt to make it more flexible, but one has to recognize
that such modifications are necessarily complications,
which detract from its basic simplicity.
REFERENCES
C.D. Manning, P. Raghavan and H. Schütze (2008).
Introduction to Information Retrieval. Cambridge
University Press, pp. 234-265.
A. McCallum and K. Nigam (1998). A comparision of
event models for Naïve Bayes text classification. Proc.
AAAI/ICML-98 Workshop on Learning for Text
Categorization, pp. 41-48.
V. Metsis, I. Androutsopoulos and G. Paliouras
(2006). Spam filtering with Naive Bayes – Which Naive
Bayes? 3rd Conf. on Email and Anti-Spam (CEAS).