+ All Categories
Home > Documents > Naïve Bayes Classifier

Naïve Bayes Classifier

Date post: 25-Feb-2016
Category:
Upload: nieve
View: 51 times
Download: 0 times
Share this document with a friend
Description:
Naïve Bayes Classifier . Ke Chen http://intranet.cs.man.ac.uk/mlo/comp20411/ Extended by Longin Jan Latecki [email protected]. COMP20411 Machine Learning. Outline. Background Probability Basics Probabilistic Classification Na ï ve Bayes Example: Play Tennis Relevant Issues - PowerPoint PPT Presentation
18
Naïve Bayes Classifier Ke Chen http://intranet.cs.man.ac.uk/mlo/comp20411 / Extended by Longin Jan Latecki [email protected] COMP20411 Machine Learning
Transcript
Page 1: Naïve Bayes Classifier

Naïve Bayes Classifier Ke Chen

http://intranet.cs.man.ac.uk/mlo/comp20411/

Extended by Longin Jan [email protected]

COMP20411 Machine Learning

Page 2: Naïve Bayes Classifier

COMP20411 Machine Learning 2

Outline

• Background• Probability Basics• Probabilistic Classification• Naïve Bayes • Example: Play Tennis• Relevant Issues• Conclusions

Page 3: Naïve Bayes Classifier

COMP20411 Machine Learning 3

Background• There are three methods to establish a classifier a) Model a classification rule directly

Examples: k-NN, decision trees, perceptron, SVM b) Model the probability of class memberships given input data Example: multi-layered perceptron with the cross-entropy cost

c) Make a probabilistic model of data within each class Examples: naive Bayes, model based classifiers

• a) and b) are examples of discriminative classification• c) is an example of generative classification• b) and c) are both examples of probabilistic

classification

Page 4: Naïve Bayes Classifier

COMP20411 Machine Learning 4

Probability Basics

• Prior, conditional and joint probability

– Prior probability: – Conditional probability: – Joint probability: – Relationship:– Independence:

• Bayesian Rule

)| ,)( 121 XP(XX|XP 2

)()()( )(

XXXP

CPC|P|CP

)(XP

) )( ),,( 22 ,XP(XPXX 11 XX)()|()()|() 2211122 XPXXPXPXXP,XP(X1

)()() ),()|( ),()|( 212121212 XPXP,XP(XXPXXPXPXXP 1

EvidencePriorLikelihoodPosterior

Page 5: Naïve Bayes Classifier

Example by Dieter Fox

Page 6: Naïve Bayes Classifier
Page 7: Naïve Bayes Classifier
Page 8: Naïve Bayes Classifier

COMP20411 Machine Learning 8

Probabilistic Classification

• Establishing a probabilistic model for classification

– Discriminative model

– Generative model • MAP classification rule

– MAP: Maximum A Posterior– Assign x to c* if

• Generative classification with the MAP rule– Apply Bayesian rule to convert:

),, , )( 1 n1L X(Xc,,cC|CP XX

),, , )( 1 n1L X(Xc,,cCC|P XX

Lc,,cccc|cCP|cCP 1** , )( )( xXxX

)()()(

)()( )( CPC|PP

CPC|P|CP XX

XX

Page 9: Naïve Bayes Classifier

Feature Histograms

x

C1 C2

P(x)

Slide by Stephen Marsland

Page 10: Naïve Bayes Classifier

Posterior Probability

x

P(C|x)

1

0Slide by Stephen Marsland

Page 11: Naïve Bayes Classifier

COMP20411 Machine Learning 11

Naïve Bayes

• Bayes classification

Difficulty: learning the joint probability • Naïve Bayes classification

– Making the assumption that all input attributes are independent

– MAP classification rule

)()|,,()()( )( 1 CPCXXPCPC|P|CP n XX

)|,,( 1 CXXP n

)|()|()|( )|,,()|(

)|,,();,,|()|,,,(

21

21

22121

CXPCXPCXPCXXPCXP

CXXPCXXXPCXXXP

n

n

nnn

Lnn ccccccPcxPcxPcPcxPcxP ,, , ),()]|()|([)()]|()|([ 1*

1***

1

Page 12: Naïve Bayes Classifier

COMP20411 Machine Learning 12

Naïve Bayes

• Naïve Bayes Algorithm (for discrete input attributes)

– Learning Phase: Given a training set S,

Output: conditional probability tables; for elements

– Test Phase: Given an unknown instance , Look up tables to assign the label c* to X’ if

; in examples with)|( estimate)|(̂

),1 ;,,1( attribute each of value attribute every For ; in examples with)( estimate)(̂

of value target each For 1

S

S

ijkjijkj

jjjk

ii

Lii

cCaXPcCaXP

N,knj xacCPcCP

)c,,c(c c

Lnn ccccccPcaPcaPcPcaPcaP ,, , ),(̂)]|(̂)|(̂[)(̂)]|(̂)|(̂[ 1*

1***

1

),,( 1 naa XLNx jj ,

Page 13: Naïve Bayes Classifier

COMP20411 Machine Learning 13

Example

• Example: Play Tennis

Page 14: Naïve Bayes Classifier

COMP20411 Machine Learning 14

Example

• Learning Phase

Outlook Play=Yes

Play=No

Sunny 2/9 3/5Overcas

t4/9 0/5

Rain 3/9 2/5

Temperature

Play=Yes Play=No

Hot 2/9 2/5Mild 4/9 2/5Cool 3/9 1/5

Humidity Play=Yes

Play=No

High 3/9 4/5Normal 6/9 1/5

Wind Play=Yes

Play=No

Strong 3/9 3/5Weak 6/9 2/5

P(Play=Yes) = 9/14P(Play=No) = 5/14

Page 15: Naïve Bayes Classifier

COMP20411 Machine Learning 15

Example

• Test Phase

– Given a new instance, x’=(Outlook=Sunny, Temperature=Cool, Humidity=High,

Wind=Strong)– Look up tables

– MAP rule

P(Outlook=Sunny|Play=No) = 3/5P(Temperature=Cool|Play==No) = 1/5P(Huminity=High|Play=No) = 4/5P(Wind=Strong|Play=No) = 3/5P(Play=No) = 5/14

P(Outlook=Sunny|Play=Yes) = 2/9P(Temperature=Cool|Play=Yes) = 3/9P(Huminity=High|Play=Yes) = 3/9P(Wind=Strong|Play=Yes) = 3/9P(Play=Yes) = 9/14

P(Yes|x’): [P(Sunny|Yes)P(Cool|Yes)P(High|Yes)P(Strong|Yes)]P(Play=Yes) = 0.0053 P(No|x’): [P(Sunny|No) P(Cool|No)P(High|No)P(Strong|No)]P(Play=No) = 0.0206

Given the fact P(Yes|x’) < P(No|x’), we label x’ to be “No”.

Page 16: Naïve Bayes Classifier

COMP20411 Machine Learning 16

Relevant Issues

• Violation of Independence Assumption

– For many real world tasks,– Nevertheless, naïve Bayes works surprisingly well

anyway!• Zero conditional probability Problem

– If no example contains the attribute value– In this circumstance, during test – For a remedy, conditional probabilities estimated with

)|()|( )|,,( 11 CXPCXPCXXP nn

0)|(̂ , ijkjjkj cCaXPaX0)|(̂)|(̂)|(̂ 1 inijki cxPcaPcxP

)1 examples, virtual"" of (number prior to weight:) of values possible for /1 (usually, estimate prior :

whichfor examples training of number : C and whichfor examples training of number :

)|(̂

mmXttpp

cCncaXn

mnmpncCaXP

j

i

ijkjc

cijkj

Page 17: Naïve Bayes Classifier

COMP20411 Machine Learning 17

Relevant Issues

• Continuous-valued Input Attributes

– Numberless values for an attribute – Conditional probability modeled with the normal

distribution

– Learning Phase: Output: normal distributions and – Test Phase:

• Calculate conditional probabilities with all the normal distributions• Apply the MAP rule to make a decision

ijji

ijji

ji

jij

jiij

cCcX

XcCXP

whichfor examples of X values attribute of deviation standard :C whichfor examples of values attribute of (avearage) mean :

2)(

exp21)|(̂ 2

2

Ln ccCXX ,, ),,,( for 11 XLn

),,( for 1 nXX XLicCP i ,,1 )(

Page 18: Naïve Bayes Classifier

COMP20411 Machine Learning 18

Conclusions• Naïve Bayes based on the independence assumption

– Training is very easy and fast; just requiring considering each attribute in each class separately

– Test is straightforward; just looking up tables or calculating conditional probabilities with normal distributions

• A popular generative model– Performance competitive to most of state-of-the-art

classifiers even in presence of violating independence assumption

– Many successful applications, e.g., spam mail filtering– Apart from classification, naïve Bayes can do more…


Recommended