/
Bayesian inference,  Naïve Bayes model Bayesian inference,  Naïve Bayes model

Bayesian inference, Naïve Bayes model - PowerPoint Presentation

tatyana-admore
tatyana-admore . @tatyana-admore
Follow
407 views
Uploaded On 2018-03-19

Bayesian inference, Naïve Bayes model - PPT Presentation

httpxkcdcom1236 Bayes Rule The product rule gives us two ways to factor a joint probability Therefore Why is this useful Can update our beliefs about A based on evidence B PA is the ID: 657395

document class rain words class document words rain word bayes probability decision likelihood spam model

Share:

Link:

Embed:

Download Presentation from below link

Download Presentation The PPT/PDF document "Bayesian inference, Naïve Bayes model" is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

Bayesian inference, Naïve Bayes model

http://xkcd.com/1236/Slide2

Bayes Rule

The product rule gives us two ways to factor a joint probability:Therefore,

Why is this useful?

Can update our beliefs about A based on evidence B

P(A) is the prior and P(A|B) is the posteriorKey tool for probabilistic inference: can get diagnostic probability from causal probability E.g., P(Cavity = true | Toothache = true) from P(Toothache = true | Cavity = true)

Rev. Thomas

Bayes

(1702-1761)Slide3

Bayes Rule example

Marie is getting married tomorrow, at an outdoor ceremony in the desert. In recent years, it has rained only 5 days each year (5/365 = 0.014). Unfortunately, the weatherman has predicted rain for tomorrow. When it actually rains, the weatherman correctly forecasts rain 90% of the time. When it doesn't rain, he incorrectly forecasts rain 10% of the time. What is the probability that it will rain on Marie's wedding? Slide4

Bayes Rule example

Marie is getting married tomorrow, at an outdoor ceremony in the desert. In recent years, it has rained only 5 days each year (5/365 = 0.014). Unfortunately, the weatherman has predicted rain for tomorrow. When it actually rains, the weatherman correctly forecasts rain 90% of the time. When it doesn't rain, he incorrectly forecasts rain 10% of the time. What is the probability that it will rain on Marie's wedding? Slide5

Law of total probabilitySlide6

Bayes Rule example

Marie is getting married tomorrow, at an outdoor ceremony in the desert. In recent years, it has rained only 5 days each year (5/365 = 0.014). Unfortunately, the weatherman has predicted rain for tomorrow. When it actually rains, the weatherman correctly forecasts rain 90% of the time. When it doesn't rain, he incorrectly forecasts rain 10% of the time. What is the probability that it will rain on Marie's wedding? Slide7

Bayes rule: Example

1% of women at age forty who participate in routine screening have breast cancer.  80% of women with breast cancer will get positive mammographies.

9.6% of women without breast cancer will also get positive

mammographies

.  A woman in this age group had a positive mammography in a routine screening.  What is the probability that she actually has breast cancer?Slide8

https://

xkcd.com/1132/

See also:

https

://xkcd.com/882/Slide9

Probabilistic inference

Suppose the agent has to make a decision about the value of an unobserved query variable X given some observed evidence variable(s)

E = e

Partially observable, stochastic, episodic environmentExamples: X = {spam, not spam}, e = email messageX = {zebra, giraffe, hippo}, e = image featuresSlide10

Bayesian decision theory

Let x be the value predicted by the agent and x* be the true value of X. The agent has a loss function, which is 0 if x = x* and 1 otherwiseExpected loss for predicting x:

What is the estimate of X that minimizes the expected loss?

The one that has the greatest posterior probability

P(x|e)This is called the Maximum a Posteriori (MAP) decisionSlide11

MAP decision

Value x of X that has the highest posterior probability given the evidence E = e:

Maximum likelihood (ML) decision:

likelihood

prior

posteriorSlide12

Naïve Bayes model

Suppose we have many different types of observations (symptoms, features) E1

, …,

E

n that we want to use to obtain evidence about an underlying hypothesis XMAP decision:If each feature Ei can take on d values, how many entries are in the (conditional) joint probability table P(E1, …, En |

X = x)?Slide13

Naïve Bayes model

Suppose we have many different types of observations (symptoms, features) E1

, …,

E

n that we want to use to obtain evidence about an underlying hypothesis XMAP decision:We can make the simplifying assumption that the different features are conditionally independent given the hypothesis:If each feature can take on d values, what is the complexity of storing the resulting distributions?Slide14

Naïve Bayes model

Posterior:

MAP decision:

likelihood

prior

posteriorSlide15

Case study:Text document classification

MAP decision: assign a document to the class with the highest posterior P(class | document)

Example: spam classification

Classify

a message as spam if P(spam | message) > P(¬spam | message)Slide16

Case study:Text document classification

MAP decision: assign a document to the class with the highest posterior P(class | document)

We have

P(class | document)

 P(document | class)P(class) To enable classification, we need to be able to estimate the likelihoods P(document | class) for all classes and priors

P(class)Slide17

Naïve Bayes Representation

Goal: estimate likelihoods P(document |

class)

and priors

P(class)Likelihood: bag of words representationThe document is a sequence of words (w1, …, wn) The order of the words in the document is not importantEach word is conditionally independent of the others given document class Slide18

Naïve Bayes Representation

Goal: estimate likelihoods P(document |

class)

and priors

P(class)Likelihood: bag of words representationThe document is a sequence of words (w1, …, wn) The order of the words in the document is not importantEach word is conditionally independent of the others given document class Slide19

Bag of words illustration

US Presidential Speeches Tag Cloud

http://chir.ag/projects/preztags/Slide20

Bag of words illustration

US Presidential Speeches Tag Cloud

http://chir.ag/projects/preztags/Slide21

Bag of words illustration

US Presidential Speeches Tag Cloud

http://chir.ag/projects/preztags/Slide22

2016 convention speeches

Clinton

Trump

SourceSlide23

2016 first presidential debate

Trump

Clinton

SourceSlide24

2016 first presidential debate

Trump unique words

Clinton unique words

SourceSlide25

Naïve Bayes Representation

Goal: estimate likelihoods P(document |

class)

and

P(class)Likelihood: bag of words representationThe document is a sequence of words (w1, … , wn) The order of the words in the document is not importantEach word is conditionally independent of the others given document class Thus, the problem is reduced to estimating marginal likelihoods of individual words P(wi | class) Slide26

Parameter estimation

Model parameters: feature likelihoods P(word | class) and

priors

P(class) How do we obtain the values of these parameters? spam: 0.33¬spam: 0.67 P(word | ¬spam)

P(word | spam)priorSlide27

Parameter estimation

Model parameters: feature likelihoods P(word | class) and

priors

P(class) How do we obtain the values of these parameters?Need training set of labeled samples from both classesThis is the maximum likelihood (ML) estimate, or estimate that maximizes the likelihood of the training data: P(word | class) =

# of occurrences of this word in docs from this classtotal # of words in docs from this class

d

: index of training document,

i

: index of a wordSlide28

Parameter estimation

Parameter estimate:

Parameter smoothing: dealing with words that were never seen or seen too few times

Laplacian

smoothing: pretend you have seen every vocabulary word one more time than you actually did P(word | class) =# of occurrences of this word in docs from this class + 1total # of words in docs from this class + V

(V: total number of unique words)

P(word | class) =

# of occurrences of this word in docs from this classtotal # of words in docs from this classSlide29

Summary: Naïve Bayes for Document Classification

Assign the document to the class with the highest posterior Model parameters:

P

(class

1)

…P(

classK

)

P(w

1 | class

1)P

(

w

2

| class

1

)

P

(

w

n

| class

1

)

Likelihood

of class 1

prior

P

(

w

1

|

class

K

)

P

(

w

2

|

class

K

)

P

(

w

n

|

class

K

)

Likelihood

of class K

…Slide30

Summary: Naïve Bayes for Document Classification

Assign the document to the class with the highest posterior Note: by convention, one typically works with logs of probabilities instead:

Can help to avoid underflowSlide31

Prediction

Learning and inference pipeline

Training Labels

Training

Samples

Training

Learning

Features

Features

Inference

Test

Sample

Learned model

Learned modelSlide32

Review: Bayesian decision making

Suppose the agent has to make decisions about the value of an unobserved query variable X based on the values of an observed evidence variable

E

Inference problem:

given some evidence E = e, what is P(X | e)?Learning problem: estimate the parameters of the probabilistic model P(X | E) given a training sample {(x1,e1), …, (xn,en)}