httpxkcdcom1236 Bayes Rule The product rule gives us two ways to factor a joint probability Therefore Why is this useful Can update our beliefs about A based on evidence B PA is the ID: 657395
Download Presentation The PPT/PDF document "Bayesian inference, Naïve Bayes model" is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.
Slide1
Bayesian inference, Naïve Bayes model
http://xkcd.com/1236/Slide2
Bayes Rule
The product rule gives us two ways to factor a joint probability:Therefore,
Why is this useful?
Can update our beliefs about A based on evidence B
P(A) is the prior and P(A|B) is the posteriorKey tool for probabilistic inference: can get diagnostic probability from causal probability E.g., P(Cavity = true | Toothache = true) from P(Toothache = true | Cavity = true)
Rev. Thomas
Bayes
(1702-1761)Slide3
Bayes Rule example
Marie is getting married tomorrow, at an outdoor ceremony in the desert. In recent years, it has rained only 5 days each year (5/365 = 0.014). Unfortunately, the weatherman has predicted rain for tomorrow. When it actually rains, the weatherman correctly forecasts rain 90% of the time. When it doesn't rain, he incorrectly forecasts rain 10% of the time. What is the probability that it will rain on Marie's wedding? Slide4
Bayes Rule example
Marie is getting married tomorrow, at an outdoor ceremony in the desert. In recent years, it has rained only 5 days each year (5/365 = 0.014). Unfortunately, the weatherman has predicted rain for tomorrow. When it actually rains, the weatherman correctly forecasts rain 90% of the time. When it doesn't rain, he incorrectly forecasts rain 10% of the time. What is the probability that it will rain on Marie's wedding? Slide5
Law of total probabilitySlide6
Bayes Rule example
Marie is getting married tomorrow, at an outdoor ceremony in the desert. In recent years, it has rained only 5 days each year (5/365 = 0.014). Unfortunately, the weatherman has predicted rain for tomorrow. When it actually rains, the weatherman correctly forecasts rain 90% of the time. When it doesn't rain, he incorrectly forecasts rain 10% of the time. What is the probability that it will rain on Marie's wedding? Slide7
Bayes rule: Example
1% of women at age forty who participate in routine screening have breast cancer. 80% of women with breast cancer will get positive mammographies.
9.6% of women without breast cancer will also get positive
mammographies
. A woman in this age group had a positive mammography in a routine screening. What is the probability that she actually has breast cancer?Slide8
https://
xkcd.com/1132/
See also:
https
://xkcd.com/882/Slide9
Probabilistic inference
Suppose the agent has to make a decision about the value of an unobserved query variable X given some observed evidence variable(s)
E = e
Partially observable, stochastic, episodic environmentExamples: X = {spam, not spam}, e = email messageX = {zebra, giraffe, hippo}, e = image featuresSlide10
Bayesian decision theory
Let x be the value predicted by the agent and x* be the true value of X. The agent has a loss function, which is 0 if x = x* and 1 otherwiseExpected loss for predicting x:
What is the estimate of X that minimizes the expected loss?
The one that has the greatest posterior probability
P(x|e)This is called the Maximum a Posteriori (MAP) decisionSlide11
MAP decision
Value x of X that has the highest posterior probability given the evidence E = e:
Maximum likelihood (ML) decision:
likelihood
prior
posteriorSlide12
Naïve Bayes model
Suppose we have many different types of observations (symptoms, features) E1
, …,
E
n that we want to use to obtain evidence about an underlying hypothesis XMAP decision:If each feature Ei can take on d values, how many entries are in the (conditional) joint probability table P(E1, …, En |
X = x)?Slide13
Naïve Bayes model
Suppose we have many different types of observations (symptoms, features) E1
, …,
E
n that we want to use to obtain evidence about an underlying hypothesis XMAP decision:We can make the simplifying assumption that the different features are conditionally independent given the hypothesis:If each feature can take on d values, what is the complexity of storing the resulting distributions?Slide14
Naïve Bayes model
Posterior:
MAP decision:
likelihood
prior
posteriorSlide15
Case study:Text document classification
MAP decision: assign a document to the class with the highest posterior P(class | document)
Example: spam classification
Classify
a message as spam if P(spam | message) > P(¬spam | message)Slide16
Case study:Text document classification
MAP decision: assign a document to the class with the highest posterior P(class | document)
We have
P(class | document)
P(document | class)P(class) To enable classification, we need to be able to estimate the likelihoods P(document | class) for all classes and priors
P(class)Slide17
Naïve Bayes Representation
Goal: estimate likelihoods P(document |
class)
and priors
P(class)Likelihood: bag of words representationThe document is a sequence of words (w1, …, wn) The order of the words in the document is not importantEach word is conditionally independent of the others given document class Slide18
Naïve Bayes Representation
Goal: estimate likelihoods P(document |
class)
and priors
P(class)Likelihood: bag of words representationThe document is a sequence of words (w1, …, wn) The order of the words in the document is not importantEach word is conditionally independent of the others given document class Slide19
Bag of words illustration
US Presidential Speeches Tag Cloud
http://chir.ag/projects/preztags/Slide20
Bag of words illustration
US Presidential Speeches Tag Cloud
http://chir.ag/projects/preztags/Slide21
Bag of words illustration
US Presidential Speeches Tag Cloud
http://chir.ag/projects/preztags/Slide22
2016 convention speeches
Clinton
Trump
SourceSlide23
2016 first presidential debate
Trump
Clinton
SourceSlide24
2016 first presidential debate
Trump unique words
Clinton unique words
SourceSlide25
Naïve Bayes Representation
Goal: estimate likelihoods P(document |
class)
and
P(class)Likelihood: bag of words representationThe document is a sequence of words (w1, … , wn) The order of the words in the document is not importantEach word is conditionally independent of the others given document class Thus, the problem is reduced to estimating marginal likelihoods of individual words P(wi | class) Slide26
Parameter estimation
Model parameters: feature likelihoods P(word | class) and
priors
P(class) How do we obtain the values of these parameters? spam: 0.33¬spam: 0.67 P(word | ¬spam)
P(word | spam)priorSlide27
Parameter estimation
Model parameters: feature likelihoods P(word | class) and
priors
P(class) How do we obtain the values of these parameters?Need training set of labeled samples from both classesThis is the maximum likelihood (ML) estimate, or estimate that maximizes the likelihood of the training data: P(word | class) =
# of occurrences of this word in docs from this classtotal # of words in docs from this class
d
: index of training document,
i
: index of a wordSlide28
Parameter estimation
Parameter estimate:
Parameter smoothing: dealing with words that were never seen or seen too few times
Laplacian
smoothing: pretend you have seen every vocabulary word one more time than you actually did P(word | class) =# of occurrences of this word in docs from this class + 1total # of words in docs from this class + V
(V: total number of unique words)
P(word | class) =
# of occurrences of this word in docs from this classtotal # of words in docs from this classSlide29
Summary: Naïve Bayes for Document Classification
Assign the document to the class with the highest posterior Model parameters:
P
(class
1)
…P(
classK
)
P(w
1 | class
1)P
(
w
2
| class
1
)
…
P
(
w
n
| class
1
)
Likelihood
of class 1
prior
P
(
w
1
|
class
K
)
P
(
w
2
|
class
K
)
…
P
(
w
n
|
class
K
)
Likelihood
of class K
…Slide30
Summary: Naïve Bayes for Document Classification
Assign the document to the class with the highest posterior Note: by convention, one typically works with logs of probabilities instead:
Can help to avoid underflowSlide31
Prediction
Learning and inference pipeline
Training Labels
Training
Samples
Training
Learning
Features
Features
Inference
Test
Sample
Learned model
Learned modelSlide32
Review: Bayesian decision making
Suppose the agent has to make decisions about the value of an unobserved query variable X based on the values of an observed evidence variable
E
Inference problem:
given some evidence E = e, what is P(X | e)?Learning problem: estimate the parameters of the probabilistic model P(X | E) given a training sample {(x1,e1), …, (xn,en)}