/
Bayesian Learning  By  Porchelvi Vijayakumar Bayesian Learning  By  Porchelvi Vijayakumar

Bayesian Learning By Porchelvi Vijayakumar - PowerPoint Presentation

desha
desha . @desha
Follow
342 views
Uploaded On 2022-06-08

Bayesian Learning By Porchelvi Vijayakumar - PPT Presentation

Cognitive Science Current Problem How do children learn and how do they get it right Connectionists and Associationists Associationism maintains that all knowledge is represented in terms of associations between ideas that complex ideas are built up from combinations of more primit ID: 915589

probability bayesian causal learning bayesian probability learning causal intervention graph variables conditional mutilated independence networks set network cognitive markov

Share:

Link:

Embed:

Download Presentation from below link

Download Presentation The PPT/PDF document "Bayesian Learning By Porchelvi Vijayak..." is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

Bayesian Learning

By

Porchelvi Vijayakumar

Slide2

Cognitive Science

Current Problem:

How do children learn and how do they get it right?

Slide3

Connectionists and Associationists

Associationism:

maintains that all knowledge is represented in terms of associations between ideas, that complex ideas are built up from combinations of more primitive ideas, which, in accordance with empiricist philosophy, are ultimately derived from the senses.

Connectionism :

is a more powerful associationist theory than its predecessors (Shanks, 1995), that seeks to model cognitive processes in a way that broadly reflects the computational style of the brain.

Slide4

Developmental Scientists

Developmental scientists believe that behavior is both abstract representation and

learning

– Inductive learning

Slide5

How do we reason?

Pure Logic

Reasoning with Beliefs (probability)

Taken From: http://www.dgp.toronto.edu/~hertzman/ibl2004

Associationists and Connectionists

Developmental Cognitive Scientists

Slide6

Pure Logic

Pure Logic:

If A is TRUE the B is also TRUE.

A: My car isn’t where I left it.

B: My car was

stolen

Taken From: http://www.dgp.toronto.edu/~hertzman/ibl2004

Slide7

Introduction to Bayesian Network

Basics:

Probability, Joint Probability, Conditional Probability.

Bayes Law

Markov Condition

Slide8

Conditional Probability, Independence

Conditional Probability

P(E|F) = P( E AND F)/ P(F)

We know that the

P(E AND F) = P(E) * P(F) when E and F are independent.

Independence

:

P(E|F) = P(E)

Conditional Independence:

P(E| F AND G) = P( E|G)

Slide9

Bayes’ Theorem

Inference :

P(E| F) =

P(F|E) * P(E)

P(F)

Likelihood

Prior Probability

Marginal Probability

Posterior Probability

Slide10

Bayesian Network

Bayesian Net:

DAG - Directed Acyclic Graph which satisfies

Markov Condition.

Nodes - Variable in the Causal System.

Edges – direct influence.

p(h1)

p(b1/h1) p(L1/h1)

p(f1|b1,l1) p(c1|l1)

From: Learning Bayesian Networks by Richard E. Neapolitan

B

L

F

H

C

Slide11

Markov Condition: If for each variable X € V {X} is conditionally independent of the set of all its non descendents, given the set of all its parents.

Bayesian Network

Slide12

Patterns in Causal Chain

A B C D

= Markov Equivalent

A B C D

These two chains have

same pattern of dependence and conditional probability.

Slide13

Learning Causal Bayesian Networks

provides an account for Inductive Inference.

defines a

Joint Probability Distribution

– thereby specifying how likely is any joint settings of the variables.

can be used to

predict

about the

variables

when the graph structure is known.

can be used to learn the graph structure when it is un know, by observing the settings of the variables tend to occur together more or less often.

Slide14

Intervention Mutilated Graph

Intervention

on particular variable

X

changes probabilistic dependencies over all the variables in the network.

Two networks that would otherwise imply identical patterns of probabilistic dependence may become distinguishable under intervention.

Mutilated Graph

in which all incoming arrows to

X

are cut.

Slide15

Intervention and mutilated Graph

A B C D =

P

attern

before intervention

A B C D

= Muti

lated

graph

A B C D = Pattern before intervention

A B C D = Mutilated graph

Thus two chains

which had similar patters of dependencies are different from each other after intervention.

This is constraint based learning

Slide16

Intervention and mutilated Graph

These algorithms can work backward to figure out the set of causal structure compatible with the constraints of the evidence. Given the observed patterns of independence and conditional independence among a set of variables perhaps under different conditions of interventions.

Slide17

Bayesian Learning

Human inclined tend to judge one causal structure more likely than another.

This degree of believe may be strongly influenced by prior expectations about which causal structures are more likely.

Example: People know Causal mechanism at work

Slide18

Bayesian Learning

H

- A space of possible causal models

d

– Some data - observations of the states of one or more variables in the causal system for different cases, individuals or situations.

P(

h|d

)

= posterior probability distribution.

P(h|d) =

Slide19

Conclusion

• Posterior probabilities

– Probability of any event given any evidence

• Most likely explanation

– Scenario that explains evidence

• Rational decision making

– Maximize expected utility

– Value of Information

• Effect of intervention

– Causal

analysis . Bayesian model may be traditionally been limited by a focus on learning representations at only a single level of abstraction.

Slide20

Referenceshttp://www.dgp.toronto.edu/~hertzman/ibl2004

Learning Bayesian Networks – by Richard E. Neapolitan

Bayesian networks, Bayesian Learning and Cognitive Development.