/
Annealing Paths for the Evaluation of Topic Models Annealing Paths for the Evaluation of Topic Models

Annealing Paths for the Evaluation of Topic Models - PowerPoint Presentation

lindy-dunigan
lindy-dunigan . @lindy-dunigan
Follow
348 views
Uploaded On 2018-09-23

Annealing Paths for the Evaluation of Topic Models - PPT Presentation

James Foulds Padhraic Smyth Department of Computer Science University of California Irvine James Foulds has recently moved to the University of California Santa Cruz Motivation Topic model extensions ID: 677267

model topic models ais topic model ais models ratio temperature

Share:

Link:

Embed:

Download Presentation from below link

Download Presentation The PPT/PDF document "Annealing Paths for the Evaluation of To..." is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

Annealing Paths for the Evaluation of Topic Models

James FouldsPadhraic SmythDepartment of Computer ScienceUniversity of California, Irvine*

*James

Foulds

has recently moved to the University of California, Santa CruzSlide2

Motivation

Topic model extensionsStructure, prior knowledge and constraintsSparse, nonparametric, correlated, tree-structured, time series, supervised, focused, determinantal…Special-purpose modelsAuthorship, scientific impact, political affiliation, conversational influence, networks, machine translation…General-purpose modelsDirichlet multinomial regression (DMR),

sparse additive generative (SAGE)…

Structural topic model (STM)

2Slide3

Motivation

Topic model extensions Structure, prior knowledge and constraintsSparse, nonparametric, correlated, tree-structured, time series, supervised, focused, determinantal…Special-purpose modelsAuthorship, scientific impact, political affiliation, conversational influence, networks, machine translation…General-purpose modelsDirichlet multinomial regression (DMR),

sparse additive generative (SAGE)…

Structural topic model (STM)

3Slide4

Motivation

Topic model extensionsStructure, prior knowledge and constraintsSparse, nonparametric, correlated, tree-structured, time series, supervised, focused, determinantal…Special-purpose modelsAuthorship, scientific impact, political affiliation, conversational influence, networks, machine translation…General-purpose modelsDirichlet multinomial regression (DMR),

sparse additive generative (SAGE), Structural topic model (STM), …

4Slide5

Motivation

Inference algorithms for topic modelsOptimizationEM, variational inference, collapsed variational inference,…SamplingCollapsed Gibbs sampling, Langevin dynamics,…

Scaling to ``big data’’Stochastic algorithms, distributed algorithms,

map reduce…

5Slide6

Motivation

Inference algorithms for topic modelsOptimizationEM, variational inference, collapsed variational inference,…SamplingCollapsed Gibbs sampling, Langevin dynamics,…

Scaling to ``big data’’Stochastic algorithms, distributed algorithms,

map reduce…

6Slide7

Motivation

Inference algorithms for topic modelsOptimizationEM, variational inference, collapsed variational inference,…SamplingCollapsed Gibbs sampling, Langevin

dynamics,…Scaling up to ``big data’’Stochastic algorithms, distributed algorithms,

map reduce, sparse data structures…

7Slide8

Motivation

Which existing techniques should we use?Is my new model/algorithm better than previous methods?8Slide9

Evaluating Topic Models

Training set

Test set

9Slide10

Evaluating Topic Models

Training set

Test set

Topic model

10Slide11

Evaluating Topic Models

Training set

Test set

Topic model

Predict:

11Slide12

Evaluating Topic Models

Training set

Test set

Topic model

Predict:

Log Pr

(

)

12Slide13

Evaluating Topic Models

Fitting these models only took a few hours on a

single

core

single core machine

.

C

reating this plot required a

cluster

13

(

Foulds

et al., 2013)Slide14

Why is this Difficult?

For every held-out document d, we need to estimateWe need to approximate possibly tens of thousands

of intractable sums/integrals!

14Slide15

Annealed Importance Sampling(Neal, 2001)

Scales up importance sampling to high dimensional data, using MCMCCorrects for MCMC convergence failures using importance weights15Slide16

Annealed Importance Sampling(Neal, 2001)

16

low “temperature”Slide17

Annealed Importance Sampling(Neal, 2001)

17

low “temperature”

high “temperature”Slide18

Annealed Importance Sampling(Neal, 2001)

18

low “temperature”

high “temperature”Slide19

Annealed Importance Sampling(Neal, 2001)

19Slide20

Annealed Importance Sampling(Neal, 2001)

Importance samples from the targetAn estimate of the ratio of partition functions

20Slide21

AIS for Evaluating Topic Models(Wallach et al., 2009)

21

Draw from the prior Anneal towards The posteriorSlide22

AIS for Evaluating Topic Models(Wallach et al., 2009)

22

Draw from the prior Anneal towards The posteriorSlide23

AIS for Evaluating Topic Models(Wallach et al., 2009)

23

Draw from the prior Anneal towards The posteriorSlide24

AIS for Evaluating Topic Models(Wallach et al., 2009)

24

Draw from the prior Anneal towards The posteriorSlide25

Insights

We are mainly interested in therelative performance of topic modelsAIS can provide estimates of the ratio of partition functions of any two distributions that we can anneal between

25Slide26

low “temperature”

high “temperature”

A standard application of Annealed

Importance

Sampling (

Neal

, 2001)

26Slide27

The Proposed Method:Ratio-AIS

Draw from Topic Model 2

Anneal towards Topic Model 1

27

medium “temperature”

medium “temperature”Slide28

The Proposed Method:Ratio-AIS

28

medium “temperature”

medium “temperature”

Draw from

Topic Model 2

Anneal towards

Topic Model 1Slide29

The Proposed Method:Ratio-AIS

29

medium “temperature”

medium “temperature”

Draw from

Topic Model 2

Anneal towards

Topic Model 1Slide30

Advantages of Ratio-AIS

Ratio-AIS avoids several sources of Monte Carlo error for comparing two models. The standard method estimates the denominator of a ratio even though it is a constant (=1),uses different z’s for both models,

and is run twice, introducing Monte Carlo noise each time.

An

easy

convergence check

: anneal in

the

reverse

direction

to compute the reciprocal.

30Slide31

Annealing Paths BetweenTopic Models

Geometric average of the two distributionsConvex combination of the parameters

31Slide32

Efficiently Plotting PerformancePer Iteration of the Learning Algorithm

32(Foulds et al., 2013)Slide33

Insights

Fsf2sfdWe can select the AIS intermediate distributions to be distributions of interestThe sequence of models we reach during training is typically amenable to annealing

The early models are often low temperatureEach successive model is similar to the previous one

33Slide34

Iteration-AIS

34Re-uses all previous computationWarm startsMore annealing temperatures, for freeImportance weights can be computed recursively

Anneal from Prior

Iteration

1

Iteration

2 …

Iteration

N

Wallach et al.

Ratio AIS

Ratio AIS

Ratio AIS

Topi

c Model at

Topi

c Model at

Topi

c Model atSlide35

Iteration-AIS

35Re-uses all previous computationWarm startsMore annealing temperatures, for freeImportance weights can be computed recursively

Anneal from Prior

Iteration

1

Iteration

2 …

Iteration

N

Wallach et al.

Ratio AIS

Ratio AIS

Ratio AIS

Topi

c Model at

Topi

c Model at

Topi

c Model atSlide36

Comparing Very Similar Topic Models (ACL Corpus)

36Slide37

Comparing Very Similar Topic Models (ACL and NIPS)

37% AccuracySlide38

Symmetric vs Asymmetric Priors(NIPS, 1000 temperatures or equiv.)

38

Correlation with longer left-to-right run

Variance of the estimate of relative log-likelihoodSlide39

Symmetric vs Asymmetric Priors(NIPS, 1000 temperatures or equiv.)

39

Correlation with longer left-to-right run

Variance of the estimate of relative log-likelihoodSlide40

Symmetric vs Asymmetric Priors(NIPS, 1000 temperatures or equiv.)

40

Correlation with longer left-to-right run

Variance of the estimate of relative log-likelihoodSlide41

Per-Iteration Evaluation, ACL Dataset

41Slide42

Per-Iteration Evaluation, ACL Dataset

42Slide43

Conclusions

Use Ratio-AIS for detailed document-level analysisRun the annealing in both directions to check for convergence failuresUse Left to Right for corpus-level analysisUse Iteration-AIS

to evaluate training algorithms

43Slide44

Future Directions

The ratio-AIS and iteration-AIS ideas can potentially be applied to other models with intractable likelihoods or partition functions (e.g. RBMs, ERGMs)Other annealing paths may be possibleEvaluating topic models remains an important, computationally challenging problem44Slide45

Thank You!

Questions?45