/
Riemannian Normalizing Flow on Variational Wasserstein Autoencoder for Text Modeling Riemannian Normalizing Flow on Variational Wasserstein Autoencoder for Text Modeling

Riemannian Normalizing Flow on Variational Wasserstein Autoencoder for Text Modeling - PowerPoint Presentation

atomexxon
atomexxon . @atomexxon
Follow
345 views
Uploaded On 2020-06-25

Riemannian Normalizing Flow on Variational Wasserstein Autoencoder for Text Modeling - PPT Presentation

Prince Wang William Wang UC Santa Barbara Outline VAE and the KL vanishing problem Motivation why Riemannian Normalizing flowWAE Details Experimental Results VAE KL vanishing ID: 786585

wae flow wasserstein normalizing flow wae normalizing wasserstein riemannian latent rnf results 2018 space curvature https vanishing vae modeling

Share:

Link:

Embed:

Download Presentation from below link

Download The PPT/PDF document "Riemannian Normalizing Flow on Variation..." is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

Riemannian Normalizing Flow on Variational Wasserstein Autoencoder for Text Modeling

Prince Wang, William WangUC Santa Barbara

‹#›

Slide2

Outline

VAE and the KL vanishing problemMotivation: why Riemannian Normalizing flow/WAE

Details

Experimental Results

‹#›

Slide3

VAE: KL vanishing

KL term, gap between posterior and prior

‹#›

Can generate sentences given latent codes z

i were to buy any groceries .

horses are to buy any groceries .

horses are to buy any animal .

horses the favorite any animal .

Slide4

Previous works

Generating sentences from Continuous Space, (2015, Bowman)Improved Variational Autoencoder for text Modeling using Dilated Convolution, (2017, Yang)Spherical Latent Spaces for Stable Variational Autoencoder, (2018, Xu)

Semi-Amortized VAE, (2018, Kim)

Cyclical Annealing Schedule: A Simple Approach to Mitigating KL Vanishing, (2019, Fu)

‹#›

Slide5

Riemannian Normalizing Flow/Wasserstein Distance

‹#›

Slide6

Normalizing Flow

https://lilianweng.github.io/lil-log/2018/10/13/flow-based-deep-generative-models.html

‹#›

Making posterior harder to collapse to a standard Gaussian prior

Slide7

Normalizing Flow

Tighter likelihood approximation

‹#›

Reconstruction

KL

Jacobian

Slide8

Why Riemannian VAE?

The Latent space is not flat Euclidean. It should be curved.

‹#›

Slide9

‹#›

Riemannian Metric

Jacobian

Rie. Metric

Slide10

Match latent manifold with input manifold

‹#›

Curve

Length

Slide11

Modeling curvature by NF

Planar Flow

Curvature:

‹#›

Slide12

Modeling curvature by NF

To match geometry of latent space with input space, we need this determinant to be large when input manifold has high curvature

Jacobian

‹#›

Slide13

Wasserstein Distance

Replace KL with Maximum Mean Discrepancy (MMD)

Wasserstein Autoencoder, (ICLR 2018,

Ilya Tolstikhin)

‹#›

Slide14

Wasserstein RNF

ReconstructionMMD loss

KLD

loss with NF

‹#›

Slide15

‹#›

Results

Language Models: Negative Log-likelihood/KL/Perplexity

Slide16

Results: KL divergence

‹#›

PTB

Yelp

Slide17

Results: Negative log-likelihood

‹#›PTB

Yelp

WAE

WAE-NF

WAE-RNF

WAE

WAE-NF

WAE-RNF

104

92

91

198

184

183

Slide18

Mutual Information

Mutual information

‹#›

Slide19

Conclusion

Propose to use Normalizing Flow and Wasserstein Distance for variational language modelDesign Riemannian Normalizing Flow to learn a smooth latent spaceEmpirical results indicate that Riemannian Normalizing Flow with Wasserstein Distance help avert KL vanishing

Code:

https://github.com/kingofspace0wzz/wae-rnf-lm

‹#›

Slide20

‹#›

Thank you! Q & A :)

Code:

https://github.com/kingofspace0wzz/wae-rnf-lm

Paper:

https://arxiv.org/abs/1904.02399