Neural Nets for NLP - Models with Latent Random Variables

Neural Nets for NLP - Models with Latent Random Variables

Graham Neubig via YouTube Direct link

Intro

1 of 23

1 of 23

Intro

Class Central Classrooms beta

YouTube playlists curated by Class Central.

Classroom Contents

Neural Nets for NLP - Models with Latent Random Variables

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Intro
  2. 2 Discriminative vs. Generative Models
  3. 3 Quiz: What Types of Variables?
  4. 4 What is Latent Random Variable Model
  5. 5 Why Latent Variable Models?
  6. 6 Deep Structured Latent Variable Models • Specify structure, but interpretable structure is often discrete e.g. POS tags, dependency parse trees
  7. 7 Examples of Deep Latent Variable Models
  8. 8 A probabilistic perspective on Variational Auto-Encoder
  9. 9 What is Our Loss Function?
  10. 10 Practice
  11. 11 Variational Inference • Variational inference approximates the true posterior poll with a family of distributions
  12. 12 Variational Inference • Variational inference approximates the true posterior polar with a family of distributions
  13. 13 Variational Auto-Encoders
  14. 14 Variational Autoencoders
  15. 15 Learning VAE
  16. 16 Problem! Sampling Breaks Backprop
  17. 17 Solution: Re-parameterization Trick
  18. 18 Difficulties in Training . Of the two components in the VAE objective, the KL divergence term is much easier to learn
  19. 19 Solution 3
  20. 20 Weaken the Decoder
  21. 21 Discrete Latent Variables?
  22. 22 Method 1: Enumeration
  23. 23 Solution 4

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.