\dm_csml_event_details
Speaker |
Zhenwen Dai |
---|---|
Affiliation |
University of Sheffield |
Date |
Friday, 15 May 2015 |
Time |
13:00-14:00 |
Location |
Zoom |
Link |
Roberts 508 (different than usual) |
Event series |
Jump Trading/ELLIS CSML Seminar Series |
Abstract |
Deep latent variable models are promising for unsupervised and semi-supervised learning, however, the development of models with continuous latent variables are left behind. We scale up a deep continuous latent variable model called a hierarchical community of experts. It contains a hierarchy of linear-Gaussian units and a mechanism for dynamically selecting an subset of these units. We derive a new variational lower bound that only needs the estimation of the variational posterior at the top layer and use a probabilistic generative model for approximating such variational posterior by directly generating samples given the inputs, in which variance reduction techniques are not necessary. We verify our new variational bound and our inference generative model by applying to SBN, and compare the performance with the literature on the MNIST dataset. With training HCE on MNIST, we show that it is able to capture sophisticated variances of characters in generated covariance matrices. |
Biography |