Publication:
But What Did You Actually Learn? Improving Inference for Non-Identifiable Deep Latent Variable Models

No Thumbnail Available

Date

2023-05-12

Published Version

Published Version

Journal Title

Journal ISSN

Volume Title

Publisher

The Harvard community has made this article openly available. Please share how this access benefits you.

Research Projects

Organizational Units

Journal Issue

Citation

Yacoby, Yaniv. 2023. But What Did You Actually Learn? Improving Inference for Non-Identifiable Deep Latent Variable Models. Doctoral dissertation, Harvard University Graduate School of Arts and Sciences.

Research Data

Abstract

Deep probabilistic and Bayesian latent variable models allow one to infer variables that have previously not been observed in the data in order to accurately model the data density. They provide an intuitive and flexible modeling paradigm, making them particularly well-suited for safety-critical applications, in which the data is often noisy and complex, requiring expressivity from deep learning, and assumptions need to be well-understood and easily interrogated to ensure safety, requiring an intuitive model specification process. However, this model class comes with several challenges. First, as model complexity grows, so does the possibility for model non-identifiability---the existence of multiple models (or parameterizations of the same models) that explain the observed data equally well. This means that, while the model specification process is intuitive, the results of inference may not be. Second, inference for such models is intractable, requiring approximations. Unfortunately, these approximations require us to make additional assumptions. But unlike the assumptions made in the original model specification, these inference assumptions are highly unintuitive. As such, in safety-critical domains, they present a significant barrier for use. Moreover, these inference assumptions may impact downstream decision-making in unpredictable (and undesirable) ways. This dissertation is focused on (a) understanding how non-identifiability compromises the quality of approximate inference for deep probabilistic and Bayesian latent variable models, and (b) developing novel approximate inference methods that mitigate or accommodate non-identifiability to achieve better performance on downstream tasks.

Description

Other Available Sources

Keywords

Bayesian Statistics, Deep Learning, Latent Variable Models, Non-Identifiability, Variational Inference, Artificial intelligence, Statistics, Computer science

Terms of Use

This article is made available under the terms and conditions applicable to Other Posted Material (LAA), as set forth at Terms of Service

Endorsement

Review

Supplemented By

Referenced By

Related Stories