DISSERTATION

A mortized inference for structured deep generative models

Abstract

The development of Artificial Intelligence aims at building systems that can learn to accomplish tasks that human beings or animals can perform. Recent advances in deep learning have demonstrated that neural networks can achieve remarkable progress when employing large amounts of data and computation. In the past ten years, we have seen the rise of various deep generative models that are particularly successful in generation tasks in the domains of computer vision and natural language processing. On the other hand, many inference tasks would involve representing specific structures from data. In the absence of any supervision, we have increasingly seen limitations of standard deep generative models when tackling such inference tasks. In addition to interpreting the structure from data, we also hope to build models that can generalize to sparse data, prevent over-confident predictions, and perform inference in a computation-efficient manner. This thesis explores one possible direction for solving problems where we need more task-specific models by combining probabilistic modeling with deep learning techniques. We achieve that by inducing inductive biases in the form of structured priors in deep generative models. Our hope is that the models incorporate useful domain knowledge that helps to improve generalization and characterize uncertainty. We start by discussing how we can disentangle different factors of variation using Variational Autoencoders (VAEs), where we develop a novel decomposition of the objective to achieve this. Then we will focus on the more structured problems where we will talk about incorporating domain knowledge by inductively biasing the deep generative models with structured priors. We propose Amortized Population Gibbs Samplers, a class of iterative amortized inference methods that can scale to structured models while standard inference methods fail. Then we will generalize this approach and propose Nested Variational Inference, a class of inference methods that combine nested importance sampling with variational inference. Also, we build Inference Combinators where we propose constructors that allow us to more easily implement advanced amortized importance samplers. We also explore a relatively orthogonal topic where we will discuss what good representations mean. We develop Conjugate Energy-Based Models, an alternative to VAEs for unsupervised representation learning. This class of models can learn representations that can align better with class labels in image data compared to baselines.--Author's abstract

Keywords:
Inference Artificial intelligence Computer science Deep learning Generative grammar Machine learning Generative model Generalization Task (project management) Domain (mathematical analysis) Probabilistic logic Artificial neural network Engineering

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
359
Refs
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Generative Adversarial Networks and Image Synthesis
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Gaussian Processes and Bayesian Inference
Physical Sciences →  Computer Science →  Artificial Intelligence
Computational Physics and Python Applications
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Human-controllable and structured deep generative models

Tran, Dieu Linh

Journal:   Spiral (Imperial College London) Year: 2021
DISSERTATION

Deep generative models for time series counterfactual inference

Li, Guangyu (author)

University:   University of Southern California Digital Library Year: 2021
JOURNAL ARTICLE

Deep generative models for reject inference in credit scoring

Rogelio A. MancisidorMichael KampffmeyerKjersti AasRobert Jenssen

Journal:   Munin Open Research Archive (The Arctic University of Norway) Year: 2020
JOURNAL ARTICLE

Causal Inference with Conditional Instruments Using Deep Generative Models

Debo ChengZiqi XuJiuyong LiLin LiuJixue LiuThuc Duy Le

Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Year: 2023 Vol: 37 (6)Pages: 7122-7130
© 2026 ScienceGate Book Chapters — All rights reserved.