Dynamical variational autoencoders: A comprehensive review

L Girin, S Leglaive, X Bie, J Diard, T Hueber… - arXiv preprint arXiv …, 2020 - arxiv.org
Variational autoencoders (VAEs) are powerful deep generative models widely used to
represent high-dimensional complex data through a low-dimensional latent space learned …

Learning transferable visual models from natural language supervision

A Radford, JW Kim, C Hallacy… - International …, 2021 - proceedings.mlr.press
State-of-the-art computer vision systems are trained to predict a fixed set of predetermined
object categories. This restricted form of supervision limits their generality and usability since …

Toward transparent ai: A survey on interpreting the inner structures of deep neural networks

T Räuker, A Ho, S Casper… - 2023 ieee conference …, 2023 - ieeexplore.ieee.org
The last decade of machine learning has seen drastic increases in scale and capabilities.
Deep neural networks (DNNs) are increasingly being deployed in the real world. However …

On disentangled representations learned from correlated data

F Träuble, E Creager, N Kilbertus… - International …, 2021 - proceedings.mlr.press
The focus of disentanglement approaches has been on identifying independent factors of
variation in data. However, the causal variables underlying real-world observations are often …

Is a caption worth a thousand images? a controlled study for representation learning

S Santurkar, Y Dubois, R Taori, P Liang… - arXiv preprint arXiv …, 2022 - arxiv.org
The development of CLIP [Radford et al., 2021] has sparked a debate on whether language
supervision can result in vision models with more transferable representations than …

When is unsupervised disentanglement possible?

D Horan, E Richardson… - Advances in Neural …, 2021 - proceedings.neurips.cc
A common assumption in many domains is that high dimensional data are a smooth
nonlinear function of a small number of independent factors. When is it possible to recover …

Unsupervised learning of disentangled representation via auto-encoding: A survey

I Eddahmani, CH Pham, T Napoléon, I Badoc… - Sensors, 2023 - mdpi.com
In recent years, the rapid development of deep learning approaches has paved the way to
explore the underlying factors that explain the data. In particular, several methods have …

Soft-introvae: Analyzing and improving the introspective variational autoencoder

T Daniel, A Tamar - … of the IEEE/CVF Conference on …, 2021 - openaccess.thecvf.com
The recently introduced introspective variational autoencoder (IntroVAE) exhibits
outstanding image generations, and allows for amortized inference using an image encoder …

Speaker adaptation using spectro-temporal deep features for dysarthric and elderly speech recognition

M Geng, X Xie, Z Ye, T Wang, G Li, S Hu… - … on Audio, Speech …, 2022 - ieeexplore.ieee.org
Despite the rapid progress of automatic speech recognition (ASR) technologies targeting
normal speech in recent decades, accurate recognition of dysarthric and elderly speech …

Disentanglement of correlated factors via hausdorff factorized support

K Roth, M Ibrahim, Z Akata, P Vincent… - arXiv preprint arXiv …, 2022 - arxiv.org
A grand goal in deep learning research is to learn representations capable of generalizing
across distribution shifts. Disentanglement is one promising direction aimed at aligning a …