Orthogonality and graph divergence losses promote disentanglement in generative models

Over the last decade, deep generative models have evolved to generate realistic and sharp images. The success of these models is often attributed to an extremely large number of trainable parameters and an abundance of training data, with limited or no understanding of the underlying data manifold....

Full description

Saved in:
Bibliographic Details
Published in:Frontiers in computer science (Lausanne) Vol. 6
Main Authors: Shukla, Ankita, Dadhich, Rishi, Singh, Rajhans, Rayas, Anirudh, Saidi, Pouria, Dasarathy, Gautam, Berisha, Visar, Turaga, Pavan
Format: Journal Article
Language:English
Published: Frontiers Media S.A 22-05-2024
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Over the last decade, deep generative models have evolved to generate realistic and sharp images. The success of these models is often attributed to an extremely large number of trainable parameters and an abundance of training data, with limited or no understanding of the underlying data manifold. In this article, we explore the possibility of learning a deep generative model that is structured to better capture the underlying manifold's geometry, to effectively improve image generation while providing implicit controlled generation by design. Our approach structures the latent space into multiple disjoint representations capturing different attribute manifolds. The global representations are guided by a disentangling loss for effective attribute representation learning and a differential manifold divergence loss to learn an effective implicit generative model. Experimental results on a 3D shapes dataset demonstrate the model's ability to disentangle attributes without direct supervision and its controllable generative capabilities. These findings underscore the potential of structuring deep generative models to enhance image generation and attribute control without direct supervision with ground truth attributes signaling progress toward more sophisticated deep generative models.
ISSN:2624-9898
2624-9898
DOI:10.3389/fcomp.2024.1274779