Show simple item record

dc.contributor.advisorSeelamantula, Chandra Sekhar
dc.contributor.advisorGhosh, Prasanta Kumar
dc.contributor.authorAsokan, Siddarth
dc.date.accessioned2023-09-20T05:34:12Z
dc.date.available2023-09-20T05:34:12Z
dc.date.submitted2023
dc.identifier.urihttps://etd.iisc.ac.in/handle/2005/6220
dc.description.abstractGenerative adversarial networks (GANs) are a popular learning framework to model the underlying distribution of images. GANs comprise a min-max game between the generator and the discriminator. While the generator transforms noise into realistic images, the discriminator learns to distinguish between the reals and the fakes. GANs are trained to either minimize a divergence function or an integral probability metrics (IPMs). In this thesis, we focus on understanding the optimality of GAN discriminator, generator, and its inputs, viewed from the perspective of Variational Calculus. Considering both divergence- and IPM-minimizing GANs, with and without gradient-based regularizers, we analyze the optimality of the GAN discriminator. We show that the optimal discriminator solves the Poisson partial differential equation, and derive solutions involving Fourier-series and radial basis function expansions. We show that providing the generator with data coming from a closely related input datasets accelerates and stabilizes training even in scenarios where there is no visual similarity between the source and target datasets. To identify closely related datasets, we propose the “signed Inception distance” (SID) as a novel GAN measure. Through the variational formulation, we demonstrate that the the optimal generator in GANs is linked to score-based Langevin diffusion and gradient flows. Leveraging these insights, we explore training GANs with flow-based and score-based costs, and diffusion models that perform discriminator-based updates.en_US
dc.language.isoen_USen_US
dc.relation.ispartofseries;ET00233
dc.rightsI grant Indian Institute of Science the right to archive and to make available my thesis or dissertation in whole or in part in all forms of media, now hereafter known. I retain all proprietary rights, such as patent rights. I also retain the right to use in future works (such as articles or books) all or part of this thesis or dissertationen_US
dc.subjectGenerative adversarial networksen_US
dc.subjectVariational Calculusen_US
dc.subjectHigh-dimensional Interpolationen_US
dc.subjectFourier analysisen_US
dc.subjectScore-based generative modelsen_US
dc.subjectContrastive learningen_US
dc.subjectKernel-based flowsen_US
dc.subjectMachine Learningen_US
dc.subjectDeep Learningen_US
dc.subjectOptimizationen_US
dc.subjectGenerative Modelingen_US
dc.subject.classificationResearch Subject Categories::TECHNOLOGY::Information technology::Signal processingen_US
dc.titleOn the Optimality of Generative Adversarial Networks — A Variational Perspectiveen_US
dc.typeThesisen_US
dc.degree.namePhDen_US
dc.degree.levelDoctoralen_US
dc.degree.grantorIndian Institute of Scienceen_US
dc.degree.disciplineEngineeringen_US


Files in this item

This item appears in the following Collection(s)

Show simple item record