dc.contributor.advisor | Seelamantula, Chandra Sekhar | |
dc.contributor.advisor | Ghosh, Prasanta Kumar | |
dc.contributor.author | Asokan, Siddarth | |
dc.date.accessioned | 2023-09-20T05:34:12Z | |
dc.date.available | 2023-09-20T05:34:12Z | |
dc.date.submitted | 2023 | |
dc.identifier.uri | https://etd.iisc.ac.in/handle/2005/6220 | |
dc.description.abstract | Generative adversarial networks (GANs) are a popular learning framework to model the underlying distribution of images. GANs comprise a min-max game between the generator and the discriminator. While the generator transforms noise into realistic images, the discriminator learns to distinguish between the reals and the fakes. GANs are trained to either minimize a divergence function or an integral probability metrics (IPMs). In this thesis, we focus on understanding the optimality of GAN discriminator, generator, and its inputs, viewed from the perspective of Variational Calculus. Considering both divergence- and IPM-minimizing GANs, with and without gradient-based regularizers, we analyze the optimality of the GAN discriminator. We show that the optimal discriminator solves the Poisson partial differential equation, and derive solutions involving Fourier-series and radial basis function expansions. We show that providing the generator with data coming from a closely related input datasets accelerates and stabilizes training even in scenarios where there is no visual similarity between the source and target datasets. To identify closely related datasets, we propose the “signed Inception distance” (SID) as a novel GAN measure. Through the variational formulation, we demonstrate that the the optimal generator in GANs is linked to score-based Langevin diffusion and gradient flows. Leveraging these insights, we explore training GANs with flow-based and score-based costs, and diffusion models that perform discriminator-based updates. | en_US |
dc.language.iso | en_US | en_US |
dc.relation.ispartofseries | ;ET00233 | |
dc.rights | I grant Indian Institute of Science the right to archive and to make available my thesis or dissertation in whole or in part in all forms of media, now hereafter known. I retain all proprietary rights, such as patent rights. I also retain the right to use in future works (such as articles or books) all or part
of this thesis or dissertation | en_US |
dc.subject | Generative adversarial networks | en_US |
dc.subject | Variational Calculus | en_US |
dc.subject | High-dimensional Interpolation | en_US |
dc.subject | Fourier analysis | en_US |
dc.subject | Score-based generative models | en_US |
dc.subject | Contrastive learning | en_US |
dc.subject | Kernel-based flows | en_US |
dc.subject | Machine Learning | en_US |
dc.subject | Deep Learning | en_US |
dc.subject | Optimization | en_US |
dc.subject | Generative Modeling | en_US |
dc.subject.classification | Research Subject Categories::TECHNOLOGY::Information technology::Signal processing | en_US |
dc.title | On the Optimality of Generative Adversarial Networks — A Variational Perspective | en_US |
dc.type | Thesis | en_US |
dc.degree.name | PhD | en_US |
dc.degree.level | Doctoral | en_US |
dc.degree.grantor | Indian Institute of Science | en_US |
dc.degree.discipline | Engineering | en_US |