Rethinking conditional GAN training: An approach using geometrically structured latent manifolds

Sameera Ramasinghe, Moshiur Farazi, Salman Khan, Nick Barnes, Stephen Gould

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    Abstract

    Conditional GANs (cGAN), in their rudimentary form, suffer from critical drawbacks such as the lack of diversity in generated outputs and distortion between the latent and output manifolds. Although efforts have been made to improve results, they can suffer from unpleasant side-effects such as the topology mismatch between latent and output spaces. In contrast, we tackle this problem from a geometrical perspective and propose a novel training mechanism that increases both the diversity and the visual quality of a vanilla cGAN, by systematically encouraging a bi-lipschitz mapping between the latent and the output manifolds. We validate the efficacy of our solution on a baseline cGAN (i.e., Pix2Pix) which lacks diversity, and show that by only modifying its training mechanism (i.e., with our proposed Pix2Pix-Geo), one can achieve more diverse and realistic outputs on a broad set of image-to-image translation tasks.
    Original languageEnglish
    Title of host publicationProceedings of the 35th Conference on Neural Information Processing Systems (NeurIPS 2021)
    EditorsM. Ranzato and A. Beygelzimer and Y. Dauphin and P.S. Liang and J. Wortman Vaugh
    Place of PublicationUSA
    PublisherIEEE
    Pages1-13
    ISBN (Print)9781713845393
    Publication statusPublished - 2021
    Event35th Conference on Neural Information Processing Systems (NeurIPS 2021) - online
    Duration: 1 Jan 2021 → …

    Conference

    Conference35th Conference on Neural Information Processing Systems (NeurIPS 2021)
    Period1/01/21 → …
    Other7-10 December

    Fingerprint

    Dive into the research topics of 'Rethinking conditional GAN training: An approach using geometrically structured latent manifolds'. Together they form a unique fingerprint.

    Cite this