Please use this identifier to cite or link to this item:
http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/1987
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Munjal, P. | - |
dc.contributor.author | Paul, A. | - |
dc.contributor.author | Krishnan, N. C. | - |
dc.date.accessioned | 2021-07-03T12:31:52Z | - |
dc.date.available | 2021-07-03T12:31:52Z | - |
dc.date.issued | 2021-07-03 | - |
dc.identifier.uri | http://localhost:8080/xmlui/handle/123456789/1987 | - |
dc.description.abstract | Recently generative models have focused on combining the advantages of variational autoencoders (VAE) and generative adversarial networks (GAN) for good reconstruction and generative abilities. In this work we introduce a novel hybrid architecture, Implicit Discriminator in Variational Autoencoder (IDVAE), that combines a VAE and a GAN, which does not need an explicit discriminator network. The fundamental premise of the IDVAE architecture is that the encoder of a VAE and the discriminator of a GAN utilize common features and therefore can be trained as a shared network, while the decoder of the VAE and the generator of the GAN can be combined to learn a single network. This results in a simple two-tier architecture that has the properties of both a VAE and a GAN. The qualitative and quantitative experiments on real-world benchmark datasets demonstrates that IDVAE perform better than the state of the art hybrid approaches. We experimentally validate that IDVAE can be easily extended to work in a conditional setting and demonstrate its performance on complex datasets. | en_US |
dc.language.iso | en_US | en_US |
dc.title | Implicit discriminator in variational autoencoder | en_US |
dc.type | Article | en_US |
Appears in Collections: | Year-2020 |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Fulltext.pdf | 2.27 MB | Adobe PDF | View/Open Request a copy |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.