{"title":"Mixture Density Hyperspherical Generative Adversarial Networks","authors":"Qinyang Li, Wentao Fan","doi":"10.1145/3529466.3529475","DOIUrl":null,"url":null,"abstract":"The Generative Adversarial Networks (GANs) are deep generative models that can generate realistic samples, but they are difficult to train in practice due to the problem of mode collapse, where the generator only repeatedly generates one mode in samples during the learning process, or only generates a small number of modes after reaching the Nash equilibrium during the adversarial training. In order to solve this issue while making the generator contains promising generation ability, we propose a mixture density hyperspherical generative model namely MDH-GAN that combines variational autoencoder (VAE) and generative adversarial network. Unlike most of the GAN-based generative models that consider a Gaussian prior, MDH-GAN adopts the von Mises-Fisher (vMF) prior defined on a unit hypersphere. Our model combines VAE with GAN by integrating the encoder of VAE with GAN to form a jointly training framework. Therefore, the generator of our model can learn data distribution with a hyperspherical latent structure, leading to an improved generative ability of the generator. Moreover, a vMF mixture model is deployed in the discriminator to form a hypersphere space to avoid mode collapse of the model. In our experiments, by calculating the Fréchet Inception distance (FID) between the generated images and real ones, we prove that MDH-GAN has a better ability to generate high-quality images with high diversity.","PeriodicalId":375562,"journal":{"name":"Proceedings of the 2022 6th International Conference on Innovation in Artificial Intelligence","volume":"146 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 6th International Conference on Innovation in Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3529466.3529475","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The Generative Adversarial Networks (GANs) are deep generative models that can generate realistic samples, but they are difficult to train in practice due to the problem of mode collapse, where the generator only repeatedly generates one mode in samples during the learning process, or only generates a small number of modes after reaching the Nash equilibrium during the adversarial training. In order to solve this issue while making the generator contains promising generation ability, we propose a mixture density hyperspherical generative model namely MDH-GAN that combines variational autoencoder (VAE) and generative adversarial network. Unlike most of the GAN-based generative models that consider a Gaussian prior, MDH-GAN adopts the von Mises-Fisher (vMF) prior defined on a unit hypersphere. Our model combines VAE with GAN by integrating the encoder of VAE with GAN to form a jointly training framework. Therefore, the generator of our model can learn data distribution with a hyperspherical latent structure, leading to an improved generative ability of the generator. Moreover, a vMF mixture model is deployed in the discriminator to form a hypersphere space to avoid mode collapse of the model. In our experiments, by calculating the Fréchet Inception distance (FID) between the generated images and real ones, we prove that MDH-GAN has a better ability to generate high-quality images with high diversity.