Abstract
Generative adversarial network (GAN) has become one of the most important neural network models for classical unsupervised machine learning. A variety of discriminator loss functions have been developed to train GAN's discriminators and they all have a common structure: a sum of real and fake losses that only depends on the actual and generated data respectively. One challenge associated with an equally weighted sum of two losses is that the training may benefit one loss but harm the other, which we show causes instability and mode collapse. In this paper, we introduce a new family of discriminator loss functions that adopts a weighted sum of real and fake parts, which we call adaptive weighted loss functions or aw-loss functions. Using the gradients of the real and fake parts of the loss, we can adaptively choose weights to train a discriminator in the direction that benefits the GAN's stability. Our method can be potentially applied to any discriminator model with a loss that is a sum of the real and fake parts. Experiments validated the effectiveness of our loss functions on unconditional and conditional image generation tasks, improving the baseline results by a significant margin on CIFAR-10, STL-10, and CIFAR-100 datasets in Inception Scores (IS) and Fréchet Inception Distance (FID) metrics.
Original language | English |
---|---|
Title of host publication | Proceedings - 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021 |
Pages | 4779-4788 |
Number of pages | 10 |
ISBN (Electronic) | 9781665445092 |
DOIs | |
State | Published - 2021 |
Event | 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021 - Virtual, Online, United States Duration: Jun 19 2021 → Jun 25 2021 |
Publication series
Name | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition |
---|---|
ISSN (Print) | 1063-6919 |
Conference
Conference | 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021 |
---|---|
Country/Territory | United States |
City | Virtual, Online |
Period | 6/19/21 → 6/25/21 |
Bibliographical note
Publisher Copyright:© 2021 IEEE
Funding
Research supported in part by NSF OIA 2040665, NIH UH3 NS100606-05, and R01 HD101508-01 grants. Research supported in part by NSF under grants DMS-1821144 and DMS-1620082. We would like to thank Devin Willmott and Xin Xing for their initial efforts and helpful advice. We thank Rebecca Calvert for reading the manuscript and providing us with many valuable comments/suggestions. We also thank the University of Kentucky Center for Computational Sciences and Information Technology Services Research Computing for their support and use of the Lipscomb Compute Cluster. *Research supported in part by NSF OIA 2040665, NIH UH3 NS100606-05, and R01 HD101508-01 grants. †Research supported in part by NSF under grants DMS-1821144 and DMS-1620082.
Funders | Funder number |
---|---|
Devin Willmott and Xin Xing | |
NSF OIA | 2040665 |
University of Kentucky Medical Center | |
National Science Foundation (NSF) | DMS-1821144, DMS-1620082 |
National Institutes of Health (NIH) | R01 HD101508-01, UH3 NS100606-05 |
ASJC Scopus subject areas
- Software
- Computer Vision and Pattern Recognition