Skip to content

amoldamare/Optimizing-GAN-using-Nestorov-Dual-averaging

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This project was done as a course project for CSE 592 Convex Optimization , Spring 2018 at Stony Brook University computer science. This repository contains the code for a saddle point optimizer proposed by Nestorov. We have applied this optimizer to train Wasserstein GAN. For details please refer to main.pdf in "report" folder

Applying Dual-Averaging Sub-gradient method to

saddle point problem of Generative adversarial

networks

Amol Damare
Department of Computer Science
Stony Brook University
SBU Id:
adamare@cs.stonybrook.edu
Arjun Krishna
Department of Computer Science
Stony Brook University
SBU Id:
arjkrishna@cs.stonybrook.edu

Abstract

Training of Generative Adversarial Networks is hard. It is one of the open research
problem in the field of deep learning. The objective of GAN is modeled as saddle
point problem. In this paper we have applied already existing algorithms for saddle
point problem from optimization theory to GAN. Specifically we have implemented
SDA method proposed by Nestorov

##Conclusion

We were able to successfully apply an existing optimization algorithm for Saddle point problem
to Wasserstein GAN with relatively simple model( 1 or 2 layer deep). But we were not able to
replicate same for cross-entropy GAN objective or a more complex GAN model like DC-GAN. When
compared with existing optimization techniques used to solve GAN we found that they perform very
similarly.

Reference

Yurii Nesterov. Primal-dual subgradient methods for convex problems. Mathematical Program-
ming, 120(1):221?259, August 2009.

About

Applying Dual averaging method to Saddle point problem in GAN

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published