moreau-yosida-f-divergences
None
view repo
Variational representations of f-divergences are central to many machine learning algorithms, with Lipschitz constrained variants recently gaining attention. Inspired by this, we generalize the so-called tight variational representation of f-divergences in the case of probability measures on compact metric spaces to be taken over the space of Lipschitz functions vanishing at an arbitrary base point, characterize functions achieving the supremum in the variational representation, propose a practical algorithm to calculate the tight convex conjugate of f-divergences compatible with automatic differentiation frameworks, define the Moreau-Yosida approximation of f-divergences with respect to the Wasserstein-1 metric, and derive the corresponding variational formulas, providing a generalization of a number of recent results, novel special cases of interest and a relaxation of the hard Lipschitz constraint. As an application of our theoretical results, we propose the Moreau-Yosida f-GAN, providing an implementation of the variational formulas for the Kullback-Leibler, reverse Kullback-Leibler, χ^2, reverse χ^2, squared Hellinger, Jensen-Shannon, Jeffreys, triangular discrimination and total variation divergences as GANs trained on CIFAR-10, leading to competitive results and a simple solution to the problem of uniqueness of the optimal critic.
READ FULL TEXT
This paper aims to bring a new perspective for understanding GANs, by de...
read it
In this paper, we propose the Lipschitz margin ratio and a new metric
le...
read it
Wasserstein GANs (WGANs), built upon the Kantorovich-Rubinstein (KR) dua...
read it
The current best algorithms for convex body chasing problem in online
al...
read it
For elliptic diffusion equations with random coefficient and source term...
read it
We generalize the Jensen-Shannon divergence by considering a variational...
read it
A simple method is shown to provide optimal variational bounds on
f-dive...
read it
None
Comments
There are no comments yet.