This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Mean Field Game GAN

Shaojun Ma
CSE, Department of Mathematics
Georgia Institute of Technology
Atlanta, GA 30332
[email protected]
&Haomin Zhou
Department of Mathematics
Georgia Institute of Technology
Atlanta, GA 30332
[email protected]
&Hongyuan Zha
School of Data Science
The Chinese University of Hong Kong
Shenzhen, Guangdong 518172
[email protected]
Abstract

We propose a novel mean field games (MFGs) based GAN(generative adversarial network) framework. To be specific, we utilize the Hopf formula in density space to rewrite MFGs as a primal-dual problem so that we are able to train the model via neural networks and samples. Our model is flexible due to the freedom of choosing various functionals within the Hopf formula. Moreover, our formulation mathematically avoids Lipschitz-1 constraint. The correctness and efficiency of our method are validated through several experiments.

1 Introduction

Over the past few years, we have witnessed a great success of generative adversarial networks (GANs) in various applications. Speaking of general settings of GANs, we have two neural networks, one is generator and the other one is discriminator. The goal is to make the generator able to produce samples from noise that matches the distribution of real data. During training process, the discriminator distinguishes the generated sample and real samples while the generator tries to ’fool’ the discriminator until an equilibrium state that the discriminator cannot tell any differences between generated samples and real samples.

Wasserstein GAN (WGAN)[3] made a breakthrough on computation performance since it adopts a more natural distance metric (Wasserstein distance) other than the Total Variation (TV) distance, the Kullback-Leibler (KL) divergence or the Jensen-Shannon (JS) divergence. However most algorithms related to WGAN is based on the Kantorovich duality form of OT(optimal transport) problem, in which we are facing Lipschitz-1 challenge in training process. Though several approximated methods have been proposed to realize Lipschitz condition in computations, we aim to find a more mathematical method.

MFGs[12, 16] are problems that model large populations of interacting agents. They have been widely used in economics, finance and data science. In mean field games, a continuum population of small rational agents play a non-cooperative differential game on a time horizon [0, TT]. At the optimum, the agents reach a Nash equilibrium (NE), where they can no longer unilaterally improve their objectives. In our case the NE leads to a PDE system consisting of continuity equation and Hamilton-Jacobi equation which represent the dynamical evolutions of the population density and the cost value respectively, following a similar pattern within the framework of Wasserstein distance[18]. Here the curse of infinite dimensionality in MFGs is overcome by applying Hopf formula in density space.

Utilizing Hopf formula to expand MFGs in density space is introduced in [9]. Base on their work, we show that under suitable choice of functionals, the MFGs framework can be reformulated to a Wasserstein-p primal-dual problem. We then handle high dimensional cases by leveraging deep neural networks. Notice that our model is a widely-generalized framework, different choices of Hamiltonian and functional settings lead to different cases. We demonstrate our approach in the context of a series of synthetic and real-world data sets. This paper is organized as following: 1) In Section 2, we review some typical GANs as well as concepts of MFGs. 2) In Section 3, we develop a new GAN formulation by applying a special Lagrangian, our model naturally and mathematically avoids Lipschitz-1 constraint. 3) In Section 4, we empirically demonstrate the effectiveness of our framework through several numerical experiments.

2 Related Works

2.1 Original GAN

Given a generator network GθG_{\theta} and a discriminator network DωD_{\omega}, the original GAN objective is to find an mapping that maps a random noise to an expected distribution, it is written as a minimax problem

infGθsupDω𝔼xρdata[logDω(x)]+𝔼zρnoise[log(1Dω(G(z))].\displaystyle\inf_{G_{\theta}}\sup_{D_{\omega}}\mathbb{E}_{x\sim\rho_{\text{data}}}\big{[}\log D_{\omega}(x)]+\mathbb{E}_{z\sim\rho_{\text{noise}}}\big{[}\log(1-D_{\omega}(G(z))]. (1)

The loss function is derived from the cross-entropy between the real and generated distributions. Since we never know the ground truth distribution, when compared with other generative modeling approaches, GAN has lots of computational advantages. In GAN framework, it is easy to do sampling without requiring Markov chain, we utilize expectations which is also easy to compute to construct differentiable loss function. Finally, the framework naturally combine the advantages of neural networks, avoiding curse of dimensionality and is extended to various GANs with different objective functions. The disadvantage of original GAN is that the training process is delicate and unstable due to the reasons theoretically investigated in [2].

2.2 Wasserstein GAN

Wasserstein distance provides a better and more natural measure to evaluate the distance between two distributions[29]. The objective function of Wasserstein GAN is constructed on the basis of Kantorovich-Rubinstein duality form [28] for distributions ρ1\rho_{1} and ρ2\rho_{2}

W(ρ1,ρ2)=supϕ,ϕ1𝔼xρ1[ϕ(x)]𝔼xρ2[ϕ(x)].\displaystyle W(\rho_{1},\rho_{2})=\sup_{\phi,\|\nabla\phi\|\leq 1}\mathbb{E}_{x\sim\rho_{1}}[\phi(x)]-\mathbb{E}_{x\sim\rho_{2}}[\phi(x)]. (2)

Given duality form, one would like to rewrite the duality form in a GAN framework by setting generator GθG_{\theta}, discriminator DωD_{\omega} and then minimizing the Wasserstein distance between target distribution and generated distribution

infGθsupDω,Dω1𝔼xρdata[Dω(x)]𝔼zρnoise[Dω(Gθ(z))].\displaystyle\inf_{G_{\theta}}\sup_{D_{\omega},\|\nabla D_{\omega}\leq 1\|}\mathbb{E}_{x\sim\rho_{data}}[D_{\omega}(x)]-\mathbb{E}_{z\sim\rho_{\text{noise}}}[D_{\omega}(G_{\theta}(z))]. (3)

Though WGAN has made a great success in various applications, it brings a well-known Lipschitz constraint problem, namely, we need to make our discriminator satisfies Dω1\|\nabla D_{\omega}\leq 1\|. Lipschitz constraint in WGAN has drawn a lot of attention of researchers, several methods such as weight clipping[3], adding gradient penalty as regularizer[13], applying weight normalization[27] and spectral normalization[24]. These methods enable convenience training process, however, besides these approximation-based method, we seek for a new formulation to avoid Lipschitz constraint in WGAN.

2.3 Other GANs

A variety of other GANs have been proposed in recent years. These GANs enlarge the application areas and provide great insights of designing generators and discriminators to facilitate the training process. One line of the work is focused on changing objective function such as f-GAN[25], Conditional-GAN[23], LS-GAN[22] and Cycle-GAN[30]. Another line is focused on developing stable structure of generator and discriminator, including DCGAN[26], Style-GAN[31] and Big-GAN[6]. Though these GANs has made a great success, there is no "best" GAN[20].

2.4 Hamilton-Jacobi Equation and Mean Field Games

Hamilton-Jacobi equation in density space (HJD) plays an important role in optimal control problems[4, 7, 14]. HJD determines the global information of the system[10, 11] and describes the time evolution of the optimal value in density space. In applications, HJD has been shown very effective at modeling population differential games, also known as MFGs which study strategic dynamical interactions in large populations by extending finite players’ differential games[16, 12]. Within this context, the agents reach a Nash equilibrium at the optimum, where they can no longer unilaterally improve their objectives. The solutions to these problems are obtained by solving the following partial differential equations (PDEs)[9]

{tρ(x,t)+x(ρ(x,t)pH(x,xΦ(x,t)))=0tΦ(x,t)+H(x,xΦ(x,t))+f(x,ρ(.,t))=0ρ(x,T)=ρ(x),Φ(x,0)=g(x,ρ(.,0)).\displaystyle\begin{cases}\partial_{t}\rho(x,t)+\nabla_{x}\cdot(\rho(x,t)\nabla_{p}H(x,\nabla_{x}\Phi(x,t)))=0\\ \partial_{t}\Phi(x,t)+H(x,\nabla_{x}\Phi(x,t))+f(x,\rho(.,t))=0\\ \rho(x,T)=\rho(x),\quad\Phi(x,0)=g(x,\rho(.,0)).\end{cases} (4)

where ρ(x,t)\rho(x,t) represents the population density of individual xx at time tt. Φ(x,t)\Phi(x,t) represents the velocity of population. We also have each player’s potential energy ff and terminal cost gg. The Hamiltonian HH is defined as

H(x,p)=supvv,pL(x,v).\displaystyle H(x,p)=\sup_{v}\langle v,p\rangle-L(x,v). (5)

where LL can be freely chosen.

Furthermore, a game is called a potential game when there exists a differentiable potential energy FF and terminal cost GG, such that the MFG can be modeled as a optimal control problem in the density space[9]

U(T,ρ)=infρ,v{0T[XL(x,v(x,t))ρ(x,t)dxF(ρ(.,t))]dt+G(ρ(.,0))}.\displaystyle U(T,\rho)=\inf_{\rho,v}\{\int_{0}^{T}[\int_{X}L(x,v(x,t))\rho(x,t)dx-F(\rho(.,t))]dt+G(\rho(.,0))\}. (6)

where the infimum is taken among all vector fields v(x,t)v(x,t) and densities ρ(x,t)\rho(x,t) subject to the continuity equation

{tρ(x,t)+(ρ(x,t)v(x,t))=00tTρ(x,T)=ρ(x).\displaystyle\begin{cases}\partial_{t}\rho(x,t)+\nabla\cdot(\rho(x,t)v(x,t))=0\quad 0\leq t\leq T\\ \rho(x,T)=\rho(x).\end{cases} (7)

Our method is derived from this formulation, the details will be introduced in the next section.

2.5 Optimal Transport

To better understand the connections between MFGs and WGAN, we also would like to have a review on optimal transport(OT). OT rises as a popular topic in recent years, OT-based theories and formulations have been widely used in fluid mechanics[1], control[8], GANs[3] as well as PDE[21]. We study OT problem because it provides an optimal map TT to pushforward one probability distribution μ\mu to another distribution ν\nu, with minimum cost cc. One formulation is Monge’s optimal transport problem[28]

infFXc(x,F(x))𝑑μ(x).\displaystyle\inf_{F}\int_{X}c(x,F(x))d\mu(x). (8)

However, the solution of Monge problem may not exist and even if the solution exists, it may not be unique. Another formulation called Kantorovich’s optimal transport problem[5] reads

{infπ𝒫(X×Y)X×Yc(x,y)𝑑π(x,y)Y𝑑π(x,y)=dμ(x),X𝑑π(x,y)=dν(y).\displaystyle\begin{cases}\underset{\pi\in\mathcal{P}(X\times Y)}{\inf}\int_{X\times Y}c(x,y)d\pi(x,y)\\ \int_{Y}d\pi(x,y)=d\mu(x),\quad\int_{X}d\pi(x,y)=d\nu(y).\end{cases} (9)

Notably, the fluid dynamic formulation of OT[5] reads

{infv,ρ0TX12v(x,t)2ρ(x,t)𝑑x𝑑tρt+(μν)=0ρ(x,0)=μ,ρ(x,T)=ν.\displaystyle\begin{cases}\underset{v,\rho}{\inf}\int_{0}^{T}\int_{X}\frac{1}{2}\|v(x,t)\|^{2}\rho(x,t)dxdt\\ \frac{\partial\rho}{\partial t}+\nabla\cdot(\mu\nu)=0\\ \rho(x,0)=\mu,\quad\rho(x,T)=\nu.\end{cases} (10)

where v(x,t)v(x,t) is the velocity vector. We solve this problem by applying Lagrange multiplier Φ(x,t)\Phi(x,t) and optimality conditions, finally we have its dual form

{supΦ,ρΦ(x,T)ρ(x,T)Φ(x,0)ρ(x,0)Φt+12Φ2=0.\displaystyle\begin{cases}\underset{\Phi,\rho}{\sup}\int\Phi(x,T)\rho(x,T)-\int\Phi(x,0)\rho(x,0)\\ \frac{\partial\Phi}{\partial t}+\frac{1}{2}\|\nabla\Phi\|^{2}=0.\end{cases} (11)

We also refer readers to another related dynamical formulation of OT [17].

3 Model Derivation

3.1 Formulation via Perspective of MFG

Let’s rewrite 6 and 7 with defining flux function m(x,t)=ρ(x,t)v(x,t)m(x,t)=\rho(x,t)v(x,t), then we have

U(T,ρ)=infρ,v{0T[XL(x,m(x,t)ρ(x,t))ρ(x,t)dxF(ρ(.,t))]dt+G(ρ(.,0))},\displaystyle U(T,\rho)=\inf_{\rho,v}\{\int_{0}^{T}[\int_{X}L(x,\frac{m(x,t)}{\rho(x,t)})\rho(x,t)dx-F(\rho(.,t))]dt+G(\rho(.,0))\}, (12)

with

{tρ(x,t)+m(x,t)=00tTρ(x,T)=ρ(x).\displaystyle\begin{cases}\partial_{t}\rho(x,t)+\nabla\cdot m(x,t)=0\quad 0\leq t\leq T\\ \rho(x,T)=\rho(x).\end{cases} (13)

We solve above primal-dual problem by applying Lagrange multiplier Φ(x,s)\Phi(x,s)

U(T,ρ)=infρ(x,t),m(x,t)supΦ(x,t)\displaystyle U(T,\rho)=\inf_{\rho(x,t),m(x,t)}\sup_{\Phi(x,t)} {0T[XL(x,m(x,t)ρ(x,t))ρ(x,t)dxF(ρ(.,t))]dt+G(ρ(.,0))\displaystyle\Biggl{\{}\int_{0}^{T}\Biggl{[}\int_{X}L\Biggl{(}x,\frac{m(x,t)}{\rho(x,t)}\Biggr{)}\rho(x,t)dx-F(\rho(.,t))\Biggr{]}dt+G(\rho(.,0))
+0TX(tρ(x,t)+m(x,t))Φ(x,t)dxdt}.\displaystyle+\int_{0}^{T}\int_{X}(\partial_{t}\rho(x,t)+\nabla\cdot m(x,t))\Phi(x,t)dxdt\Biggr{\}}. (14)

After integration by parts and reordering it leads to

U(T,ρ)=supΦ(x,t)infρ(x,t)\displaystyle U(T,\rho)=\sup_{\Phi(x,t)}\inf_{\rho(x,t)} {0TXρ(x,t)H(x,Φ(x,t))dxdt0TXρ(x,t)tΦ(x,t)dxdt\displaystyle\Biggl{\{}-\int_{0}^{T}\int_{X}\rho(x,t)H(x,\nabla\Phi(x,t))dxdt-\int_{0}^{T}\int_{X}\rho(x,t)\partial_{t}\Phi(x,t)dxdt
0TF(ρ(.,t))dt+G(ρ(.,0))Xρ(x,0)Φ(x,0)dx\displaystyle-\int_{0}^{T}F(\rho(.,t))dt+G(\rho(.,0))-\int_{X}\rho(x,0)\Phi(x,0)dx
+Xρ(x,t)Φ(x,t)dx},\displaystyle+\int_{X}\rho(x,t)\Phi(x,t)dx\Biggr{\}}, (15)

where

H(x,Φ(x,s))=supvΦvL(x,v).\displaystyle H(x,\nabla\Phi(x,s))=\sup_{v}\nabla\Phi\cdot v-L(x,v). (16)

Now let L(x,v)=vppL(x,v)=\frac{\|v\|^{p}}{p} where \|\cdot\| can be any norm in Euclidean space. For example, choose =2\|\cdot\|=\|\cdot\|_{2}, namely, v=v2=v12+v22++vn2\|v\|=\|v\|_{2}=\sqrt{v_{1}^{2}+v_{2}^{2}+...+v_{n}^{2}}. For arbitrary norm ss, we have the minimizer of 16 and corresponding HH as

Φ\displaystyle\nabla\Phi =vspsvs1\displaystyle=\|v\|_{s}^{p-s}\cdot v^{s-1} (17)
H(v)\displaystyle H(v) =vspvspp=1qvsp.\displaystyle=\|v\|_{s}^{p}-\frac{\|v\|_{s}^{p}}{p}=\frac{1}{q}\|v\|_{s}^{p}. (18)

Furthermore we have

Φsq\displaystyle\|\nabla\Phi\|_{s^{\prime}}^{q} =(vsps)qvs1sq\displaystyle=(\|v\|_{s}^{p-s})^{q}\|v^{s-1}\|_{s^{\prime}}^{q}
=[(|v|s𝑑x)pss]pp1(|vs1|ss1𝑑x)pp1s1s\displaystyle=[(\int|v|^{s}dx)^{\frac{p-s}{s}}]^{\frac{p}{p-1}}(\int|v^{s-1}|^{\frac{s}{s-1}}dx)^{\frac{p}{p-1}\frac{s-1}{s}}
=(|v|s𝑑x)1s(ps)pp1+1s(s1)pp1\displaystyle=(\int|v|^{s}dx)^{\frac{1}{s}(p-s)\frac{p}{p-1}+\frac{1}{s}(s-1)\frac{p}{p-1}}
=(|v|s𝑑x)ps\displaystyle=(\int|v|^{s}dx)^{\frac{p}{s}}
=vsp\displaystyle=\|v\|_{s}^{p} (19)
H(Φ)\displaystyle H(\nabla\Phi) =1qΦsqwhere1p+1q=1,1s+1s=1.\displaystyle=\frac{1}{q}\|\nabla\Phi\|_{s^{\prime}}^{q}\quad\text{where}\quad\frac{1}{p}+\frac{1}{q}=1,\quad\frac{1}{s}+\frac{1}{s^{\prime}}=1. (20)

Now we are able to plug H(Φ)H(\nabla\Phi) back into 15, let T=1T=1 and make FF and GG as 0, then

U(1,ρ)=supΦ(x,t)infρ(x,t)\displaystyle U(1,\rho)=\sup_{\Phi(x,t)}\inf_{\rho(x,t)} {01X(sΦ(x,t)+1qΦsq)ρ(x,t)dxdt\displaystyle\Biggl{\{}\int_{0}^{1}\int_{X}-(\partial_{s}\Phi(x,t)+\frac{1}{q}\|\nabla\Phi\|_{s^{\prime}}^{q})\rho(x,t)dxdt
Xρ(x,0)Φ(x,0)dx+Xρ(x,1)Φ(x,1)dx}.\displaystyle-\int_{X}\rho(x,0)\Phi(x,0)dx+\int_{X}\rho(x,1)\Phi(x,1)dx\Biggr{\}}. (21)

Next, we extend 21 to a GAN framework. We let ρ(z,t)\rho(z,t) be the generator that pushforward Gaussian noise distribution to target distribution. Specially, we set ρ1\rho_{1} as the distribution of our ground truth data. We treat Φ(x,t)\Phi(x,t) as discriminator thus it leads to our objective function

LMFG-GAN=infρ(x,t)supΦ(x,t){\displaystyle L_{\text{MFG-GAN}}=\inf_{\rho(x,t)}\sup_{\Phi(x,t)}\Biggl{\{} 𝔼zp(z),tUnif[0,1][tΦ(ρ(z,t),t)+1qΦ(ρ(z,t),t)sq]\displaystyle-\mathbb{E}_{z\sim p(z),t\sim Unif[0,1]}[\partial_{t}\Phi(\rho(z,t),t)+\frac{1}{q}\|\nabla\Phi(\rho(z,t),t)\|_{s^{\prime}}^{q}]
+𝔼xρ1[Φ(x,1)]𝔼zp(z)[Φ(ρ(z,0),0)]}.\displaystyle+\mathbb{E}_{x\sim\rho_{1}}[\Phi(x,1)]-\mathbb{E}_{z\sim p(z)}[\Phi(\rho(z,0),0)]\Biggr{\}}. (22)

We solve this minimax problem by algorithm 1. Notably, if we extend 22 to a general case with a general Hamiltonian HH, we revise 22 as

L=infρ(x,t)supΦ(x,t){\displaystyle L=\inf_{\rho(x,t)}\sup_{\Phi(x,t)}\Biggl{\{} 𝔼zp(z),tUnif[0,1][tΦ(G(z,t),t)+H(x,Φ,Φ)]\displaystyle-\mathbb{E}_{z\sim p(z),t\sim Unif[0,1]}[\partial_{t}\Phi(G(z,t),t)+H(x,\Phi,\nabla\Phi)]
+𝔼xρ1[Φ(x,1)]𝔼zp(z)[Φ(ρ(z,0),0)]}.\displaystyle+\mathbb{E}_{x\sim\rho_{1}}[\Phi(x,1)]-\mathbb{E}_{z\sim p(z)}[\Phi(\rho(z,0),0)]\Biggr{\}}. (23)
Algorithm 1
1:Initialize generator ρθ\rho_{\theta}, discriminator Φω\Phi_{\omega}
2:for ρθ\rho_{\theta} steps do
3:     Sample xx from real distribution
4:     Sample zz from standard normal distribution
5:     for Φω\Phi_{\omega} steps do
6:         Do gradient ascent on LMFG-GANL_{\text{MFG-GAN}} to update Φω\Phi_{\omega}
7:     end for
8:     Do gradient descent on LMFG-GANL_{\text{MFG-GAN}} to update ρθ\rho_{\theta}
9:end for

3.2 Formulation via Perspective of OT

Now we are going to derive the same objective function from the perspective of OT. Base on the formulation 11, we directly apply Lagrange multiplier ρ(x,s)\rho(x,s) on the constraint to reformulate it as a saddle problem

infρsupΦ{01X(Φt+12Φ2)ρ(x,t)dxdt+XΦ(x,1)ρ(x,1)𝑑xXΦ(x,0)ρ(x,0)𝑑x}.\displaystyle\inf_{\rho}\sup_{\Phi}\Biggl{\{}\int_{0}^{1}\int_{X}-\Biggl{(}\frac{\partial\Phi}{\partial t}+\frac{1}{2}\|\nabla\Phi\|^{2}\Biggr{)}\rho(x,t)dxdt+\int_{X}\Phi(x,1)\rho(x,1)dx-\int_{X}\Phi(x,0)\rho(x,0)dx\Biggr{\}}. (24)

We extend 24 to a GAN framework: we let ρ(z,s)\rho(z,s) be the generator and Φ(x,s)\Phi(x,s) as the discriminator thus finally we achieve the same problem

LOT-GAN=infρ(x,t)supΦ(x,t){\displaystyle L_{\text{OT-GAN}}=\inf_{\rho(x,t)}\sup_{\Phi(x,t)}\Biggl{\{} 𝔼zp(z),tUnif[0,1][tΦ(ρ(z,t),t)+12Φ(ρ(z,t),t)2]\displaystyle-\mathbb{E}_{z\sim p(z),t\sim Unif[0,1]}[\partial_{t}\Phi(\rho(z,t),t)+\frac{1}{2}\|\nabla\Phi(\rho(z,t),t)\|^{2}]
+𝔼xρ1[Φ(x,1)]𝔼zp(z)[Φ(ρ(z,0),0)]},\displaystyle+\mathbb{E}_{x\sim\rho_{1}}[\Phi(x,1)]-\mathbb{E}_{z\sim p(z)}[\Phi(\rho(z,0),0)]\Biggr{\}}, (25)

which has the same structure with 22, thus 22 can also be seen as minimizing Wasserstein-2 distance. A similar formulation is introduced in our recent work [19], where we consider two known distributions and compute their Wasserstein geodesic.

4 Experiments

Notice that for all experiments we set s=s=2s=s^{\prime}=2, we adopt fully connected neural networks for both generator and discriminator. In terms of training process, for all synthetic and realistic cases we use the Adam optimizer [15] with learning rate 10410^{-4}.

Refer to caption
(a) generated by (z,0)
Refer to caption
(b) generated by (z,1)
Refer to caption
(c) generated by (z,0)
Refer to caption
(d) generated by (z,1)
Figure 1: Results on 2D synthetic data set
Refer to caption
(a) Ground truth
Refer to caption
(b) generated by (z,0)
Refer to caption
(c) generated by (z,1)
Refer to caption
(d) Ground truth
Refer to caption
(e) generated by (z,0)
Refer to caption
(f) generated by (z,1)
Refer to caption
(g) Ground truth
Refer to caption
(h) generated by (z,0)
Refer to caption
(i) generated by (z,1)
Refer to caption
(j) Ground truth
Refer to caption
(k) generated by (z,0)
Refer to caption
(l) generated by (z,1)
Figure 2: Results on 10D synthetic data set

4.1 Synthetic

Syn-1: In this case we test our algorithm on 2D case, the target distribution is a Gaussian 𝒩(μ,Σ)\mathcal{N}(\mu,\Sigma), where μ\mu=(5,5) and Σ\Sigma= [[1,0],[0,1]], here we choose q=10,p=1.1q=10,p=1.1. The generated distributions are shown in Figure 1 (a) and (b).

Syn-2: Here and in all following tests we set q=2,p=2q=2,p=2, in this case we still generate a 2D Gaussian distribution, with the same mean and covariance matrix we used in Syn-1. The generated distributions are shown in Figure 1 (c) and (d).

Syn-3: In this case we generate a 10D Gaussian as our target distribution. Some of 2D projections of the true distributions and generated distributions are shown in Figure 2.

Refer to caption
(a) iteration 18000
Refer to caption
(b) iteration 52000
Refer to caption
(c) iteration 146400
Figure 3: Generated handwritten digits

4.2 Realistic

MNIST: We first test our algorithm on MNIST data set. The generated handwritten digits are shown in Figure 3. We believe more carefully designed neural networks such as CNN with extra time input dimension will improve the quality of generated pictures.

5 Conclusion and Future Directions

In this paper we derive a new GAN framework based on MFG formulation, by choosing special Hamiltonian and Lagrangian. The same formulation via OT perspective is also presented. The framework avoids Lipschitz-1 constraint and is validated through several synthetic and realistic data sets. We didn’t thoroughly investigate the optimal structure of neural networks, we only tried fully connected layers. Some of future works include improving structures of our neural networks, or providing a way to embed time tt into CNN.

References

  • [1] M. Agueh, B. Khouider, and L.P. Saumier. Optimal transport for particle image velocimetry. In Communications in Mathematical Sciences, 2016.
  • [2] M. Arjovsky and L. Bottou. Towards principled methods for training generative adversarial networks. arXiv: 1701.04862, 2017.
  • [3] Martin Arjovsky, Soumith Chintala, and Léon Bottou. Wasserstein gan. In arXiv preprint arXiv:1701.07875, 2017.
  • [4] J.D. Benamou and Y. Brenier. A computational fluid mechanics solution to the monge-kantorovich mass transfer problem. In Numerische Mathematik, 2000.
  • [5] J.D. Benamou, B.D. Froese, and A.M. Oberman. Two numerical methods for the elliptic monge-ampere equation. ESAIM: Mathematical Modelling and Numerical Analysis, 44(4):737–758, 2010.
  • [6] A. Brock, J. Donahue, and K. Simonyan. A style-based generator architecture for generative adversarial networks. In arXiv preprint arXiv:1809.11096, 2018.
  • [7] P. Cardaliaguet, F. Delarue, J.M. Lasry, and P.L. Lions. The master equation and the convergence problem in mean field games. In arXiv preprint arXiv:1509.02505, 2015.
  • [8] Y. Chen, T.T. Georgiou, and M. Pavon. On the relation between optimal transport and schrödinger bridges: A stochastic control viewpoint. In Journal of Optimization Theory and Applications, 2016.
  • [9] Y.T. Chow, W. Li, S. Osher, and W. Yin. Algorithm for hamilton–jacobi equations in density space via a generalized hopf formula. Journal of Scientific Computing, 80(2):1195–1239, 2019.
  • [10] W. Gangbo, T. Nguyen, and A. Tudorascu. Hamilton-jacobi equations in the wasserstein space. In Methods and Applications of Analysis, 2008.
  • [11] W. Gangbo and A. Swiech. Existence of a solution to an equation arising from the theory of mean field games. In Journal of Differential Equations, 2015.
  • [12] O. Gueant, J.M. Lasry, and P.L. Lions. Mean field games and applications. In In Paris-Princeton lectures on mathematical finance, 2010.
  • [13] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A.C. Courville. Improved training of wasserstein gans. In Advances in Neural Information Processing Systems, 2017.
  • [14] M. Huang, R.P. Malhamé, and P.E. Caines. Large population stochastic dynamic games: closed-loop mckean-vlasov systems and the nash certainty equivalence principle. In Communications in Information & Systems, 2006.
  • [15] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization, 2014.
  • [16] J.M. Lasry and P.L. Lions. Mean field games. In Japanese journal of mathematics, 2007.
  • [17] W. Li and S. Osher. Constrained dynamical optimal transport and its lagrangian formulation. In arXiv preprint arXiv:1807.00937, 2018.
  • [18] A.T. Lin, S.W. Fung, W. Li, L. Nurbekyan, and S.J. Osher. Apac-net: Alternating the population and agent control via two neural networks to solve high-dimensional stochastic mean field games. In arXiv preprint arXiv:2002.10113, 2020.
  • [19] S. Liu, S. Ma, Y. Chen, H. Zha, and H. Zhou. Learning high dimensional wasserstein geodesics. In arXiv preprint arXiv:2102.02992, 2021.
  • [20] M. Lucic, K. Kurach, M. Michalski, S. Gelly, and O. Bousquet. Are gans created equal? a large-scale study. In Advances in Neural Information Processing Systems, 2018.
  • [21] S. Ma, S. Liu, H. Zha, and H. Zhou. Learning stochastic behaviour of aggregate data. In arXiv preprint arXiv:2002.03513, 2020.
  • [22] X. Mao, Q. Li, H. Xie, Lau R.Y., Wang Z., and S. Paul Smolley. Least squares generative adversarial networks. In International Conference on Computer Vision, 2017.
  • [23] M. Mirza and S. Osindero. Conditional generative adversarial nets. In arXiv preprint arXiv:1411.1784, 2014.
  • [24] T. Miyato, T. Kataoka, M. Koyama, and Y. Yoshida. Spectral normalization for generative adversarial networks. In arXiv preprint arXiv:1802.05957, 2018.
  • [25] S. Nowozin, B. Cseke, and R. Tomioka. f-gan: Training generative neural samplers using variational divergence minimization. In Advances in Neural Information Processing Systems, pages 271–279, 2016.
  • [26] A. Radford, L. Metz, and S. Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. In arXiv preprint arXiv:1511.06434, 2015.
  • [27] T. Salimans and D.P. Kingma. Weight normalization: A simple reparameterization to accelerate training of deep neural networks. In Advances in Neural Information Processing Systems, 2016.
  • [28] C. Villani. Optimal transport: old and new, volume 338. Springer Science & Business Media, 2008.
  • [29] Cédric Villani. Topics in optimal transportation. Number 58. American Mathematical Soc., 2003.
  • [30] J.Y. Zhu, T. Park, P. Isola, and A.A. Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In International Conference on Computer Vision, 2017.
  • [31] T. ZKarras, S. Laine, and T. Aila. A style-based generator architecture for generative adversarial networks. In International Conference on Computer Vision and Pattern Recognition, 2019.