This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On length measures of planar closed curves and the comparison of convex shapes

Nicolas Charon1 1 Department of Applied Mathematics and Statistics, Johns Hopkins University. [email protected]  and  Thomas Pierron2 2 Department of Mathematics, ENS Paris-Saclay. [email protected]
Abstract.

In this paper, we revisit the notion of length measures associated to planar closed curves. These are a special case of area measures of hypersurfaces which were introduced early on in the field of convex geometry. The length measure of a curve is a measure on the circle 𝕊1\mathbb{S}^{1} that intuitively represents the length of the portion of curve which tangent vector points in a certain direction. While a planar closed curve is not characterized by its length measure, the fundamental Minkowski-Fenchel-Jessen theorem states that length measures fully characterize convex curves modulo translations, making it a particularly useful tool in the study of geometric properties of convex objects. The present work, that was initially motivated by problems in shape analysis, introduces length measures for the general class of Lipschitz immersed and oriented planar closed curves, and derives some of the basic properties of the length measure map on this class of curves. We then focus specifically on the case of convex shapes and present several new results. First, we prove an isoperimetric characterization of the unique convex curve associated to some length measure given by the Minkowski-Fenchel-Jessen theorem, namely that it maximizes the signed area among all the curves sharing the same length measure. Second, we address the problem of constructing a distance with associated geodesic paths between convex planar curves. For that purpose, we introduce and study a new distance on the space of length measures that corresponds to a constrained variant of the Wasserstein metric of optimal transport, from which we can induce a distance between convex curves. We also propose a primal-dual algorithm to numerically compute those distances and geodesics, and show a few simple simulations to illustrate the approach. Keywords: planar curves, length measures, convex domains, isoperimetric inequality, metrics on shape spaces, optimal transport, primal-dual scheme.

1. Introduction

There is a long history of interactions between geometric analysis and measure theory that goes back to the early 20th century alongside the development of convex geometry [2, 23] and later on, in the 1960s, with the emergence of a whole new field known as geometric measure theory [22] from the works of Federer, Fleming and their students. Since then, ideas from geometric measure theory have found their way into applied mathematics most notably in areas such as computational geometry [16, 1, 10, 31] and shape analysis [29, 26, 14, 41] with applications in physics, image analysis and reconstruction or computer vision among many others.

A key element explaining the success of measure representations in geometric analysis and processing is their ability to encompass geometric structures of various regularities being either smooth immersed or embedded manifolds or discrete objects such as polytopes, polyhedra or simplicial complexes. They usually provide a comprehensive and efficient framework to capture the most essential geometric features in those objects. One of the earliest and most illuminating example is the use of the concept of area measures [2] in convex geometry which has proved instrumental to the theory of mixed volumes and the derivation of a multitude of geometric inequalities, in particular of isoperimetric or Brunn-Minkowski types [24, 42]. Area measures are typically defined for convex domains in n\mathbb{R}^{n} through the classical notion of support function of a convex set and can then be interpreted as a measure on the sphere 𝕊n1\mathbb{S}^{n-1} that represents the distribution of area of the domain’s boundary along its different normal directions. A fundamental property of the area measure, that resulted from a series of works by Minkowski, Alexandrov, Fenchel and Jessen, is that there is a one-to-one correspondence between convex sets (modulo translations) and area measures: in particular the area measure characterizes a convex set up to translation and therefore fully encodes its geometry.

In this paper, we are interested in area measures in the special and simplest case of planar curves. These are more commonly referred to as length measures and are finite measures on the unit circle 𝕊1\mathbb{S}^{1} that represents the distribution of length of the curve along its different tangent directions in the plane. Yet, unlike most previous works such as [34] which are mainly focused on convex objects, one of our purpose is to first introduce and investigate length measures of general rectifiable closed curves in the plane. Our objective is also to provide as elementary and self-contained of an introduction as possible to these notions from the point of view and framework of shape analysis without requiring preliminary concepts from convex geometry. Although, in sharp contrast to the convex case, there are infinitely many rectifiable curves that share the same length measure, we will emphasize some simple geometric properties of the underlying curve that can be recovered or connected to those of its length measure. Furthermore, this approach will allow us to formulate and prove a new characterization of the unique convex curve associated to a fixed measure on 𝕊1\mathbb{S}^{1} given by the Minkowski-Fenchel-Jessen theorem in the form of an isoperimetric inequality. Specifically, we show that this convex curve is the unique maximizer of the signed area among all the rectifiable curves with the same length measure. This extends some previously known results about polygons [8] from the field of discrete geometry.

One of the fundamental problem of shape analysis is the construction of relevant metrics on spaces of shapes such as curves, surfaces, images… which is often the first and crucial step to subsequently extend statistical methods on those highly nonlinear and infinite dimensional spaces. Considering the shape space of convex curves, the one-to-one representation provided by the length measure can be particularly advantageous in that regard. Indeed, various types of generalized measure representations of shapes such as currents [27, 19], varifolds [14, 32] or normal cycles [41] have been used in the recent past to define notions of distances between geometric shapes from corresponding metrics on those measure spaces. A common downside of all these frameworks however is that the mapping that associates a shape to its measure is only injective but not surjective. This typically prevents those metrics to be directly associated to geodesics on the shape space without adding more constraints and/or exterior regularization, for example through deformation models. In contrast, the bijectivity of the length measure map from the space of convex curves modulo translations to the space of all measures on 𝕊1\mathbb{S}^{1} satisfying only a simple linear closure constraint hints at the possibility of constructing a geodesic distance based on the length measure. In this paper, we propose an approach that relies on a constrained variant of the Wasserstein metric of optimal transport for probability measures on 𝕊1\mathbb{S}^{1} which, as we show, turns the set of convex curves of length 1 modulo translations into a geodesic space. We also adapt and implement a primal-dual algorithm to numerically estimate the distance and geodesics.

The paper is organized as follows. In Section 2, we define the length measure of a Lipschitz regular closed oriented curve of the plane and provide a few general geometric and approximation properties. In Section 3, we discuss specifically the case of convex curves and recap some of the fundamental connections between convex geometry and length measures, in particular the Minkowski-Fenchel-Jessen theorem. Section 4 is dedicated to the statement and proof of an isoperimetric inequality for curves of prescribed length measure, from which we can also find again the classical isoperimetric inequality for planar curves. Section 5 focuses on the construction of geodesic distances on the space of convex curves and in particular on our proposed constrained Wasserstein distance and its numerical computation. We also present and compare a few examples of estimated geodesics. Finally, Section 6 concludes the paper by discussing some current limitations of this work and potential avenues for future improvements and extensions.

2. Length measures of Lipschitz closed curves

2.1. Definitions

In all the paper, we will identify the 2D plane with the space of complex numbers \mathbb{C}. We start by introducing the space of closed, immersed oriented Lipschitz parametrized curves in the plane which we define as:

𝒞{cLip(𝕊1,)|c(θ)0fora.e.θ𝕊1}.\mathcal{C}\doteq\{c\in\text{Lip}(\mathbb{S}^{1},\mathbb{C})\ |\ c^{\prime}(\theta)\neq 0\ for\ a.e.\ \theta\in\mathbb{S}^{1}\}.

Depending on the context, we will identify 𝕊1\mathbb{S}^{1} either with the interval [0,2π)[0,2\pi)\subset\mathbb{R} or the circle {eiθ|θ}\{e^{i\theta}\ |\ \theta\in\mathbb{R}\}\subset\mathbb{C}. We recall that Lipschitz continuous curves are indeed differentiable almost everywhere and that the derivative is integrable on 𝕊1\mathbb{S}^{1}, which implies that the length L(c)=𝕊1|c(θ)|𝑑θL(c)=\int_{\mathbb{S}^{1}}|c^{\prime}(\theta)|d\theta is always finite. Note that we do not assume a priori that curves are simple. In anticipation to what follows, we also introduce the space of unparametrized oriented curves up to translations which we define as the quotient space

𝒞~𝒞/\tilde{\mathcal{C}}\doteq\mathcal{C}/\sim

the equivalence relation being that for c1,c2𝒞c_{1},c_{2}\in\mathcal{C}, c1c2c_{1}\sim c_{2} if and only if there exists z0z_{0}\in\mathbb{C} such that Im(c2)=Im(c1)+z0\text{Im}(c_{2})=\text{Im}(c_{1})+z_{0} and the orientations of Im(c2)\text{Im}(c_{2}) and Im(c1)+z0\text{Im}(c_{1})+z_{0} coincide. This will constitute the actual space of shapes, since objects in 𝒞~\tilde{\mathcal{C}} are essentially planar curves modulo reparametrizations and translations or in other words oriented rectifiable closed curves modulo translations in the plane. Furthermore, for any parametrization cc of a curve in 𝒞\mathcal{C} or 𝒞~\tilde{\mathcal{C}}, we let Tc:𝕊1𝕊1T_{c}:\mathbb{S}^{1}\rightarrow\mathbb{S}^{1} be the Gauss map, i.e. for all θ\theta, Tc(θ)=c(θ)/|c(θ)|T_{c}(\theta)=c^{\prime}(\theta)/|c^{\prime}(\theta)| which is well-defined for almost all θ𝕊1\theta\in\mathbb{S}^{1}. Lastly, we will write dsc=|c(θ)|dθds_{c}=|c^{\prime}(\theta)|d\theta the arc-length measure of cc, which is a positive measure on 𝕊1\mathbb{S}^{1}. In all this paper, we will denote by +(𝕊1)\mathcal{M}^{+}(\mathbb{S}^{1}) the set of all positive finite Radon measures on 𝕊1\mathbb{S}^{1} and (𝕊1)\mathcal{M}(\mathbb{S}^{1}) the set of signed finite Radon measures on 𝕊1\mathbb{S}^{1}. Also, we recall that the pushforward of a measure μ+(𝕊1)\mu\in\mathcal{M}^{+}(\mathbb{S}^{1}) by a mapping τ:𝕊1𝕊1\tau:\mathbb{S}^{1}\rightarrow\mathbb{S}^{1} is the measure denoted τμ\tau_{*}\mu such that (τμ)(B)=μ(τ1(B))(\tau_{*}\mu)(B)=\mu(\tau^{-1}(B)) for all Borel subset B𝕊1B\subset\mathbb{S}^{1}. Finally, we will denote by λn\lambda^{n} the Lebesgue measure on n\mathbb{R}^{n}. We can now introduce the notion of length measure which is the focus of this work.

Definition 2.1 (Length measure of a curve).

For c𝒞c\in\mathcal{C}, we define the length measure of cc that we write μc\mu_{c}, as the positive Radon measure on 𝕊1\mathbb{S}^{1} obtained as the pushforward of dscds_{c} by the Gauss map TcT_{c} i.e. μc(Tc)dsc\mu_{c}\doteq(T_{c})_{*}ds_{c}. In other words, for all Borel set BB of 𝕊1\mathbb{S}^{1}:

(1) μc(B)=dsc(Tc1(B))=Tc1(B)|c(θ)|𝑑θ.\mu_{c}(B)=ds_{c}(T_{c}^{-1}(B))=\int_{T_{c}^{-1}(B)}|c^{\prime}(\theta)|d\theta.

For any c1,c2c_{1},c_{2} such that [c1]=[c2][c_{1}]=[c_{2}] in 𝒞~\tilde{\mathcal{C}}, one has μc1=μc2\mu_{c_{1}}=\mu_{c_{2}} leading to the well-defined mapping M:cμcM:c\mapsto\mu_{c} from 𝒞~\tilde{\mathcal{C}} to the space of positive measures on 𝕊1\mathbb{S}^{1}.

The invariance of μc\mu_{c} to the equivalence relation \sim in Definition 2.1 is immediate from the fact that the arclength measure and Gauss map both only depend on the geometric image of the curve and are invariant to translations in the plane. In fact, the length measure of [c]𝒞~[c]\in\tilde{\mathcal{C}} can be also interpreted as the pushforward of the Hausdorff measure 1\mathcal{H}^{1} on the plane by the mapping T:Im(c)𝕊1T:\text{Im}(c)\rightarrow\mathbb{S}^{1} such that for any xIm(c)x\in\text{Im}(c), T(x)T(x) is the direction of the tangent vector of the curve at xx. Note however that μc\mu_{c} does depend on the orientation of cc: more specifically, if the orientation is reversed, the associated length measure of the resulting cˇ\check{c} is the reflection of μc\mu_{c} that is μcˇ(B)=μ(B)\mu_{\check{c}}(B)=\mu(-B) for all Borel set BB of 𝕊1\mathbb{S}^{1} (or equivalently if 𝕊1\mathbb{S}^{1} is identified with [0,2π)[0,2\pi), μcˇ(B)=μ((B+π)mod 2π)\mu_{\check{c}}(B)=\mu((B+\pi)\ \text{mod}\ 2\pi)). Thus μc\mu_{c} is a geometric quantity associated to unparametrized oriented curves modulo translations of the plane. For simplicity, we will still write μc\mu_{c} to denote the length measure associated to whole equivalence class [c]𝒞~[c]\in\tilde{\mathcal{C}}.

Remark 2.2.

The measure μc\mu_{c} can be intuitively understood as the distribution of length along the different directions of tangents to the curve cc, as we will further illustrate below. We point out that our definition of length measure departs slightly from the traditional concept of length measure (or perimetric measure as it is sometimes called) introduced initially for convex objects by Alexandrov, Fenchel and Jessen [2, 23], in that we consider here the direction of unit tangent vectors rather than unit normal vectors to the planar curve. Note that these two definitions only differ by a simple global rotation of the measure by an angle of π/2\pi/2 and so have a straightforward relation to one another. The reason for choosing this alternative convention is that this work focuses on length measures of planar curves and not area measures for general hypersurfaces, and our presentation will become a little simpler in this setting.

Lastly, thanks to the classical Riesz-Markov-Kakutani representation theorem, we also recall that signed measures on 𝕊1\mathbb{S}^{1} can be alternatively interpreted as elements of the dual to the space of continuous functions on 𝕊1\mathbb{S}^{1}. In the case of a length measure, the measure μc\mu_{c} acts on any fC(𝕊1)f\in C(\mathbb{S}^{1}) as follows:

(2) (μc|f)=𝕊1f(c(θ)|c(θ)|)|c(θ)|𝑑θ.(\mu_{c}|f)=\int_{\mathbb{S}^{1}}f\left(\frac{c^{\prime}(\theta)}{|c^{\prime}(\theta)|}\right)|c^{\prime}(\theta)|d\theta.

By straightforward extension, we can also consider the action of μc\mu_{c} on continuous complex-valued functions given by the same expression as in (2). Before looking into the properties of length measures, we add a final remark on the definition itself.

Remark 2.3.

Length measures can be also connected to some other concept of geometric measure theory namely the 1-varifolds as introduced in [3] and more precisely the oriented varifold representation of curves which is discussed and analyzed for instance in [32]. Indeed, the oriented varifold VcV_{c} associated to a curve c𝒞c\in\mathcal{C} is by definition the positive Radon measure on the product space 2×𝕊1\mathbb{R}^{2}\times\mathbb{S}^{1} given for all continuous compactly supported function ω\omega on 2×𝕊1\mathbb{R}^{2}\times\mathbb{S}^{1} by:

(Vc|ω)𝕊1ω(c(θ),c(θ)|c(θ)|)|c(θ)|𝑑θ.(V_{c}|\omega)\doteq\int_{\mathbb{S}^{1}}\omega\left(c(\theta),\frac{c^{\prime}(\theta)}{|c^{\prime}(\theta)|}\right)|c^{\prime}(\theta)|d\theta.

and VcV_{c} is once again independent of the choice of cc in the equivalence class [c]𝒞~[c]\in\tilde{\mathcal{C}}. By comparison to (2), the 1-varifold VcV_{c} can be thought as a spatially localized version of the length measure μc\mu_{c}, that is a distribution of unit directions at different positions in the plane. Equivalently, the length measure is obtained by marginalizing VcV_{c} with respect to its spatial component. This loss in spatial localization explains the lack of injectivity of the length measure representation (even after quotienting out translations) that is discussed below.

2.2. Basic geometric properties

Let us now examine more closely the most immediate properties of the length measure, in particular how it relates to various geometric quantities and transformations of the underlying curve. For a general smooth mapping ϕ:\phi:\mathbb{C}\rightarrow\mathbb{C}, we will write ϕc\phi\cdot c the curve θ𝕊1ϕ(c(θ))\theta\in\mathbb{S}^{1}\mapsto\phi(c(\theta)).

Proposition 2.4.

Let c𝒞c\in\mathcal{C} and μc\mu_{c} its length measure. Then

  1. (1)

    For all θ1,θ2𝕊1=[0,2π)\theta_{1},\theta_{2}\in\mathbb{S}^{1}=[0,2\pi):

    μc([θ1,θ2])=Length({c(θ)|θ1angle(c(θ))θ2}).\mu_{c}([\theta_{1},\theta_{2}])=\text{Length}\left(\left\{c(\theta)\ |\ \theta_{1}\leq\text{angle}(c^{\prime}(\theta))\leq\theta_{2}\right\}\right).

    In particular, the total length of cc is L(c)=μc(𝕊1)L(c)=\mu_{c}(\mathbb{S}^{1}).

  2. (2)

    For any rotation RθSO(2)R_{\theta}\in SO(2) of angle θ𝕊1\theta\in\mathbb{S}^{1}, μRθc=(Rθ)μc\mu_{R_{\theta}\cdot c}=(R_{\theta})_{*}\mu_{c} namely for all B𝕊1B\subset\mathbb{S}^{1}, μRθc(B)=μc((Bθ)mod 2π)\mu_{R_{\theta}\cdot c}(B)=\mu_{c}((B-\theta)\ \text{mod}\ 2\pi).

  3. (3)

    For any λ>0\lambda>0, μλc=λμc\mu_{\lambda c}=\lambda\mu_{c}, where λc\lambda c denotes the rescaling of cc by a factor λ\lambda.

Note that property (1) above implies in particular that the length measures of curves of length one are probability measures on 𝕊1\mathbb{S}^{1}. Combined with property (3) on the action of scalings, one could further define a mapping from 𝒞~/Scal\tilde{\mathcal{C}}/\text{Scal}, the space of curves modulo translation and scaling, into the space of probability measures on 𝕊1\mathbb{S}^{1} by essentially renormalizing curves to have length one. In all cases, by identifying 𝕊1\mathbb{S}^{1} with [0,2π)[0,2\pi), a convenient way to represent and visualize μc\mu_{c} is through its cumulative distribution function (cdf):

Fμc:[0,2π)\displaystyle F_{\mu_{c}}:[0,2\pi) [0,Lc)\displaystyle\longrightarrow[0,L_{c})
θ\displaystyle\theta μc([0,θ])\displaystyle\longmapsto\mu_{c}([0,\theta])

which is always a non-decreasing and right-continuous function.

Refer to caption Refer to caption Refer to caption
Refer to caption
Figure 1. Three planar curves with the same length measure μc\mu_{c} which c.d.f is plotted on the second row.

There are some further constraints that length measures must satisfy. Most notably, since the curve cc is closed, we obtain from (2) that:

(μc|eiθ)=𝕊1eiθ𝑑μc(θ)=𝕊1eiangle(c(θ))|c(θ)|𝑑θ=𝕊1c(θ)𝑑θ=0.\displaystyle\left(\mu_{c}|e^{i\theta}\right)=\int_{\mathbb{S}^{1}}e^{i\theta}d\mu_{c}(\theta)=\int_{\mathbb{S}^{1}}e^{i\text{angle}(c^{\prime}(\theta))}|c^{\prime}(\theta)|d\theta=\int_{\mathbb{S}^{1}}c^{\prime}(\theta)d\theta=0.

In other words, all length measures are such that the expectation of eiθe^{i\theta} on 𝕊1\mathbb{S}^{1} vanishes. Together with the above, this leads us to introduce the following subset of +(𝕊1)\mathcal{M}^{+}(\mathbb{S}^{1}):

Definition 2.5.

We denote by 0+(𝕊1)\mathcal{M}_{0}^{+}(\mathbb{S}^{1}) (resp. 0(𝕊1)\mathcal{M}_{0}(\mathbb{S}^{1})) the space of all positive (resp. signed) Radon measures μ\mu on 𝕊1\mathbb{S}^{1} which are such that:

𝕊1eiθ𝑑μ(θ)=0𝕊1cos(θ)𝑑μ(θ)=𝕊1sin(θ)𝑑μ(θ)=0.\int_{\mathbb{S}^{1}}e^{i\theta}d\mu(\theta)=0\Leftrightarrow\ \int_{\mathbb{S}^{1}}\cos(\theta)d\mu(\theta)=\int_{\mathbb{S}^{1}}\sin(\theta)d\mu(\theta)=0.

We then define the mapping M:𝒞~0+M:\ \tilde{\mathcal{C}}\rightarrow\mathcal{M}^{+}_{0} by M([c])=μcM([c])=\mu_{c} which does not depend on the choice of cc in the equivalence class [c][c].

Again, we will often replace an element of the quotient [c]𝒞~[c]\in\tilde{\mathcal{C}} by one of its representant c𝒞c\in\mathcal{C} and then write M(c)M(c) instead of M([c])M([c]).

Remark 2.6.

Note that any positive measure μ\mu on 𝕊1\mathbb{S}^{1} such that μ(B)=μ(B)\mu(-B)=\mu(B) for all Borel subsets of 𝕊1\mathbb{S}^{1} belongs to 0+\mathcal{M}^{+}_{0}. Indeed the assumption implies that:

02πeiθ𝑑μ(θ)=0πeiθ𝑑μ(θ)+0+ππ+πeiθ𝑑μ(θ)=0πeiθ𝑑μ(θ)0πeiθ𝑑μ(θ)=0\int_{0}^{2\pi}e^{i\theta}d\mu(\theta)=\int_{0}^{\pi}e^{i\theta}d\mu(\theta)+\int_{0+\pi}^{\pi+\pi}e^{i\theta}d\mu(\theta)=\int_{0}^{\pi}e^{i\theta}d\mu(\theta)-\int_{0}^{\pi}e^{i\theta}d\mu(\theta)=0

where the third equality follows from the change of variable θθ+π\theta\leftarrow\theta+\pi and the fact that dμ(θ+π)=dμ(θ)d\mu(\theta+\pi)=d\mu(\theta) by assumption. As a consequence, given any positive measure μ\mu on 𝕊1\mathbb{S}^{1}, one can always symmetrize it by defining μ¯(B)=μ(B)+μ(B)\bar{\mu}(B)=\mu(B)+\mu(-B) and obtain a measure of 0+\mathcal{M}^{+}_{0}. As a side note, convex curves for which the length measure satisfy μc(B)=μc(B)\mu_{c}(-B)=\mu_{c}(B) are called central-symmetric and are an important class of objects in convex geometry.

It is obvious that the mapping MM in Definition 2.5 cannot be injective on 𝒞~\tilde{\mathcal{C}}. In fact, given c𝒞~c\in\tilde{\mathcal{C}}, there are infinitely many other curves that share the same length measure as cc. In Figure 1, we show several examples of curves having the same length measure which c.d.f is plotted underneath. On the other hand, MM is surjective as we shall see in the next section. For now, we will just illustrate the different types of length measures depending on he nature of the underlying curve with a few examples.

Refer to caption Refer to caption
Refer to caption Refer to caption
Figure 2. Examples of closed curves (left) with their corresponding c.d.f (right).
Example 2.7.

Assume that c(θ)=eiθc(\theta)=e^{i\theta} for θ[0,2π)\theta\in[0,2\pi) is the unit circle in \mathbb{C}. Then dsc(θ)=dθds_{c}(\theta)=d\theta and Tc(θ)=θT_{c}(\theta)=\theta thus μc=dθ\mu_{c}=d\theta is the uniform measure on 𝕊1\mathbb{S}^{1}.

Example 2.8.

Consider a closed polygon with nn vertices z1,z2,,znz_{1},z_{2},\ldots,z_{n} and the nn faces [z1,z2],[z2,z3],[z_{1},z_{2}],[z_{2},z_{3}], ,[zn,z0]\ldots,[z_{n},z_{0}]. For all j=1,,nj=1,\ldots,n, let αj=angle(zj+1zj)\alpha_{j}=\text{angle}(z_{j+1}-z_{j}), lj=|zj+1zj|l_{j}=|z_{j+1}-z_{j}| (by convention zn+1=z0z_{n+1}=z_{0}) and L=j=1nljL=\sum_{j=1}^{n}l_{j}. Then, one can construct a parametrization c𝒞c\in\mathcal{C} of the polygon as follows. We define θj=2π(j1)/n\theta_{j}=2\pi(j-1)/n for j=1,,n+1j=1,\ldots,n+1. Then, on [θj,θj+1][\theta_{j},\theta_{j+1}], j=1,,nj=1,\ldots,n, take c(θ)(θθj)n12πljeiαjc(\theta)\doteq(\theta-\theta_{j})\frac{n-1}{2\pi}l_{j}e^{i\alpha_{j}}. It follows that c(θ)=n12πljeiαjc^{\prime}(\theta)=\frac{n-1}{2\pi}l_{j}e^{i\alpha_{j}} for θ[θj,θj+1]\theta\in[\theta_{j},\theta_{j+1}] and one can easily check that it leads to μc=j=1nljδeiαj\mu_{c}=\sum_{j=1}^{n}l_{j}\delta_{e^{i\alpha_{j}}}. The length measure of a polygon is thus always a sum of Dirac masses, c.f Figure 2 for an illustration.

Example 2.9.

Lastly, we consider an example of a singular continuous length measure. To construct it, let’s introduce the standard Cantor distribution that we rescale to the interval [0,2π][0,2\pi] and denote it by σ\sigma. Its support is the Cantor set and its cumulative distribution function FσF_{\sigma} is the well-known devil’s staircase function on [0,2π][0,2\pi]. The measure σ\sigma is not in 0+\mathcal{M}^{+}_{0} but following Remark 2.6, we can consider instead its symmetrization σ¯0\bar{\sigma}\in\mathcal{M}_{0} which c.d.f is given by Fσ¯(θ)=Fσ(θ)+σ(([0,θ]+π)mod 2π)F_{\bar{\sigma}}(\theta)=F_{\sigma}(\theta)+\sigma(([0,\theta]+\pi)\ \text{mod}\ 2\pi). Now, letting Fσ¯(1)F_{\bar{\sigma}}^{(-1)} being the pseudo-inverse of Fσ¯F_{\bar{\sigma}}, such that for all θ[0,2π]\theta\in[0,2\pi] by Fσ¯(1)(θ)=inf{θ|Fσ¯(θ)θ}F_{\bar{\sigma}}^{(-1)}(\theta)=\inf\{\theta^{\prime}\ |\ F_{\bar{\sigma}}(\theta^{\prime})\geq\theta\}, we define the curve c:[0,2π)c:[0,2\pi)\rightarrow\mathbb{C}:

c(θ)=0θeiFσ¯(1)(θ)𝑑θc(\theta)=\int_{0}^{\theta}e^{iF_{\bar{\sigma}}^{(-1)}(\theta^{\prime})}d\theta^{\prime}

which is C1C^{1} and satisfies |c(θ)|=1|c^{\prime}(\theta)|=1 and Tc(θ)=eiFσ¯(1)(θ)T_{c}(\theta)=e^{iF_{\bar{\sigma}}^{(-1)}(\theta)}. As we will show more in details and in general in the proof of Theorem 3.2, it follows that we have Fμc=Fσ¯F_{\mu_{c}}=F_{\bar{\sigma}} and therefore μc=σ¯\mu_{c}=\bar{\sigma} is the above symmetrized Cantor distribution σ¯\bar{\sigma}. We simulated such a curve using an approximation at level 12 of the Cantor measure, which is shown in Figure 2.

Those examples show that length measures of curves in 𝒞\mathcal{C} may have density with respect to the Lebesgue measure on 𝕊1\mathbb{S}^{1}, be singular discrete measures but also singular continuous measures as the last example shows. Furthermore, we have the following connection between the length measure density of a sufficiently regular curve and its curvature:

Proposition 2.10.

If c𝒞c\in\mathcal{C} is in addition twice differentiable with bounded and a.e non-vanishing curvature then μc=ρ(θ)dθ\mu_{c}=\rho(\theta)d\theta where the density ρ(θ)\rho(\theta) is given for a.e θ𝕊1\theta\in\mathbb{S}^{1} by:

(3) ρ(θ)=uTc1({θ})1κc(u)\rho(\theta)=\sum_{u\in T_{c}^{-1}(\{\theta\})}\frac{1}{\kappa_{c}(u)}

where κc\kappa_{c} is the curvature of cc.

Proof.

First, we notice that Tc(θ)=|c(θ)|ddscTc(θ)=|c(θ)|κ(θ)T_{c}^{\prime}(\theta)=|c^{\prime}(\theta)|\frac{d}{ds_{c}}T_{c}(\theta)=|c^{\prime}(\theta)|\kappa(\theta). Now, let ff be a continuous test function 𝕊1\mathbb{S}^{1}\rightarrow\mathbb{R}. By definition, we have:

(μc|f)=𝕊1fTc(θ)|c(θ)|𝑑θ=𝕊1fTc(θ)1κ(θ)Tc(θ)𝑑θ.\displaystyle(\mu_{c}|f)=\int_{\mathbb{S}^{1}}f\circ T_{c}(\theta)|c^{\prime}(\theta)|d\theta=\int_{\mathbb{S}^{1}}f\circ T_{c}(\theta)\frac{1}{\kappa(\theta)}T_{c}^{\prime}(\theta)d\theta.

Since TcT_{c} is a Lipschitz function, by the coarea formula ([4] Theorem 2.93), the above integral can be rewritten as:

(μc|f)=𝕊1(uTc1({θ})f(θ)1κ(u)𝑑0(u))𝑑θ=𝕊1(uTc1({θ})1κ(u)𝑑0(u))f(θ)𝑑θ.(\mu_{c}|f)=\int_{\mathbb{S}^{1}}\left(\int_{u\in T_{c}^{-1}(\{\theta\})}f(\theta)\frac{1}{\kappa(u)}d\mathcal{H}^{0}(u)\right)d\theta=\int_{\mathbb{S}^{1}}\left(\int_{u\in T_{c}^{-1}(\{\theta\})}\frac{1}{\kappa(u)}d\mathcal{H}^{0}(u)\right)f(\theta)d\theta.

and furthermore for almost all θ𝕊1\theta\in\mathbb{S}^{1}, Tc1({θ})T_{c}^{-1}(\{\theta\}) is a finite set so we obtain:

(μc|f)=𝕊1(uTc1({θ})1κc(u))=ρ(θ)f(θ)𝑑θ.(\mu_{c}|f)=\int_{\mathbb{S}^{1}}\underbrace{\left(\sum_{u\in T_{c}^{-1}(\{\theta\})}\frac{1}{\kappa_{c}(u)}\right)}_{=\rho(\theta)}f(\theta)d\theta.

A consequence is that the c.d.f of the length measure of a curve satisfying the assumptions of Proposition 2.10 is given by Fμc(θ)=0θρ(θ)𝑑θF_{\mu_{c}}(\theta)=\int_{0}^{\theta}\rho(\theta^{\prime})d\theta^{\prime} and thus does not have any jumps. Such jumps can only occur with the presence of flat faces in the curve, in particular for polygons as in Example 2.8. Note also that if the curve is smooth and strictly convex, the curvature is non-vanishing everywhere and the mapping TcT_{c} is a bijection which implies in this case that ρ(θ)=1κc(Tc1(θ))\rho(\theta)=\frac{1}{\kappa_{c}(T_{c}^{-1}(\theta))}.

2.3. Convergence of length measures

It will be useful for the rest of the paper to examine the topological properties of the mapping M:cμcM:c\mapsto\mu_{c}. Specifically, we want to determine under what notion of convergence in the space of curves, we can recover convergence of the corresponding length measures. We remind that a sequence of measures μn(𝕊1)\mu_{n}\in\mathcal{M}(\mathbb{S}^{1}) is said to converge weakly to μ(𝕊1)\mu\in\mathcal{M}(\mathbb{S}^{1}), which will be written μnμ\mu_{n}\rightharpoonup\mu, when for all fC(𝕊1)f\in C(\mathbb{S}^{1}), (μn|f)(μ|f)(\mu_{n}|f)\rightarrow(\mu|f) as n+n\rightarrow+\infty.

First, it is clear that uniform convergence cncc_{n}\rightarrow c in 𝒞\mathcal{C} (or equivalently convergence in Hausdorff distance of the unparametrized curves in C~\tilde{C}) does not imply that μcn\mu_{c_{n}} converges to μc\mu_{c} even weakly. Indeed it suffices to consider a sequence of staircase curves as the one displayed in Figure 3 which converges in Hausdorff distance to the diamond curve in blue; however, for all nn, the length measure is identical and equal to μcn=2(δ0+δπ/2+δπ+δ3π/2)\mu_{c_{n}}=2(\delta_{0}+\delta_{\pi/2}+\delta_{\pi}+\delta_{3\pi/2}) whereas μc=2(δπ/4+δπ/4+δ3π/4+δ5π/4)\mu_{c}=\sqrt{2}(\delta_{\pi/4}+\delta_{-\pi/4}+\delta_{3\pi/4}+\delta_{5\pi/4}).

Refer to caption
Figure 3. Two curves close in Hausdorff distance but for which the length measures are respectively 2(δπ/4+δπ/4+δ3π/4+δ5π/4)\sqrt{2}(\delta_{\pi/4}+\delta_{-\pi/4}+\delta_{3\pi/4}+\delta_{5\pi/4}) (blue curve) and 2(δ0+δπ/2+δπ+δ3π/2)2(\delta_{0}+\delta_{\pi/2}+\delta_{\pi}+\delta_{3\pi/2}) (red curve).

However, assuming in addition some convergence of the derivatives, we have the following:

Proposition 2.11.

Let (cn)(c_{n}) be a sequence of 𝒞\mathcal{C} with uniformly bounded Lipschitz constant such that there exists c𝒞c\in\mathcal{C} for which cn(θ)c(θ)c_{n}^{\prime}(\theta)\rightarrow c^{\prime}(\theta) for almost every θ𝕊1\theta\in\mathbb{S}^{1}. Then (μcn)(\mu_{c_{n}}) converges weakly to μc\mu_{c} and for all Borel set B𝕊1B\subset\mathbb{S}^{1} with μc(B)=0\mu_{c}(\partial B)=0 one has μcn(B)μc(B)\mu_{c_{n}}(B)\rightarrow\mu_{c}(B).

Proof.

Let ff be a continuous real-valued function on 𝕊1\mathbb{S}^{1}.

(μcn|f)=𝕊1fTcn(θ)|cn(θ)|𝑑θ=𝕊1f(cn(θ)|cn(θ)|)|cn(θ)|𝑑θ.(\mu_{c_{n}}|f)=\int_{\mathbb{S}^{1}}f\circ T_{c_{n}}(\theta)|c_{n}^{\prime}(\theta)|d\theta=\int_{\mathbb{S}^{1}}f\left(\frac{c_{n}^{\prime}(\theta)}{|c_{n}^{\prime}(\theta)|}\right)|c_{n}^{\prime}(\theta)|d\theta.

By the assumptions above, outside a set of measure 0 in 𝕊1\mathbb{S}^{1}, we have cn(θ)c(θ)c_{n}^{\prime}(\theta)\rightarrow c^{\prime}(\theta) with c(θ)0c^{\prime}(\theta)\neq 0 and cn(θ)0c_{n}^{\prime}(\theta)\neq 0 for all nn. Thus for almost all θ\theta in 𝕊1\mathbb{S}^{1}, cn(θ)/|cn(θ)|c(θ)/|c(θ)|c_{n}^{\prime}(\theta)/|c_{n}^{\prime}(\theta)|\rightarrow c^{\prime}(\theta)/|c^{\prime}(\theta)| and thus

f(cn(θ)|cn(θ)|)|cn(θ)|n+f(c(θ)|c(θ)|)|c(θ)|f\left(\frac{c_{n}^{\prime}(\theta)}{|c_{n}^{\prime}(\theta)|}\right)|c_{n}^{\prime}(\theta)|\xrightarrow[n\rightarrow+\infty]{}f\left(\frac{c^{\prime}(\theta)}{|c^{\prime}(\theta)|}\right)|c^{\prime}(\theta)|

since ff is continuous. Furthermore, as ff is bounded, there exists α>0\alpha>0, such that f(cn(θ)/|cn(θ)|)αf(c_{n}^{\prime}(\theta)/|c_{n}^{\prime}(\theta)|)\leq\alpha for all nn and by assumption, there also exists β>0\beta>0 such that |cn(θ)|β|c_{n}^{\prime}(\theta)|\leq\beta. It results from Lebesgue dominated convergence theorem that:

(μcn|f)n+𝕊1f(c(θ)|c(θ)|)|c(θ)|𝑑θ=(μc|f).(\mu_{c_{n}}|f)\xrightarrow[n\rightarrow+\infty]{}\int_{\mathbb{S}^{1}}f\left(\frac{c^{\prime}(\theta)}{|c^{\prime}(\theta)|}\right)|c^{\prime}(\theta)|d\theta=(\mu_{c}|f).

and so μcnμc\mu_{c_{n}}\rightharpoonup\mu_{c}. The second statement is a classical consequence of weak convergence of Radon measures (see [20] Theorem 1.40). ∎

Another important question is whether polygonal approximation of a curve leads to consistent length measures which we answer in the particular case of piecewise smooth curves:

Proposition 2.12.

Let cc be a curve of 𝒞\mathcal{C} which is assumed in addition to be piecewise twice differentiable with bounded second derivative. If (cn)(c_{n}) is a sequence of polygonal curves with knk_{n} vertices given by c(θn,0),c(θn,1),,c(θn,kn)c(\theta_{n,0}),c(\theta_{n,1}),\ldots,c(\theta_{n,k_{n}}) with 0=θn,0<θn,1<<θn,kn=2π0=\theta_{n,0}<\theta_{n,1}<\ldots<\theta_{n,k_{n}}=2\pi and maxi{θi+1,nθi,n}0\max_{i}\{\theta_{i+1,n}-\theta_{i,n}\}\rightarrow 0 as n+n\rightarrow+\infty, then (μcn)(\mu_{c_{n}}) converges weakly to μc\mu_{c}.

Proof.

Using Proposition 2.11, we simply need to show that cnc_{n}^{\prime} converges pointwise to cc^{\prime} a.e and bound the Lipschitz constant of cnc_{n} uniformly. As cc is piecewise smooth, we can treat each of the intervals separately and fix θ𝕊1\theta\in\mathbb{S}^{1} with cc being twice differentiable at θ\theta. For n1n\geq 1, let us denote hn=maxi{θi+1,nθi,n}h_{n}=\max_{i}\{\theta_{i+1,n}-\theta_{i,n}\}. For nn large enough, let θj,nθ<θj+1,n\theta_{j,n}\leq\theta<\theta_{j+1,n} with cc twice differentiable on [θj,n,θj+1,n)[\theta_{j,n},\theta_{j+1,n}). Then we have by definition cn(θ)=c(θj+1,n)c(θj,n)θj+1,nθj,n=c(ξ)c_{n}^{\prime}(\theta)=\frac{c(\theta_{j+1,n})-c(\theta_{j,n})}{\theta_{j+1,n}-\theta_{j,n}}=c^{\prime}(\xi) for a certain ξ(θj,n,θj+1,n)\xi\in(\theta_{j,n},\theta_{j+1,n}) by Taylor’s theorem which, using the boundedness of the second derivative of cc, gives

|cn(θ)c(θ)|c′′L|ξθ|c′′Lhnn+0|c_{n}^{\prime}(\theta)-c^{\prime}(\theta)|\leq\|c^{\prime\prime}\|_{L^{\infty}}|\xi-\theta|\leq\|c^{\prime\prime}\|_{L^{\infty}}h_{n}\xrightarrow[n\rightarrow+\infty]{}0

Thus cnc_{n}^{\prime} converges to cc^{\prime} pointwise except maybe on the finite set of points where cc is not twice differentiable. Furthermore, the above equation also implies that Lip(cn)c′′Lhn+Lip(c)\text{Lip}(c_{n})\leq\|c^{\prime\prime}\|_{L^{\infty}}h_{n}+\text{Lip}(c) which is uniformly bounded in nn and therefore Proposition 2.11 leads to the conclusion. ∎

Note that the above approximation property requires more regularity on cc. We will see in the next section that the result holds for all convex curves as well.

3. Convex curves and length measures

As already pointed out, the length measure μc\mu_{c} does not characterize the curve itself since there are in fact infinitely many curves of 𝒞~\tilde{\mathcal{C}} in the fiber M1({μc})M^{-1}(\{\mu_{c}\}) for any given cc. Yet, remarkably, it is the case if one restricts to convex curves in 𝒞~\tilde{\mathcal{C}}. This follows from the fundamental Minkowski-Fenchel-Jessen theorem (also in part due to Alexandrov) established in [23, 2] that shows in general dimension the uniqueness of a convex domain associated to any given area measure. For the specific case of planar curves which is the focus of this paper, we provide a more direct and constructive proof of this result in the following section before highlighting a few other well-known connections with convex geometry.

3.1. Characterization of convex curves by their length measure

In all the following, we will denote by 𝒞~conv\tilde{\mathcal{C}}_{conv} the set of curves in 𝒞~\tilde{\mathcal{C}} which are simple, convex and positively oriented. For technical reasons that will appear later, will adopt the convention that degenerate convex curves made of two opposite segments (which length measures are of the form r(δθ0+δθ0+π)r(\delta_{\theta_{0}}+\delta_{\theta_{0}+\pi})) belong to 𝒞~conv\tilde{\mathcal{C}}_{conv}. For simplicity, we shall still write M:𝒞~conv0M:\tilde{\mathcal{C}}_{conv}\rightarrow\mathcal{M}_{0} for the restriction of the previous length measure mapping to 𝒞~conv\tilde{\mathcal{C}}_{conv}. Before stating the main connection between curves in 𝒞~conv\tilde{\mathcal{C}}_{conv} and length measures, let us start with the following lemma.

Lemma 3.1.

Let c𝒞c\in\mathcal{C} such that there exists 0θ1<θ2<2π0\leq\theta_{1}<\theta_{2}<2\pi with c(θ1)=c(θ2)c(\theta_{1})=c(\theta_{2}). Then there exist θ,θ~[θ1,θ2]\theta,\tilde{\theta}\in[\theta_{1},\theta_{2}] such that |angle(c(θ))angle(c(θ~))|π|\text{angle}(c^{\prime}(\theta))-\text{angle}(c^{\prime}(\tilde{\theta}))|\geq\pi with strict equality unless c([θ1,θ2])c([\theta_{1},\theta_{2}]) is a segment.

Proof.

By contradiction, let’s assume that given θ1,θ2\theta_{1},\theta_{2} as above, we have |angle(c(θ))angle(c(θ~))|<π|\text{angle}(c^{\prime}(\theta))-\text{angle}(c^{\prime}(\tilde{\theta}))|<\pi for all θ,θ~[θ1,θ2]\theta,\tilde{\theta}\in[\theta_{1},\theta_{2}] where c(θ)c^{\prime}(\theta) is defined. Up to a rotation and translation of the curve, we may assume that c(θ1)=c(θ2)=0c(\theta_{1})=c(\theta_{2})=0 and that angle(c(θ))[0,π)\text{angle}(c^{\prime}(\theta))\in[0,\pi) for θ[θ1,θ2]\theta\in[\theta_{1},\theta_{2}]. Assuming that angle(c(θ))=0\text{angle}(c^{\prime}(\theta))=0 a.e on [θ1,θ2][\theta_{1},\theta_{2}] would lead to c(θ2)=θ1θ2|c(θ)|𝑑θ>0c(\theta_{2})=\int_{\theta_{1}}^{\theta_{2}}|c^{\prime}(\theta)|d\theta>0 which is impossible. On the other hand, if angle(c(θ))0\text{angle}(c^{\prime}(\theta))\neq 0 on a subset of [θ1,θ2][\theta_{1},\theta_{2}] of non-zero measure, then we would have:

(4) Imag(c(θ2))=θ1θ2|c(θ)|sin(angle(c(θ)))𝑑θ\text{Imag}(c(\theta_{2}))=\int_{\theta_{1}}^{\theta_{2}}|c^{\prime}(\theta)|\sin(\text{angle}(c^{\prime}(\theta)))d\theta

strictly positive which is again a contradiction. It results that we can find θ,θ~[θ1,θ2]\theta,\tilde{\theta}\in[\theta_{1},\theta_{2}] such that |angle(c(θ))angle(c(θ~))|π|\text{angle}(c^{\prime}(\theta))-\text{angle}(c^{\prime}(\tilde{\theta}))|\geq\pi. Furthermore, if maxθ[θ1,θ2]|angle(c(θ))angle(c(θ~))|=π\max_{\theta\in[\theta_{1},\theta_{2}]}|\text{angle}(c^{\prime}(\theta))-\text{angle}(c^{\prime}(\tilde{\theta}))|=\pi, it can be easily seen from a similar reasoning as above and (4) that in this case, one must have angle(c(θ))=0\text{angle}(c^{\prime}(\theta))=0 or π\pi a.e on [θ1,θ2][\theta_{1},\theta_{2}] and therefore the curve is a subset of the horizontal line. ∎

The following result is a reformulation of the Minkowski-Fenchel-Jessen theorem in the special setting of this paper. The usual proof of this theorem in general dimensions is quite involved and requires many preliminary results and notions from convex geometry. We propose here an alternative proof specific to the case of curves which is more elementary and constructive.

Theorem 3.2.

The length measure mapping MM is a bijection from C~conv\tilde{C}_{conv} to 0+\mathcal{M}^{+}_{0}.

Proof.

1. We first show the surjectivity of MM. Thus, taking μ0+\mu\in\mathcal{M}^{+}_{0}, we want to construct c𝒞c\in\mathcal{C}, a parametrization of a curve in 𝒞~conv\tilde{\mathcal{C}}_{conv} such that M(c)=μc=μM(c)=\mu_{c}=\mu. To do so, let us first assume, without loss of generality thanks to Proposition 2.4 on the action of rescaling on length measures, that μ(𝕊1)=2π\mu(\mathbb{S}^{1})=2\pi. Then let Fμ:[0,2π][0,2π]F_{\mu}:[0,2\pi]\rightarrow[0,2\pi] be the c.d.f of μ\mu as we defined previously and let’s introduce its pseudo-inverse Fμ(1):[0,2π][0,2π]F_{\mu}^{(-1)}:[0,2\pi]\rightarrow[0,2\pi]:

Fμ(1)(s)=inf{θ[0,2π]|Fμ(θ)s}.F_{\mu}^{(-1)}(s)=\inf\{\theta\in[0,2\pi]\ |\ F_{\mu}(\theta)\geq s\}.

Note that, as FμF_{\mu}, the pseudo-inverse Fμ(1)F_{\mu}^{(-1)} is a non-decreasing function with Fμ(1)(0)=0F_{\mu}^{(-1)}(0)=0. We now define c:[0,2π]c:[0,2\pi]\rightarrow\mathbb{C} as follows:

(5) c(θ)=0θeiFμ(1)(s)𝑑s.c(\theta)=\int_{0}^{\theta}e^{iF_{\mu}^{(-1)}(s)}ds.

We first see that c𝒞c\in\mathcal{C}. Indeed, by construction, cc is differentiable almost everywhere with c(θ)=eiFμ(1)(θ)c^{\prime}(\theta)=e^{iF_{\mu}^{(-1)}(\theta)} giving |c(θ)|=1|c^{\prime}(\theta)|=1 thus cLip(𝕊1,)c\in\text{Lip}(\mathbb{S}^{1},\mathbb{C}) and cc is a Lipschitz immersion. Next, we obtain that μc=μ\mu_{c}=\mu by checking the equality Fμc=FμF_{\mu_{c}}=F_{\mu}. This is simply a consequence of the fact that Fμc(θ)=λ1({α| 0Fμ(1)(α)θ})F_{\mu_{c}}(\theta)=\lambda^{1}(\{\alpha\ |\ 0\leq F_{\mu}^{(-1)}(\alpha)\leq\theta\}) and the easy verification that Fμ(1)(α)θαFμ(θ)F_{\mu}^{(-1)}(\alpha)\leq\theta\Leftrightarrow\alpha\leq F_{\mu}(\theta) so that

Fμc(θ)=λ1({α| 0αFμ(θ)})=Fμ(θ).F_{\mu_{c}}(\theta)=\lambda^{1}(\{\alpha\ |\ 0\leq\alpha\leq F_{\mu}(\theta)\})=F_{\mu}(\theta).

Incidentally, this also shows that the curve is indeed closed as:

c(2π)c(0)=02πc(θ)𝑑θ=02πeiθ𝑑μc(θ)=02πeiθ𝑑μ(θ)=0.c(2\pi)-c(0)=\int_{0}^{2\pi}c^{\prime}(\theta)d\theta=\int_{0}^{2\pi}e^{i\theta}d\mu_{c}(\theta)=\int_{0}^{2\pi}e^{i\theta}d\mu(\theta)=0.

One still needs to verify that the image of cc belongs to 𝒞~conv\tilde{\mathcal{C}}_{conv}. As c(θ)=eiFμ(1)(θ)c^{\prime}(\theta)=e^{iF_{\mu}^{(-1)}(\theta)} and Fμ(1)F_{\mu}^{(-1)} is non-decreasing, cc is locally convex and we only need to show that cc is simple or is supported by a straight segment. By contradiction, let’s assume that there exists 0θ1<θ2<2π0\leq\theta_{1}<\theta_{2}<2\pi such that c(θ1)=c(θ2)c(\theta_{1})=c(\theta_{2}). We have to consider the two following cases:

  • If c([0,θ1])c([0,\theta_{1}]) is not a segment then, since angle(c(θ))\text{angle}(c^{\prime}(\theta)) is non-decreasing on [0,θ1][0,\theta_{1}], there is a limit α=limθθ1angle(c(θ))[0,2π)\alpha=\lim_{\theta\uparrow\theta_{1}}\text{angle}(c^{\prime}(\theta))\in[0,2\pi) and we have that c(0)c(0) is strictly above the line passing by c(θ1)c(\theta_{1}) and directed by eiαe^{i\alpha}. Furthermore, using Lemma 3.1, we deduce that for all θθ2\theta\geq\theta_{2}, we have α+πangle(c(θ))<2π\alpha+\pi\leq\text{angle}(c^{\prime}(\theta))<2\pi which implies that for all θθ2\theta\geq\theta_{2}, c(θ)c(\theta) is below the line passing by c(θ2)=c(θ1)c(\theta_{2})=c(\theta_{1}) directed by eiαe^{i\alpha} which contradicts the fact that limθ2πc(θ)=c(0)\lim_{\theta\rightarrow 2\pi^{-}}c(\theta)=c(0).

  • If c([0,θ1])c([0,\theta_{1}]) is a segment, then by Lemma 3.1 and the fact that θangle(c(θ))\theta\mapsto\text{angle}(c^{\prime}(\theta)) is non-decreasing, we have that angle(c(θ))angle(c(0))+π\text{angle}(c^{\prime}(\theta))\geq\text{angle}(c^{\prime}(0))+\pi for all θ2θ<2π\theta_{2}\leq\theta<2\pi with even strict inequality if c([θ1,θ2])c([\theta_{1},\theta_{2}]) is not a segment of the same direction as c([0,θ1])c([0,\theta_{1}]). The latter case is not possible since we would then find that for θ>θ2\theta>\theta_{2}, c(θ)c(\theta) is strictly below the line containing the segment c([0,θ1])c([0,\theta_{1}]) and thus limθ2πc(θ)c(0)\lim_{\theta\rightarrow 2\pi^{-}}c(\theta)\neq c(0). Finally, by a similar argument, we find that the image of cc on [θ2,2π)[\theta_{2},2\pi) is again a segment of the same direction, which shows that Im(c)\text{Im}(c) is eventually a segment.

Thus the curve is either simple, convex and positively oriented or is a segment, and in all cases belong to 𝒞~conv\tilde{\mathcal{C}}_{conv}. 2. Let us now prove the injectivity of MM. Specifically, we show that if cc is the arclength parametrization of a curve in C~conv\tilde{C}_{conv}, which we assume once again to have length 2π2\pi, and μc=μ0+\mu_{c}=\mu\in\mathcal{M}^{+}_{0} then for almost all s[0,2π)s\in[0,2\pi), c(s)=eiFμ(1)(s)c^{\prime}(s)=e^{iF_{\mu}^{(-1)}(s)}. By assumption, we have |c(s)|=1|c^{\prime}(s)|=1 and since the curve is convex and positively oriented, up to a shifting of the parameter, we may assume that sangle(c(s))s\mapsto\text{angle}(c^{\prime}(s)) is non-decreasing from [0,2π)[0,2\pi) to [0,2π)[0,2\pi). Let us denote ν(s)angle(c(s))[0,2π)\nu(s)\doteq\text{angle}(c^{\prime}(s))\in[0,2\pi). From Fμc=FμF_{\mu_{c}}=F_{\mu} we get that Fμ(θ)=λ1({s| 0ν(s)θ})F_{\mu}(\theta)=\lambda^{1}(\{s\ |\ 0\leq\nu(s)\leq\theta\}). We need to show that ν(s)=Fμ(1)(s)=inf{θ|Fμ(θ)s}\nu(s)=F_{\mu}^{(-1)}(s)=\inf\{\theta\ |\ F_{\mu}(\theta)\geq s\}. On the one hand, we have Fμ(ν(s))=λ1({s~| 0ν(s~)ν(s)})sF_{\mu}(\nu(s))=\lambda^{1}(\{\tilde{s}\ |\ 0\leq\nu(\tilde{s})\leq\nu(s)\})\geq s since for any s~[0,s]\tilde{s}\in[0,s] we have 0ν(s~)ν(s)0\leq\nu(\tilde{s})\leq\nu(s). This leads to Fμ(1)(s)ν(s)F_{\mu}^{(-1)}(s)\leq\nu(s). On the other hand, for any θ<ν(s)\theta<\nu(s) we have that Fμ(θ)<sF_{\mu}(\theta)<s by definition of ν(s)\nu(s) and consequently Fμ(1)(s)θF_{\mu}^{(-1)}(s)\geq\theta for all θ<ν(s)\theta<\nu(s) leading to Fμ(1)(s)ν(s)F_{\mu}^{(-1)}(s)\geq\nu(s). ∎

Remark 3.3.

Note that from the above proof, one can technically reconstruct a convex curve up to translation from its length measure directly based on (5). When the measure is discrete i.e. μ=i=1Nljδαj\mu=\sum_{i=1}^{N}l_{j}\delta_{\alpha_{j}} where lj>0l_{j}>0 and αj[0,2π)\alpha_{j}\in[0,2\pi) for all j=1,,Nj=1,\ldots,N, the reconstruction becomes particularly simple. In this case, one can see that the corresponding convex polygon is obtained by selecting an initial vertex (e.g. at the origin) and ordering the edges lj1eiαj1,,ljNeiαjNl_{j_{1}}e^{i\alpha_{j_{1}}},\ldots,l_{j_{N}}e^{i\alpha_{j_{N}}} such that the angles 0αj1αjN<2π0\leq\alpha_{j_{1}}\leq\ldots\leq\alpha_{j_{N}}<2\pi are in ascending order, which is a well-known algorithm for convex planar objects. However, this reconstruction is a significantly more difficult problem for area measures in higher dimensions, c.f. the discussion in Section 6.

3.2. Length measures and Minkowski sum of convex sets

The above correspondence between convex shapes and length measures has many interesting consequences and applications, in particular for the study of mixed areas and Brunn-Minkowski theory, as developed for example in [34, 30, 42]. We will not go over all of these in detail but only recap in this section a few results which shall be relevant for the rest of the paper.

First, in the category of planar convex curves, one has the following stronger version of the approximation result of Proposition 2.12:

Proposition 3.4.

Let CC be a planar convex domain with boundary c=CC~convc=\partial C\in\tilde{C}_{conv} and (Pn)(P_{n}) a sequence of convex polygons of boundary pn=Pnp_{n}=\partial P_{n} that converges in Hausdorff distance to CC. Then μpnμc\mu_{p_{n}}\rightharpoonup\mu_{c} as n+n\rightarrow+\infty. Furthermore, the area of PnP_{n} converges to the area of CC i.e. λ2(Pn)λ2(C)\lambda^{2}(P_{n})\rightarrow\lambda^{2}(C).

This is classical property of convex sets and length measures which proof can be found in [42] (Theorem 1.8.16 and Theorem 4.1.1).

We now recall the definition of the Minkowski sum. If C1C_{1} and C2C_{2} are two convex planar domains, their Minkowski sum (also known as dilation in mathematical morphology) is defined by C1+C2={x1+x2|x1C1,x2C2}C_{1}+C_{2}=\{x_{1}+x_{2}\ |\ x_{1}\in C_{1},x_{2}\in C_{2}\} which is also a convex planar domain. More generally, one can define the Minkowski combination a1C1+a2C2a_{1}C_{1}+a_{2}C_{2} for a1,a20a_{1},a_{2}\geq 0 as a1C1+a2C2={a1x1+a2x2|x1C1,x2C2}a_{1}C_{1}+a_{2}C_{2}=\{a_{1}x_{1}+a_{2}x_{2}\ |\ x_{1}\in C_{1},x_{2}\in C_{2}\}. This allows to view the set of all convex domains as a convex cone for this Minkowski addition. The length measure mapping MM has the following interesting property ([34] Theorem 3.2):

Proposition 3.5.

Let C1C_{1} and C2C_{2} be two convex domains and c1,c2c_{1},c_{2} their boundary curves. Then, denoting c𝒞c\in\mathcal{C} a parametrization of the boundary of C1+C2C_{1}+C_{2}, it holds that μc=μc1+μc2\mu_{c}=\mu_{c_{1}}+\mu_{c_{2}}.

Proof.

This is easily shown by approximating the convex domains C1C_{1} and C2C_{2} by sequences of convex polygons and using Proposition 3.4. The fact that the result holds for two convex polygons is well-known and is actually used algorithmically for the computation of the Minkowski sum of polygons in the plane in linear time, c.f. for example [44] (Chap. 13). ∎

By combining the above with Theorem 3.2 and Proposition 2.4 (3), we can summarize the properties of the length measure mapping MM as follows:

Corollary 3.6.

The map M:CμCM:C\mapsto\mu_{\partial C} is an isomorphism of convex cones between C~conv\tilde{C}_{conv} and 0\mathcal{M}_{0}.

Refer to caption Refer to caption Refer to caption Refer to caption
λ=0\lambda=0 λ=1/3\lambda=1/3 λ=2/3\lambda=2/3 λ=1\lambda=1
Figure 4. Minkowski sum (1λ)C1+λC2(1-\lambda)C_{1}+\lambda C_{2} for different values of λ\lambda of a triangle and a disk computed based on the addition of their length measures.

This implies that if C1C_{1} and C2C_{2} are two convex domains with length measures μC1\mu_{\partial C_{1}} and μC2\mu_{\partial C_{2}}, their Minkowski sum C=C1+C2C=C_{1}+C_{2} is such that C=M1(μC1+μC2)\partial C=M^{-1}(\mu_{\partial C_{1}}+\mu_{\partial C_{2}}) which could be directly computed using the inversion formula (5) or, in the case of discrete measures and polygons, by adequately sorting the Diracs appearing in μC1+μC2\mu_{\partial C_{1}}+\mu_{\partial C_{2}} with angles in ascending order as explained earlier in Remark 3.3. We show an example of Minkowski sum computed with this approach in Figure 4.

4. An isoperimetric characterization

The previous section showed that there is a unique convex curve of positive orientation in the preimage M1({μ})𝒞~M^{-1}(\{\mu\})\subset\tilde{\mathcal{C}} for any measure μ0+\mu\in\mathcal{M}^{+}_{0}. Several previous works on length and area measures have investigated variational characterizations of these convex objects in the context of shape optimization and geometric inequalities, see for instance the survey of [24] and [12]. These are typically expressing some variational property within the set of convex shapes only. We prove here a distinct characterization which can be rather interpreted as an isoperimetric inequality in each of the fiber M1({μ})M^{-1}(\{\mu\}), namely the convex curve of M1({μ})M^{-1}(\{\mu\}) is also the one of maximal signed area among all the curves in 𝒞~\tilde{\mathcal{C}} of length measure μ\mu. Our result extends to the whole class of Lipschitz regular curves some related results on polytopes in discrete geometry that were stated in [8]. We will also see how the classical isoperimetric inequality on curves can be recovered as a consequence.

Let us first recall that the signed area of a curve in 𝒞~\tilde{\mathcal{C}} with Lipschitz parametrization c𝒞c\in\mathcal{C} is given by (cf [46] Chap 1.10):

(6) Area(c)=12𝕊1c(θ),Nc(θ)|c(θ)|𝑑θ=12𝕊1det(c(θ),c(θ))dθ\text{Area}(c)=-\frac{1}{2}\int_{\mathbb{S}^{1}}\langle c(\theta),N_{c}(\theta)\rangle|c^{\prime}(\theta)|d\theta=\frac{1}{2}\int_{\mathbb{S}^{1}}\det(c(\theta),c^{\prime}(\theta))d\theta

where NcN_{c} denotes the unit normal vector to the curve. Note that for a simple and positively oriented curve, (6) is the usual area enclosed by this curve. We begin by reminding a few preliminary properties of the signed area. The first one is that the signed area is additive with respect to the “gluing” of two cycles, namely:

Lemma 4.1.

Let c𝒞c\in\mathcal{C} and 0θ1<θ2<2π0\leq\theta_{1}<\theta_{2}<2\pi. Consider any given Lipschitz open curve γ:[θ1,θ2]\gamma:[\theta_{1},\theta_{2}]\rightarrow\mathbb{C} with γ(θ1)=c(θ1)\gamma(\theta_{1})=c(\theta_{1}) and γ(θ2)=c(θ2)\gamma(\theta_{2})=c(\theta_{2}) and denote γˇ\check{\gamma} the same curve but with opposite orientation. Define c1c_{1}, c2c_{2} the two closed curves in 𝒞\mathcal{C} obtained by respectively concatenating c(𝕊1\[θ1,θ2])c(\mathbb{S}^{1}\backslash[\theta_{1},\theta_{2}]) with γ\gamma and c([θ1,θ2])c([\theta_{1},\theta_{2}]) with γˇ\check{\gamma}. Then:

Area(c)=Area(c1)+Area(c2)\text{Area}(c)=\text{Area}(c_{1})+\text{Area}(c_{2})
Proof.

This is just a direct verification from the definition (6). Indeed, we have γˇ(θ)=γ(θ1+θ2θ)\check{\gamma}(\theta)=\gamma(\theta_{1}+\theta_{2}-\theta) for θ[θ1,θ2]\theta\in[\theta_{1},\theta_{2}] and:

2Area(c)=𝕊1\[θ1,θ2]det(c(θ),c(θ))dθ+θ1θ2det(c(θ),c(θ))dθ\displaystyle 2\text{Area}(c)=\int_{\mathbb{S}^{1}\backslash[\theta_{1},\theta_{2}]}\det(c(\theta),c^{\prime}(\theta))d\theta+\int_{\theta_{1}}^{\theta_{2}}\det(c(\theta),c^{\prime}(\theta))d\theta
=𝕊1\[θ1,θ2]det(c(θ),c(θ))dθ+θ1θ2det(γ(θ),γ(θ))dθ+θ1θ2det(c(θ),c(θ))dθθ1θ2det(γ(θ),γ(θ))dθ\displaystyle=\int_{\mathbb{S}^{1}\backslash[\theta_{1},\theta_{2}]}\det(c(\theta),c^{\prime}(\theta))d\theta+\int_{\theta_{1}}^{\theta_{2}}\det(\gamma(\theta),\gamma^{\prime}(\theta))d\theta+\int_{\theta_{1}}^{\theta_{2}}\det(c(\theta),c^{\prime}(\theta))d\theta-\int_{\theta_{1}}^{\theta_{2}}\det(\gamma(\theta),\gamma^{\prime}(\theta))d\theta
=𝕊1\[θ1,θ2]det(c(θ),c(θ))dθ+θ1θ2det(γ(θ),γ(θ))dθ=2Area(c1)+θ1θ2det(c(θ),c(θ))dθ+θ1θ2det(γˇ(θ),γˇ(θ))dθ=2Area(c2)\displaystyle=\underbrace{\int_{\mathbb{S}^{1}\backslash[\theta_{1},\theta_{2}]}\det(c(\theta),c^{\prime}(\theta))d\theta+\int_{\theta_{1}}^{\theta_{2}}\det(\gamma(\theta),\gamma^{\prime}(\theta))d\theta}_{=2\text{Area}(c_{1})}+\underbrace{\int_{\theta_{1}}^{\theta_{2}}\det(c(\theta),c^{\prime}(\theta))d\theta+\int_{\theta_{1}}^{\theta_{2}}\det(\check{\gamma}(\theta),\check{\gamma}^{\prime}(\theta))d\theta}_{=2\text{Area}(c_{2})}

which leads to the result. ∎

We also have the following well-known approximation property of the signed area:

Lemma 4.2.

Let c𝒞c\in\mathcal{C}. There exists a sequence (pn)(p_{n}) of polytopes (i.e. piecewise linear curves) in 𝒞\mathcal{C} such that Area(pn)Area(c)\text{Area}(p_{n})\rightarrow\text{Area}(c) and μpn\mu_{p_{n}} converges weakly to μc\mu_{c}.

Proof.

Let c𝒞c\in\mathcal{C} and by invariance to translation, let’s also assume that c(0)=0c(0)=0. Then since cc is Lipschitz regular, we have cL(𝕊1,)L1(𝕊1,)c^{\prime}\in L^{\infty}(\mathbb{S}^{1},\mathbb{C})\subset L^{1}(\mathbb{S}^{1},\mathbb{C}). Using the density of step functions in L1L^{1}, we can construct a sequence (ρn)n(\rho_{n})_{n\in\mathbb{N}} of step functions such that ρncL10\|\rho_{n}-c^{\prime}\|_{L^{1}}\rightarrow 0 as n+n\rightarrow+\infty. By the converse of Lebesgue’s dominated convergence theorem ([9] Theorem 4.9), up to extraction of a subsequence, we can assume that ρn(θ)\rho_{n}(\theta) converges to c(θ)c^{\prime}(\theta) almost everywhere in 𝕊1\mathbb{S}^{1} and that there exists hL1(𝕊1)h\in L^{1}(\mathbb{S}^{1}) such that ρn(θ)h(θ)\rho_{n}(\theta)\leq h(\theta) for almost all θ𝕊1\theta\in\mathbb{S}^{1}. Let us then define pn=0θρn(α)𝑑αp_{n}=\int_{0}^{\theta}\rho_{n}(\alpha)d\alpha which is a piecewise linear curve in 𝒞\mathcal{C} such that:

|pn(θ)c(θ)|0θ|ρn(α)c(α)|𝑑α𝕊1|ρn(α)c(α)|𝑑α=ρncL1\displaystyle|p_{n}(\theta)-c(\theta)|\leq\int_{0}^{\theta}|\rho_{n}(\alpha)-c^{\prime}(\alpha)|d\alpha\leq\int_{\mathbb{S}^{1}}|\rho_{n}(\alpha)-c^{\prime}(\alpha)|d\alpha=\|\rho_{n}-c^{\prime}\|_{L^{1}}

showing that pncL0\|p_{n}-c\|_{L^{\infty}}\rightarrow 0 as n+n\rightarrow+\infty and as a consequence there exists M>0M>0 such that pnLM\|p_{n}\|_{L^{\infty}}\leq M for all nn. Now, by Proposition 2.11, we deduce that μpn\mu_{p_{n}} weakly converges to μc\mu_{c}. Furthermore, for any nn\in\mathbb{N}, we have:

Area(pn)=12𝕊1pn(θ),Npn(θ)|pn(θ)|𝑑θ=12𝕊1pn(θ),Npn(θ)|ρn(θ)|𝑑θ\text{Area}(p_{n})=-\frac{1}{2}\int_{\mathbb{S}^{1}}\langle p_{n}(\theta),N_{p_{n}}(\theta)\rangle|p_{n}^{\prime}(\theta)|d\theta=-\frac{1}{2}\int_{\mathbb{S}^{1}}\langle p_{n}(\theta),N_{p_{n}}(\theta)\rangle|\rho_{n}(\theta)|d\theta

and for almost all θ𝕊1\theta\in\mathbb{S}^{1}, pn(θ)c(θ)p_{n}(\theta)\rightarrow c(\theta), ρn(θ)c(θ)\rho_{n}(\theta)\rightarrow c^{\prime}(\theta), Npn(θ)=Rπ/2(ρn(θ)/|ρn(θ)|)Nc(θ)N_{p_{n}}(\theta)=R_{\pi/2}(\rho_{n}(\theta)/|\rho_{n}(\theta)|)\rightarrow N_{c}(\theta) as n+n\rightarrow+\infty. In addition,

|pn(θ),Npn(θ)||ρn(θ)||pn(θ)||ρn(θ)|Mh(θ).|\langle p_{n}(\theta),N_{p_{n}}(\theta)\rangle||\rho_{n}(\theta)|\leq|p_{n}(\theta)||\rho_{n}(\theta)|\leq Mh(\theta).

As hL1(𝕊1)h\in L^{1}(\mathbb{S}^{1}), Lebesgue dominated convergence theorem leads to:

Area(pn)n+12𝕊1c(θ),Nc(θ)|c(θ)|𝑑θ=Area(c).\text{Area}(p_{n})\xrightarrow[n\rightarrow+\infty]{}-\frac{1}{2}\int_{\mathbb{S}^{1}}\langle c(\theta),N_{c}(\theta)\rangle|c^{\prime}(\theta)|d\theta=\text{Area}(c).

We will also need the following which is a particular case of our main result for polytopes with fixed number of edges, which proof is adapted from the one outlined in [21, 8]. We call a polytope non-degenerate if it does not lie entirely along a single line (in other words if its length measure is not a sum of two oppositely oriented Diracs).

Refer to caption Refer to caption Refer to caption
Figure 5. Illustration of the proof of Lemma 4.3: the edges of a non-convex polytope can be rearranged to obtain a polytope of larger signed area.
Lemma 4.3.

Let pp be a non-degenerate polytope. Then the convex polygon p¯\bar{p} such that μp¯=μp\mu_{\bar{p}}=\mu_{p} satisfies Area(p)Area(p¯)\text{Area}(p)\leq\text{Area}(\bar{p}).

Proof.

Let pp be a polytope that we represent by its ordered list of vertices v1,v2,,vNv_{1},v_{2},\ldots,v_{N}. Its corresponding list of edges are denoted by =(e1,e2,,eN)\mathcal{E}=(e_{1},e_{2},\ldots,e_{N}) where we assume the directions of the eie_{i} to be consecutively distinct. Recall that, from Theorem 3.2, there is a unique convex and positively oriented polygon p¯\bar{p} such that μp¯=μp\mu_{\bar{p}}=\mu_{p}. Let us define SS_{\mathcal{E}} as the set of polygonal curves obtained by all the different permutations of the edges in \mathcal{E}. Note that for any polygonal curve p~S\tilde{p}\in S_{\mathcal{E}}, we have μp~=μp\mu_{\tilde{p}}=\mu_{p}. Let us show that:

(7) Area(p¯)=maxp~SArea(p~)\text{Area}(\bar{p})=\max_{\tilde{p}\in S_{\mathcal{E}}}\ \ \text{Area}(\tilde{p})

which will imply in particular that Area(p)Area(p¯)\text{Area}(p)\leq\text{Area}(\bar{p}). Since SS_{\mathcal{E}} is finite, the maximum is attained: let p^\hat{p} be a curve in SS_{\mathcal{E}} of maximal signed area. By contradiction, assume that p^\hat{p} is not the convex positively oriented polygon p¯\bar{p} (modulo translations).

Let us first consider the trivial cases. When N=3N=3, there are only two triangles up to translation in SS_{\mathcal{E}} and thus p^\hat{p} is the negatively oriented one which signed are is clearly strictly smaller than the one of p¯\bar{p}. For N=4N=4, the polytope p^\hat{p} being distinct from p¯\bar{p}, one can see that, up to a cyclic permutation of its edges, we can assume without loss of generality that det(e1,e2)<0\det(e_{1},e_{2})<0, i.e. the angle is decreasing between the first and second edge. Now the signed area (6) of the piecewise linear curve p^\hat{p} can be written more simply as (cf [46] Chap 1.10):

Area(p^)=12k=1Ndet(vvk,ek)\text{Area}(\hat{p})=\frac{1}{2}\sum_{k=1}^{N}\det(\overrightarrow{vv_{k}},e_{k})

where v1=O,v2=e1,v3=e1+e2,v4=e1+e2+e3v_{1}=O,v_{2}=e_{1},v_{3}=e_{1}+e_{2},v_{4}=e_{1}+e_{2}+e_{3} are the consecutive vertices of p^\hat{p} and vv is any reference point in the plane. The above expression is independent of the choice of this reference point, thus choosing v=v1v=v_{1}, we obtain after simplifications:

Area(p^)=12[det(e1,e2)+det(e1,e3)+det(e2,e3)].\text{Area}(\hat{p})=\frac{1}{2}[\det(e_{1},e_{2})+\det(e_{1},e_{3})+\det(e_{2},e_{3})].

Then we can define the quadrilateral p~\tilde{p} in which the ordering of the edges e1e_{1} and e2e_{2} is switched. This is still a polytope of SS_{\mathcal{E}} and we have:

Area(p~)\displaystyle\text{Area}(\tilde{p}) =12[det(e2,e1)+det(e2,e3)+det(e1,e3)]\displaystyle=\frac{1}{2}[\det(e_{2},e_{1})+\det(e_{2},e_{3})+\det(e_{1},e_{3})]
=12[det(e1,e2)+det(e1,e3)+det(e2,e3)]\displaystyle=\frac{1}{2}[-\det(e_{1},e_{2})+\det(e_{1},e_{3})+\det(e_{2},e_{3})]
>12[det(e1,e2)+det(e1,e3)+det(e2,e3)]=Area(p^).\displaystyle>\frac{1}{2}[\det(e_{1},e_{2})+\det(e_{1},e_{3})+\det(e_{2},e_{3})]=\text{Area}(\hat{p}).

which contradicts the fact that Area(p^)\text{Area}(\hat{p}) is maximal.

Let us now assume N5N\geq 5. Then, by the assumption on p^\hat{p}, we can find four vertices vi,vi+1,vj,vj+1v_{i},v_{i+1},v_{j},v_{j+1} (i<j1i<j-1) such that the corresponding quadrilateral with edges vivi+1,vi+1vj,vjvj+1,vj+1vi\overrightarrow{v_{i}v_{i+1}},\overrightarrow{v_{i+1}v_{j}},\overrightarrow{v_{j}v_{j+1}},\overrightarrow{v_{j+1}v_{i}} is not both convex and positively oriented. Let us write p1p_{1} this quadrilateral and introduce in addition the polytope p2p_{2} with successive vertices v1,,vi,vj+1,vj+2,,vNv_{1},\ldots,v_{i},v_{j+1},v_{j+2},\ldots,v_{N} and p3p_{3} the polytope with vertices vi+1,vi+2,vjv_{i+1},v_{i+2},\ldots v_{j}. This amounts in decomposing the original polygon into three distinct cycles as depicted in Figure 5 (left). By the additivity property of the signed area from Lemma 4.1, this implies that Area(p^)=Area(p1)+Area(p2)+Area(p3)\text{Area}(\hat{p})=\text{Area}(p_{1})+\text{Area}(p_{2})+\text{Area}(p_{3}). Now, from the case N=4N=4 treated above, we can find a permutation of the edges of p1p_{1} giving the convex positively oriented quadrilateral p¯1\bar{p}_{1} which satisfies Area(p¯1)>Area(p1)\text{Area}(\bar{p}_{1})>\text{Area}(p_{1}). We then obtain the three polytopes p¯1\bar{p}_{1}, p2p_{2} and p3p_{3} shown respectively in green, blue and red in the example of Figure 5. By translation of p3p_{3}, we can superpose the edge vi+1vj\overrightarrow{v_{i+1}v_{j}} in p¯1\bar{p}_{1} with the edge vjvi+1\overrightarrow{v_{j}v_{i+1}} in p3p_{3}. Similarly, by translation of p1p_{1}, we match the edge vj+1vi\overrightarrow{v_{j+1}v_{i}} of p1p_{1} on the edge vivj+1\overrightarrow{v_{i}v_{j+1}} of p2p_{2}. Then, removing the trivial back and forth edges resulting from this superposition, one obtains a new polytope p~\tilde{p} as shown in the right image in Figure 5. By construction, this polytope has the same list of edge vectors as p^\hat{p} and thus belongs to SS_{\mathcal{E}}. Moreover, its signed area is:

Area(p~)=Area(p¯1)+Area(p2)+Area(p3)>Area(p1)+Area(p2)+Area(p3)=Area(p^)\text{Area}(\tilde{p})=\text{Area}(\bar{p}_{1})+\text{Area}(p_{2})+\text{Area}(p_{3})>\text{Area}(p_{1})+\text{Area}(p_{2})+\text{Area}(p_{3})=\text{Area}(\hat{p})

which contradicts the maximality of Area(p^)\text{Area}(\hat{p}) among polytopes of SS_{\mathcal{E}}. ∎

The convex polygon p¯\bar{p} is sometimes referred to as the convexification of pp (which is distinct from the convex hull of pp). We now state and prove the main result of this section.

Theorem 4.4.

Let μ0+\mu\in\mathcal{M}^{+}_{0} such that the support of μ\mu is not of the form {θ,θ+π}\{\theta,\theta+\pi\} for some θ𝕊1\theta\in\mathbb{S}^{1}. Among all curves in M1({μ})M^{-1}(\{\mu\}), the convex positively oriented curve is the unique maximum of the signed area and this maximum is given by:

(8) Amax(μ)=1402π02πsin|θα|dμ(θ)𝑑μ(α).A_{max}(\mu)=\frac{1}{4}\int_{0}^{2\pi}\int_{0}^{2\pi}\sin|\theta-\alpha|d\mu(\theta)d\mu(\alpha).
Proof.

Let μ\mu be a measure in 0+\mathcal{M}^{+}_{0} satisfying the above assumption on its support. This means that for any curve cc in M1({μ})M^{-1}(\{\mu\}), the image of cc does not lie within a single straight line. Furthermore, we have Length(c)=μ(𝕊1)L\text{Length}(c)=\mu(\mathbb{S}^{1})\doteq L. By the standard isoperimetric inequality, this implies that the area of any simple closed curve in M1({μ})M^{-1}(\{\mu\}) is bounded by L24π\frac{L^{2}}{4\pi} and thus the supremum of these areas AmaxA_{max} is indeed finite. Let us first show that this maximal area is achieved and by the convex positively oriented curve of M1({μ})M^{-1}(\{\mu\}). We let (cn)n(c_{n})_{n\in\mathbb{N}} be a maximizing sequence i.e cnM1({μ})c_{n}\in M^{-1}(\{\mu\}) and Area(cn)Amax\text{Area}(c_{n})\rightarrow A_{max} as n+n\rightarrow+\infty. 1. As a first step, we want to replace this maximizing sequence by a sequence of piecewise linear curves. For a fixed nn\in\mathbb{N}, by Lemma 4.2, we can construct a sequence of polytopes (p~n,m)m(\tilde{p}_{n,m})_{m\in\mathbb{N}} such that Area(p~n,m)Area(cn)\text{Area}(\tilde{p}_{n,m})\rightarrow\text{Area}(c_{n}) and μp~n,m\mu_{\tilde{p}_{n,m}} weakly converges to μcn=μ\mu_{c_{n}}=\mu as m+m\rightarrow+\infty. From this, let us construct a sequence of polytopes pnp_{n} such that μpnμc\mu_{p_{n}}\rightharpoonup\mu_{c} and Area(pn)Amax\text{Area}(p_{n})\rightarrow A_{max} as n+n\rightarrow+\infty. Indeed, recall that the weak convergence of finite measures of 𝕊1\mathbb{S}^{1} is metrizable, for instance by the bounded Lipschitz distance dBL(μ,ν)supfLip1(𝕊1)|(μ|f)(ν|f)|d^{BL}(\mu,\nu)\doteq\sup_{f\in\text{Lip}^{1}(\mathbb{S}^{1})}|(\mu|f)-(\nu|f)| ([45] Chap. 6). Thus, we have for all nn\in\mathbb{N}, dBL(μp~n,m,μ)0d^{BL}(\mu_{\tilde{p}_{n,m}},\mu)\rightarrow 0 as m+m\rightarrow+\infty. Since Area(cn)Amax\text{Area}(c_{n})\rightarrow A_{max}, for any nn\in\mathbb{N}, we can find an increasing function ϕ:\phi:\mathbb{N}\rightarrow\mathbb{N} such that |Area(cϕ(n))Amax|1n|\text{Area}(c_{\phi(n)})-A_{max}|\leq\frac{1}{n}. In addition, we also obtain an increasing function ψ:\psi:\mathbb{N}\rightarrow\mathbb{N} such that for any nn\in\mathbb{N}, we have dBL(μp~n,ψ(n),μ)1/nd^{BL}(\mu_{\tilde{p}_{n,\psi(n)}},\mu)\leq 1/n and |Area(p~n,ψ(n))Area(cn)|1/n|\text{Area}(\tilde{p}_{n,\psi(n)})-\text{Area}(c_{n})|\leq 1/n. We then set pnp~ϕ(n),ψ(ϕ(n))p_{n}\doteq\tilde{p}_{\phi(n),\psi(\phi(n))} which gives on the one hand:

dBL(μp~n,μ)=dBL(μp~ϕ(n),ψ(ϕ(n)),μ)1ϕ(n)1nd^{BL}(\mu_{\tilde{p}_{n}},\mu)=d^{BL}(\mu_{\tilde{p}_{\phi(n),\psi(\phi(n))}},\mu)\leq\frac{1}{\phi(n)}\leq\frac{1}{n}

and thus μpnμ\mu_{p_{n}}\rightharpoonup\mu. And on the other hand:

|Area(pn)Amax|\displaystyle|\text{Area}(p_{n})-A_{max}| +|Area(p~ϕ(n),ψ(ϕ(n)))Area(cϕ(n))|+|Area(cϕ(n))Amax|\displaystyle\leq+|\text{Area}(\tilde{p}_{\phi(n),\psi(\phi(n))})-\text{Area}(c_{\phi(n)})|+|\text{Area}(c_{\phi(n)})-A_{max}|
1ϕ(n)+1n2n\displaystyle\leq\frac{1}{\phi(n)}+\frac{1}{n}\leq\frac{2}{n}

and therefore Area(pn)Amax\text{Area}(p_{n})\rightarrow A_{max}. 2. Now, using Lemma 4.3, we obtain a sequence of convex, positively oriented polygons (p¯n)(\bar{p}_{n}) such that μp¯n=μpnμ\mu_{\bar{p}_{n}}=\mu_{p_{n}}\rightharpoonup\mu and for all nn, Area(p¯n)Area(pn)\text{Area}(\bar{p}_{n})\geq\text{Area}(p_{n}). Using once again the invariance to translation, we can further assume that each p¯n\bar{p}_{n} passes through the origin. We then point out that the length L(p¯n)=μp¯n(𝕊1)=μpn(𝕊1)L(\bar{p}_{n})=\mu_{\bar{p}_{n}}(\mathbb{S}^{1})=\mu_{p_{n}}(\mathbb{S}^{1}) converges to μ(𝕊1)<+\mu(\mathbb{S}^{1})<+\infty as n+n\rightarrow+\infty since μpnμ\mu_{p_{n}}\rightharpoonup\mu. This implies that L(p¯n)L(\bar{p}_{n}) is bounded uniformly in nn by some constant M>0M>0. If we denote by P¯n\bar{P}_{n} the polygonal domain delimited by p¯n\bar{p}_{n} i.e. P¯n=p¯n\partial\bar{P}_{n}=\bar{p}_{n}, it is then easy to see that for all nn\in\mathbb{N}, P¯n\bar{P}_{n} is included in the fixed ball B(0,M)B(0,M). In other words, the sequence (P¯n)(\bar{P}_{n}) is bounded in the space of compact subsets of the plane equipped with the Hausdorff metric. 3. We can therefore apply Blaschke selection theorem [7] which allows to assume, up to extraction of a subsequence, that the sequence of convex polygonal domains P¯n\bar{P}_{n} converges in Hausdorff distance to a convex domain CC which oriented boundary curve we write C=cC~conv\partial C=c\in\tilde{C}_{conv}. Then by Proposition 3.4, we deduce that μp¯n\mu_{\bar{p}_{n}} weakly converges to μc\mu_{c}. As a consequence, μc=μ\mu_{c}=\mu and cc is thus the (unique) convex curve associated to μ\mu given by Theorem 3.2. In addition, still from Proposition 3.4, we get that λ2(P¯n)λ2(C)\lambda^{2}(\bar{P}_{n})\rightarrow\lambda^{2}(C) as n+n\rightarrow+\infty and since the boundary curves p¯n\bar{p}_{n} and cc are simple and positively oriented it follows that Area(p¯n)Area(c)\text{Area}(\bar{p}_{n})\rightarrow\text{Area}(c) as n+n\rightarrow+\infty. Now, since for all nn\in\mathbb{N}, Area(p¯n)Area(pn)Amax\text{Area}(\bar{p}_{n})\geq\text{Area}(p_{n})\rightarrow A_{max}, we conclude that Area(c)Amax\text{Area}(c)\geq A_{max} and therefore cc achieves the maximal area among all curves in M1({μ})M^{-1}(\{\mu\}).

Refer to caption Refer to caption
Figure 6. Illustration of step 4 in the proof of Theorem 4.4. By rearranging the different subcycles in the curve c~\tilde{c}, one obtains a new curve in M1({μ})M^{-1}(\{\mu\}) of larger signed area.

4. The uniqueness of the maximum can be now showed by essentially adapting the argument used in the proof of Lemma 4.3 for polytopes. Assume by contradiction that c~M1(μ)\tilde{c}\in M^{-1}(\mu) is non-convex and that Area(c~)=Amax(μ)\text{Area}(\tilde{c})=A_{max}(\mu). Then we can find 0θ1<θ2<θ3<θ4<2π0\leq\theta_{1}<\theta_{2}<\theta_{3}<\theta_{4}<2\pi such that the quadrilateral QQ with successive vertices c~(θ1),c~(θ2),c~(θ3),c~(θ4)\tilde{c}(\theta_{1}),\tilde{c}(\theta_{2}),\tilde{c}(\theta_{3}),\tilde{c}(\theta_{4}) is not convex positively oriented and is not degenerate. Let us further introduce c1c_{1} defined as the concatenation of the portion of curve c~([0,θ1])\tilde{c}([0,\theta_{1}]), the oriented segment [c(θ1),c(θ4)][c(\theta_{1}),c(\theta_{4})] and c~([θ4,2π])\tilde{c}([\theta_{4},2\pi]). In addition, let c2c_{2} be the concatenation of c~([θ1,θ2])\tilde{c}([\theta_{1},\theta_{2}]) and the segment [c(θ2),c(θ1)][c(\theta_{2}),c(\theta_{1})], c3c_{3} the curve obtained by concatenating c~([θ2,θ3])\tilde{c}([\theta_{2},\theta_{3}]) and the segment [c~(θ3),c~(θ2)][\tilde{c}(\theta_{3}),\tilde{c}(\theta_{2})] and c4c_{4} the concatenation of c~([θ3,θ4])\tilde{c}([\theta_{3},\theta_{4}]) and [c~(θ4),c~(θ3)][\tilde{c}(\theta_{4}),\tilde{c}(\theta_{3})]. See Figure 6 for an illustration. Now from Lemma 4.1, we deduce that:

Area(c~)=Area(Q)+Area(c1)+Area(c2)+Area(c3)+Area(c4).\text{Area}(\tilde{c})=\text{Area}(Q)+\text{Area}(c_{1})+\text{Area}(c_{2})+\text{Area}(c_{3})+\text{Area}(c_{4}).

By reordering of the edges in QQ, we obtain the convexified quadrilateral Q¯\bar{Q} which, with the same argument used in the proof of Lemma 4.3, is such that Area(Q)<Area(Q¯)\text{Area}(Q)<\text{Area}(\bar{Q}). We can then rearrange the different cycles introduced above accordingly, as shown in Figure 6 (right). This is done by translating c1,c2,c3,c4c_{1},c_{2},c_{3},c_{4} to match the corresponding edges in Q¯\bar{Q}. As a result, we obtain a new closed curve c¯\bar{c} which by construction still belongs to M1({μ})M^{-1}(\{\mu\}) since it is just obtained by interchanging sections of c~\tilde{c}. Then applying again Lemma 4.1:

Area(c¯)=Area(Q¯)+Area(c1)+Area(c2)+Area(c3)+Area(c4)>Area(c~).\text{Area}(\bar{c})=\text{Area}(\bar{Q})+\text{Area}(c_{1})+\text{Area}(c_{2})+\text{Area}(c_{3})+\text{Area}(c_{4})>\text{Area}(\tilde{c}).

This contradicts the fact that Area(c~)=Amax(μ)\text{Area}(\tilde{c})=A_{max}(\mu). 5. Finally, the expression of the area Amax=Area(c)A_{max}=\text{Area}(c) of the convex curve cc with respect to its length measure is well-known and can be recovered for example from [34], although the definition and presentation of length measures is slightly different than in the present paper. For completeness, we provide a direct proof. It suffices to show (8) for a convex polygon and the general case will directly follow from the above approximation argument. Let p¯\bar{p} be a convex polygon with ordered list of edges =(e1,e2,,eN)\mathcal{E}=(e_{1},e_{2},\ldots,e_{N}) and vertices v1=O,v2=e1,,vN=e1+e2++eN1v_{1}=O,v_{2}=e_{1},\ldots,v_{N}=e_{1}+e_{2}+\ldots+e_{N-1}. Its length measure is then μ=i=1Nliδαi\mu=\sum_{i=1}^{N}l_{i}\delta_{\alpha_{i}} where li=eil_{i}=\|e_{i}\| and αi=eiei𝕊1\alpha_{i}=\frac{e_{i}}{\|e_{i}\|}\in\mathbb{S}^{1}. Since the polygon is convex and positively oriented, the successive angles of the edges are non-decreasing so we may assume without loss of generality that 0α1<α2<<αN<2π0\leq\alpha_{1}<\alpha_{2}<\ldots<\alpha_{N}<2\pi. Therefore the right hand side of (8) is:

1402π02πsin|θα|dμ(θ)𝑑μ(α)=14i=1Nj=1Nliljsin|αiαj|.\frac{1}{4}\int_{0}^{2\pi}\int_{0}^{2\pi}\sin|\theta-\alpha|d\mu(\theta)d\mu(\alpha)=\frac{1}{4}\sum_{i=1}^{N}\sum_{j=1}^{N}l_{i}l_{j}\sin|\alpha_{i}-\alpha_{j}|.

By splitting the inner sum into two sums from j=1,,ij=1,\ldots,i and j=i+1,,Nj=i+1,\ldots,N, after an easy calculation, we obtain:

1402π02πsin|θα|dμ(θ)𝑑μ(α)=12i=1Nj=1i1liljsin|αiαj|=12i=1Nj=1i1liljsin(αiαj)\frac{1}{4}\int_{0}^{2\pi}\int_{0}^{2\pi}\sin|\theta-\alpha|d\mu(\theta)d\mu(\alpha)=\frac{1}{2}\sum_{i=1}^{N}\sum_{j=1}^{i-1}l_{i}l_{j}\sin|\alpha_{i}-\alpha_{j}|=\frac{1}{2}\sum_{i=1}^{N}\sum_{j=1}^{i-1}l_{i}l_{j}\sin(\alpha_{i}-\alpha_{j})

since αi>αj\alpha_{i}>\alpha_{j} for all j=1,,i1j=1,\ldots,i-1. This is in turn leads to:

(9) 1402π02πsin|θα|dμ(θ)𝑑μ(α)=12i=1Nj=1i1det(ej,ei)=12i=1Ndet(Ovi,ei)\frac{1}{4}\int_{0}^{2\pi}\int_{0}^{2\pi}\sin|\theta-\alpha|d\mu(\theta)d\mu(\alpha)=\frac{1}{2}\sum_{i=1}^{N}\sum_{j=1}^{i-1}\det(e_{j},e_{i})=\frac{1}{2}\sum_{i=1}^{N}\det(Ov_{i},e_{i})

since j=1i1ej=Ovi\sum_{j=1}^{i-1}e_{j}=\overrightarrow{Ov_{i}}. Now the right hand side of (9) is exactly the signed area of p¯\bar{p}. ∎

Another useful expression of the maximal area Amax(μ)A_{max}(\mu) is through the Fourier coefficients of the measure μ\mu. Let us define:

μ^(n)=12π𝕊1einθ𝑑μ(θ).\hat{\mu}(n)=\frac{1}{2\pi}\int_{\mathbb{S}^{1}}e^{-in\theta}d\mu(\theta).
Proposition 4.5.

For all μ0+\mu\in\mathcal{M}^{+}_{0}, the following holds:

(10) Amax(μ)=14πnn±1|μ^(n)|21n2A_{max}(\mu)=\frac{1}{4\pi}\sum_{\begin{subarray}{c}n\in\mathbb{Z}\\ n\neq\pm 1\end{subarray}}\frac{|\hat{\mu}(n)|^{2}}{1-n^{2}}
Proof.

Let θ[0,2π)\theta\in[0,2\pi) be for now fixed and fθ(α)=sin|θα|f_{\theta}(\alpha)=\sin|\theta-\alpha|. After some calculations that we skip for the sake of concision, one can show that the Fourier coefficients of fθf_{\theta} are given by:

fθ^(n)=12π02πeinθfθ(α)dα={1π(1n2)einθ+12π(eiθn1eiθn+1)for n±1eiθ4π+(14π+iθπ2π)eiθfor n=1eiθ4π+(14π+iπθ2π)eiθfor n=1\widehat{f_{\theta}}(n)=\frac{1}{2\pi}\int_{0}^{2\pi}e^{-in\theta}f_{\theta}(\alpha)d\alpha=\left\{\begin{aligned} &\frac{1}{\pi(1-n^{2})}e^{-in\theta}+\frac{1}{2\pi}\left(\frac{e^{-i\theta}}{n-1}-\frac{e^{i\theta}}{n+1}\right)\ \ \text{for }n\neq\pm 1\\ &-\frac{e^{i\theta}}{4\pi}+\left(\frac{1}{4\pi}+i\frac{\theta-\pi}{2\pi}\right)e^{-i\theta}\ \ \text{for }n=1\\ &-\frac{e^{-i\theta}}{4\pi}+\left(\frac{1}{4\pi}+i\frac{\pi-\theta}{2\pi}\right)e^{i\theta}\ \ \text{for }n=-1\end{aligned}\right.

Now, writing fθ(α)=nfθ^(n)einαf_{\theta}(\alpha)=\sum_{n\in\mathbb{Z}}\widehat{f_{\theta}}(n)e^{in\alpha} which converges everywhere on 𝕊1\mathbb{S}^{1} and recalling that 𝕊1eiθ𝑑μ(θ)=𝕊1eiθ𝑑μ(θ)=0\int_{\mathbb{S}^{1}}e^{i\theta}d\mu(\theta)=\int_{\mathbb{S}^{1}}e^{-i\theta}d\mu(\theta)=0, we get that:

Amax(μ)=1402π02πnn±11π(1n2)einθeinαdμ(θ)dμ(α)\displaystyle A_{max}(\mu)=\frac{1}{4}\int_{0}^{2\pi}\int_{0}^{2\pi}\sum_{\begin{subarray}{c}n\in\mathbb{Z}\\ n\neq\pm 1\end{subarray}}\frac{1}{\pi(1-n^{2})}e^{-in\theta}e^{in\alpha}d\mu(\theta)d\mu(\alpha)

and since the series of functions in the above equation is uniformly converging on 𝕊1×𝕊1\mathbb{S}^{1}\times\mathbb{S}^{1}:

Amax(μ)\displaystyle A_{max}(\mu) =14nn±11π(1n2)02π(02πeinθ𝑑μ(θ))einα𝑑μ(α)\displaystyle=\frac{1}{4}\sum_{\begin{subarray}{c}n\in\mathbb{Z}\\ n\neq\pm 1\end{subarray}}\frac{1}{\pi(1-n^{2})}\int_{0}^{2\pi}\left(\int_{0}^{2\pi}e^{-in\theta}d\mu(\theta)\right)e^{in\alpha}d\mu(\alpha)
=14nn±11π(1n2)(2π)2μ^(n)μ^(n)¯\displaystyle=\frac{1}{4}\sum_{\begin{subarray}{c}n\in\mathbb{Z}\\ n\neq\pm 1\end{subarray}}\frac{1}{\pi(1-n^{2})}(2\pi)^{2}\hat{\mu}(n)\overline{\hat{\mu}(n)}
=πnn±1|μ^(n)|2(1n2)\displaystyle=\pi\sum_{\begin{subarray}{c}n\in\mathbb{Z}\\ n\neq\pm 1\end{subarray}}\frac{|\hat{\mu}(n)|^{2}}{(1-n^{2})}

Note that as a direct corollary of Theorem 4.4 and Proposition 4.5, we recover the standard isoperimetric inequality in its general form, that is:

Corollary 4.6.

For any curve cc in 𝒞~\tilde{\mathcal{C}}, we have:

Area(c)L(c)24π\text{Area}(c)\leq\frac{L(c)^{2}}{4\pi}

with equality if and only if cc is a circle.

Proof.

By adequate rescaling, we can first restrict the proof to the case where L(c)=μc(𝕊1)=2πL(c)=\mu_{c}(\mathbb{S}^{1})=2\pi. Then, from Theorem 4.4, we have that Area(c)sup{Amax(μ)|μ0+,μ(𝕊1)=2π}\text{Area}(c)\leq\sup\{A_{max}(\mu)\ |\ \mu\in\mathcal{M}^{+}_{0},\mu(\mathbb{S}^{1})=2\pi\}. Now, for any μ0+\mu\in\mathcal{M}^{+}_{0} such that μ(𝕊1)=2π\mu(\mathbb{S}^{1})=2\pi, we have μ^(0)=μ(𝕊1)=2π\hat{\mu}(0)=\mu(\mathbb{S}^{1})=2\pi and:

Amax(μ)=14πnn±1|μ^(n)|21n214π|μ^(0)|2=π.A_{max}(\mu)=\frac{1}{4\pi}\sum_{\begin{subarray}{c}n\in\mathbb{Z}\\ n\neq\pm 1\end{subarray}}\frac{|\hat{\mu}(n)|^{2}}{1-n^{2}}\leq\frac{1}{4\pi}|\hat{\mu}(0)|^{2}=\pi.

Therefore, Area(c)π=L(c)24π\text{Area}(c)\leq\pi=\frac{L(c)^{2}}{4\pi} with equality if and only if Amax(μ)=πA_{max}(\mu)=\pi which implies that in the above we have μ^(n)=0\hat{\mu}(n)=0 for all |n|1|n|\neq 1. Therefore, μ\mu is the uniform measure on 𝕊1\mathbb{S}^{1} and as Area(c)=Amax(μ)\text{Area}(c)=A_{max}(\mu), by the uniqueness of the maximizer in Theorem 4.4, we obtain that cc is the unit circle. ∎

5. A geodesic space structure on convex sets of the plane

The correspondence between convex curves and length measures that is given by Theorem 3.2 also suggests the idea of comparing convex sets through their associated length measures. In other words, one can transpose the construction of metrics on the set of convex domains to that of building metrics on the measure space 0+\mathcal{M}^{+}_{0}. What makes this advantageous is that there are already many existing and well-known distances between measures that can be introduced to that end, and several past works [47, 1] have exploited this idea for various purposes. Yet most of these works are focused on the computation and/or mathematical properties of the distance only. In the field of interest of the authors, namely shape analysis, an often equally important aspect is to define distances (typically Riemannian or sub-Riemannian) which also lead to relevant geodesics on the shape space: such geodesics indeed provide a way of interpolating between two shapes and are crucial to extend many statistical or machine learning tools to analyze shape datasets. In this section, we will therefore be interested in building numerically computable geodesic distances on 0+\mathcal{M}^{+}_{0} which will in turn induce metrics and corresponding geodesics on C~conv\tilde{C}_{conv}. Note that although we discuss this problem for planar curves as it is the focus of this paper, most of what follows can be adapted to larger dimensions by considering the general area measures of convex domains in n\mathbb{R}^{n}. We first review several standard metrics on the space of measures of 𝕊1\mathbb{S}^{1} and analyze their potential shortcomings when it comes to comparing planar convex curves. We then propose a new constrained optimal transport distance.

5.1. Kernel metrics

As a dual space, the space of measures on 𝕊1\mathbb{S}^{1} can be endowed with dual metrics based on a choice of norm on a set of test functions. These include classical measure metrics such as the Lévy-Prokhorov or the bounded Lipschitz distance. Those distances metrizes the weak convergence between measures and have many appealing mathematical properties. However, they are typically challenging to compute or even approximate in practice. An alternative is the class of metrics derived from reproducing kernel Hilbert spaces (RKHS) which have been widely used in shape analysis [26, 19, 32] and in statistics [28, 35]. We briefly recap and discuss such metrics in our context. The starting point is a continuous and positive definite kernel K:𝕊1×𝕊1K:\mathbb{S}^{1}\times\mathbb{S}^{1}\rightarrow\mathbb{R} to which, by Aronszajn theorem [5], corresponds a unique RKHS of functions on 𝕊1\mathbb{S}^{1}. Let us denote this space by \mathcal{H} and by \mathcal{H}^{*} its dual. Now taking \mathcal{H} as our space of test functions, one can introduce the norm \|\cdot\|_{\mathcal{H}^{*}} on 0+\mathcal{M}^{+}_{0} defined by:

(11) μ=sup{(μ|f)|f,f=1}\|\mu\|_{\mathcal{H}^{*}}=\sup\{(\mu|f)\ |\ f\in\mathcal{H},\ \|f\|_{\mathcal{H}}=1\}

and the corresponding distance d(μ0,μ1)=μ1μ0d_{\mathcal{H}^{*}}(\mu_{0},\mu_{1})=\|\mu_{1}-\mu_{0}\|_{\mathcal{H}^{*}}. In general, (11) only gives a pseudo-distance on 0+\mathcal{M}^{+}_{0} but is shown in [43] Theorem 6 that a necessary and sufficient condition to recover a true distance is for the kernel KK to satisfy a property known as C0C_{0}-universality, which includes several families of well-known kernels such as Gaussian, Cauchy… An important advantage of this RKHS framework is that, thanks to the reproducing kernel property, the distance can be directly expressed based on KK as follows:

d(μ0,μ1)2=\displaystyle d_{\mathcal{H}^{*}}(\mu_{0},\mu_{1})^{2}=
𝕊1×𝕊1K(θ,θ)𝑑μ0(θ)𝑑μ0(θ)2𝕊1×𝕊1K(θ,θ)𝑑μ0(θ)𝑑μ1(θ)+𝕊1×𝕊1K(θ,θ)𝑑μ1(θ)𝑑μ1(θ).\displaystyle\iint_{\mathbb{S}^{1}\times\mathbb{S}^{1}}K(\theta,\theta^{\prime})d\mu_{0}(\theta)d\mu_{0}(\theta^{\prime})-2\iint_{\mathbb{S}^{1}\times\mathbb{S}^{1}}K(\theta,\theta^{\prime})d\mu_{0}(\theta)d\mu_{1}(\theta^{\prime})+\iint_{\mathbb{S}^{1}\times\mathbb{S}^{1}}K(\theta,\theta^{\prime})d\mu_{1}(\theta)d\mu_{1}(\theta^{\prime}).

For discrete measures μ0\mu_{0} and μ1\mu_{1}, the above integrals become double sums and can be all evaluated in closed form once the kernel KK is specified which makes these distances easy to compute in practice. However, since all are essentially dual metrics to some Hilbert space, it is easy to see that the resulting metric space is flat, namely the constant speed geodesic between two positive measures μ0\mu_{0} and μ1\mu_{1} on 𝕊1\mathbb{S}^{1} is given by μ(t)=(1t)μ0+tμ1\mu(t)=(1-t)\mu_{0}+t\mu_{1} for t[0,1]t\in[0,1]. Furthermore, if both μ0\mu_{0} and μ1\mu_{1} belong to 0+\mathcal{M}^{+}_{0} then so does μ(t)\mu(t) for all t[0,1]t\in[0,1] since:

𝕊1eiθ𝑑μ(t)(θ)=(1t)𝕊1eiθ𝑑μ0(θ)+t𝕊1eiθ𝑑μ1(θ)=0.\int_{\mathbb{S}^{1}}e^{i\theta}d\mu(t)(\theta)=(1-t)\int_{\mathbb{S}^{1}}e^{i\theta}d\mu_{0}(\theta)+t\int_{\mathbb{S}^{1}}e^{i\theta}d\mu_{1}(\theta)=0.

Therefore 0+\mathcal{M}^{+}_{0} is a totally geodesic space for the metric \|\cdot\|_{\mathcal{H}^{*}}.

Refer to caption Refer to caption Refer to caption Refer to caption
t=0t=0 t=1/3t=1/3 t=2/3t=2/3 t=1t=1
Refer to caption Refer to caption Refer to caption Refer to caption
Figure 7. Geodesic for the kernel metrics. On top, the intermediate convex curves c(t)C~convc(t)\in\tilde{C}_{conv} and in the bottom row are shown the associated measures μ(t)0+\mu(t)\in\mathcal{M}^{+}_{0}.

Now if we look at the metric on C~conv\tilde{C}_{conv} that results from the identification of convex curves with their length measure in 0+\mathcal{M}^{+}_{0}, we see that, by the isomorphism property of the mapping MM given by Corollary 3.6, the geodesic c(t)=M1(μ(t))c(t)=M^{-1}(\mu(t)) is simply the Minkowski combination of the two convex curves associated to μ0\mu_{0} and μ1\mu_{1}. Note that although the distance will depend on the choice of kernel KK, the geodesics are however independent of KK. While the Minkowski sum may seem like a natural way to interpolate between two convex sets, it may also not be the most optimal from the point of view of shape comparison. Figure 7 shows an example of what a geodesic looks like both in the space of convex sets and in the space of measures. It is for instance clear from this example that the measure geodesic do not involve actual transportation of mass which in this case would be a more natural behavior. This is the shortcoming that the metrics of the following sections will attempt to address.

5.2. Wasserstein metrics

A natural class of distances between probability measures is given by optimal transport and the Wasserstein metrics [45]. Let us consider the usual geodesic distance on 𝕊1\mathbb{S}^{1} defined by d(θ,θ)=min{|θθ|,2π|θθ|}d(\theta,\theta^{\prime})=\min\{|\theta^{\prime}-\theta|,2\pi-|\theta^{\prime}-\theta|\}. Given two probability measures μ0,μ1𝒫(𝕊1)\mu_{0},\mu_{1}\in\mathcal{P}(\mathbb{S}^{1}), the 2-Wasserstein distance between them is defined by:

W2(μ0,μ1)=min{𝕊1×𝕊1d(θ,θ)2𝑑γ|γΠ(μ0,μ1)}W_{2}(\mu_{0},\mu_{1})=\min\left\{\iint_{\mathbb{S}^{1}\times\mathbb{S}^{1}}d(\theta,\theta^{\prime})^{2}d\gamma\ |\ \gamma\in\Pi(\mu_{0},\mu_{1})\right\}

where Π(μ0,μ1)\Pi(\mu_{0},\mu_{1}) is the set of all transport plans between μ0\mu_{0} and μ1\mu_{1} i.e.

Π(μ0,μ1)={γ𝒫(𝕊1×𝕊1)|γ(A×𝕊1)=μ0(A),γ(𝕊1×B)=μ1(B),for all Borel sets A,B𝕊1}.\Pi(\mu_{0},\mu_{1})=\{\gamma\in\mathcal{P}(\mathbb{S}^{1}\times\mathbb{S}^{1})\ |\ \gamma(A\times\mathbb{S}^{1})=\mu_{0}(A),\ \gamma(\mathbb{S}^{1}\times B)=\mu_{1}(B),\ \text{for all Borel sets }A,B\subset\mathbb{S}^{1}\}.

An optimal transport plan γ\gamma can be shown to exist and it is known that the Wasserstein metric makes 𝒫(𝕊1)\mathcal{P}(\mathbb{S}^{1}) into a length space with the corresponding geodesic being given by μ(t)=(πt)γ\mu(t)=(\pi_{t})_{\sharp}\gamma where for all θ,θ𝕊1\theta,\theta^{\prime}\in\mathbb{S}^{1}, tπt(θ,θ)t\mapsto\pi_{t}(\theta,\theta^{\prime}) is a unit constant speed geodesic between θ\theta and θ\theta^{\prime}, that is specifically:

πt(θ,θ)={(1t)θ+tθ,if πθθπ(1t)θ+t(θ2π),if π<θθ<2π(1t)θ+t(θ+2π),if 2π<θθ<π\pi_{t}(\theta,\theta^{\prime})=\left\{\begin{aligned} &(1-t)\theta+t\theta^{\prime},\ \text{if }-\pi\leq\theta^{\prime}-\theta\leq\pi\\ &(1-t)\theta+t(\theta^{\prime}-2\pi),\ \text{if }\pi<\theta^{\prime}-\theta<2\pi\\ &(1-t)\theta+t(\theta^{\prime}+2\pi),\ \text{if }-2\pi<\theta^{\prime}-\theta<-\pi\end{aligned}\right.

Alternatively, this means that for all continuous function f:𝕊1f:\mathbb{S}^{1}\rightarrow\mathbb{R}:

(μ(t)|f)=𝕊1×𝕊1f(πt(θ,θ))𝑑γ(θ,θ).(\mu(t)|f)=\iint_{\mathbb{S}^{1}\times\mathbb{S}^{1}}f(\pi_{t}(\theta,\theta^{\prime}))d\gamma(\theta,\theta^{\prime}).

Unlike in the previous kernel framework, the Wasserstein distance and geodesics cannot be expressed in closed form but there are many well-established and efficient approaches to numerically estimate those, such as combinatorial methods [18] or Sinkhorn algorithm [17] in the situation of discrete measures as well as methods based on the dynamical formulation of optimal transport [6, 37] for densities.

Refer to caption Refer to caption Refer to caption Refer to caption
t=0t=0 t=1/3t=1/3 t=2/3t=2/3 t=1t=1
Refer to caption Refer to caption Refer to caption Refer to caption
Figure 8. Geodesic for the Wasserstein metric computed with the Sinkhorn algorithm. On top, the intermediate convex curves c(t)C~convc(t)\in\tilde{C}_{conv} and in the bottom row are shown the associated measures μ(t)\mu(t). Note that the intermediate measures μ(t)\mu(t) do not belong to 0+\mathcal{M}^{+}_{0} which results in an opening of the reconstructed curves.

When it comes to the comparison of convex shapes, the use of Wasserstein metrics on the length or area measures was proposed for instance in [47]. Note that since W2W_{2} is defined between probability measures, this amounts in restricting to convex curves of length 11 i.e. comparing curves modulo rescaling which is often natural in shape analysis. However, an important downside of the Wasserstein metric for this problem is that the subspace of probability measures in 0+\mathcal{M}^{+}_{0} is not totally geodesic for that metric, namely the above path of measures μ(t)\mu(t) connecting μ0\mu_{0} to μ1\mu_{1} does not generally stay in 0+\mathcal{M}^{+}_{0} when μ0,μ1𝒫(𝕊1)0+\mu_{0},\mu_{1}\in\mathcal{P}(\mathbb{S}^{1})\cap\mathcal{M}^{+}_{0}. This means that the intermediate measures μ(t)\mu(t) cannot be canonically associated to a convex curve (or even to a closed curve as a matter of fact). We illustrate this in Figure 8 that shows the Wasserstein geodesic between the same two discrete measures as in Figure 7 and the curves obtained from the reconstruction procedure of Remark 3.3. This issue could be addressed a posteriori: for example the authors in [47] propose to consider a slightly modified path defined from the optimal transport plan. Specifically, they introduce μ~(t)\tilde{\mu}(t) defined by:

(μ~(t)|f)=𝕊1×𝕊1f((1t)eiθ+teiθ|(1t)eiθ+teiθ|)|(1t)eiθ+teiθ|𝑑γ(θ,θ)(\tilde{\mu}(t)|f)=\iint_{\mathbb{S}^{1}\times\mathbb{S}^{1}}f\left(\frac{(1-t)e^{i\theta}+te^{i\theta^{\prime}}}{|(1-t)e^{i\theta}+te^{i\theta^{\prime}}|}\right)|(1-t)e^{i\theta}+te^{i\theta^{\prime}}|d\gamma(\theta,\theta^{\prime})

for which it is straightforward to check that μ~(t)0+\tilde{\mu}(t)\in\mathcal{M}^{+}_{0} for all t[0,1]t\in[0,1]. A clear remaining downside however is that, despite relying on the optimal transport between μ0\mu_{0} and μ1\mu_{1}, the paths of measures and corresponding convex curves are not actual geodesics associated to a metric.

5.3. A constrained Wasserstein distance on 0+\mathcal{M}^{+}_{0}

In this section we shall instead modify the formulation of the Wasserstein metric so as to directly enforce the constraint defining 0+\mathcal{M}^{+}_{0}. To do so, we will need to shift our focus to the so called dynamical formulation of optimal transport which was first introduced by the works of Benamou and Brenier in [6] and derive a primal-dual scheme adapted from the work of [37] to tackle the corresponding optimization problem.

5.3.1. Mathematical formulation

Formally, the above Wasserstein metric can be obtained by minimizing 01𝕊1|vt|2𝑑μ(t)𝑑t\int_{0}^{1}\int_{\mathbb{S}^{1}}|v_{t}|^{2}d\mu(t)dt over all vector fields vt()v_{t}(\cdot) and path of measures μ(t)\mu(t)\in\mathcal{M} that satisfy the boundary constraints μ(0)=μ0\mu(0)=\mu_{0}, μ(1)=μ1\mu(1)=\mu_{1} and the continuity equation on 𝕊1\mathbb{S}^{1} tμ(t)+θ(vtμ(t))=0\partial_{t}\mu(t)+\partial_{\theta}(v_{t}\mu(t))=0. To make this formulation more rigorous, one has to introduce the following definitions.

Definition 5.1.

Let (ρ,m)(\rho,m) be a couple of measures on [0,1]×𝕊1[0,1]\times\mathbb{S}^{1} which we will write ρ,m([0,1]×𝕊1)\rho,m\in\mathcal{M}([0,1]\times\mathbb{S}^{1}). We say that (ρ,m)(\rho,m) satisfy the continuity equation tρ+θm=0\partial_{t}\rho+\partial_{\theta}m=0 with boundary conditions ρ(0)=μ0\rho(0)=\mu_{0} and ρ(1)=μ1\rho(1)=\mu_{1} in the distribution sense if:

(12) [0,1]×𝕊1tϕdρ+[0,1]×𝕊1θϕdm=𝕊1ϕ(1,)𝑑μ1𝕊1ϕ(0,)𝑑μ0.\int_{[0,1]\times\mathbb{S}^{1}}\partial_{t}\phi\ d\rho+\int_{[0,1]\times\mathbb{S}^{1}}\partial_{\theta}\phi\ dm=\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\mu_{1}-\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\mu_{0}.

for all ϕC1([0,1]×𝕊1)\phi\in C^{1}([0,1]\times\mathbb{S}^{1}).

We recall the following usual property of the continuity equation

Property 5.2.

If (ρ,m)(\rho,m) satisfies the above continuity equation with boundary conditions μ0\mu_{0} and μ1\mu_{1}, then the time marginal of ρ\rho is the Lebesgue measure on [0,1][0,1]. In other words the measure ρ\rho disintegrates as ρ=ρtdt\rho=\rho_{t}\otimes dt.

Proof.

By applying the disintegration theorem on measures of [0,1]×𝕊1[0,1]\times\mathbb{S}^{1} (cf Theorem 2.28 in [4]), we have the existence of finite measures ρt\rho_{t} on 𝕊1\mathbb{S}^{1} for all t[0,1]t\in[0,1] such that we can write ρ=ρtν\rho=\rho_{t}\otimes\nu with ν\nu the measure on [0,1][0,1] given by ν(A)=ρ(A×𝕊1)\nu(A)=\rho(A\times\mathbb{S}^{1}). Taking ϕ(t,θ)=ψ(t)\phi(t,\theta)=\psi(t) with ψC1([0,1])\psi\in C^{1}([0,1]), we get from (12):

[0,1]ψ(t)𝑑ν(t)=ψ(1)ψ(0).\int_{[0,1]}\psi^{\prime}(t)d\nu(t)=\psi(1)-\psi(0).

Since this holds for any ψC1([0,1])\psi\in C^{1}([0,1]), we obtain that ν\nu is the Lebesgue measure on [0,1][0,1]. ∎

Next we define the set B={(a,b)2|a+12|b|20}B=\{(a,b)\in\mathbb{R}^{2}\ |\ a+\frac{1}{2}|b|^{2}\leq 0\} and we will denote ιB\iota_{B} its convex indicator function: ιB(a,b)=0\iota_{B}(a,b)=0 for (a,b)B(a,b)\in B and ιB(a,b)=+\iota_{B}(a,b)=+\infty otherwise. We also define f:×+f:\mathbb{R}\times\mathbb{R}\rightarrow\mathbb{R}_{+} by:

f(d,m)={|m|22dif d>00if (d,m)=(0,0)+otherwisef(d,m)=\left\{\begin{aligned} &\frac{|m|^{2}}{2d}\ \ \text{if }d>0\\ &0\ \ \text{if }(d,m)=(0,0)\\ &+\infty\ \ \text{otherwise}\end{aligned}\right.

Note that ff is a 1-homogeneous function and is the convex conjugate of ιB\iota_{B}. Then it has been shown that given two measures μ0\mu_{0} and μ1\mu_{1}, their Wasserstein distance is equal to:

(13) W2(μ0,μ1)=infν=(ρ,m){[0,1]×𝕊1f(dρdλ)𝑑λ|subj. to ρ0=μ0,ρ1=μ1,tρ+θm=0}.W_{2}(\mu_{0},\mu_{1})=\inf_{\nu=(\rho,m)}\left\{\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\rho}{d\lambda}\right)d\lambda\ |\ \text{subj. to }\rho_{0}=\mu_{0},\ \rho_{1}=\mu_{1},\ \partial_{t}\rho+\partial_{\theta}m=0\right\}.

where λ\lambda is any measure on [0,1]×𝕊1[0,1]\times\mathbb{S}^{1} such that |ν|λ|\nu|\ll\lambda and dνdλ\frac{d\nu}{d\lambda} denotes the Radon-Nykodym derivative (the choice of such λ\lambda does not affect the value in (13) due to the homogeneity of ff). Furthermore, the existence of optimal measures ρ\rho and mm can be proved and the path of measures tρtt\mapsto\rho_{t} given by Property 5.2 is a geodesic between μ0\mu_{0} and μ1\mu_{1}.

We now modify the above expression so as to enforce that the path ρt\rho_{t} remains in the subspace 0+\mathcal{M}^{+}_{0}.

Definition 5.3.

Let μ0\mu_{0} and μ1\mu_{1} be two probability measures in 0+\mathcal{M}^{+}_{0}. With the same conventions as above, we define the constrained Wasserstein distance on 0+\mathcal{M}^{+}_{0} by:

(14) W¯2(μ0,μ1)=infν=(ρ,m){[0,1]×𝕊1f(dνdλ)dλ|subj. to {tρ+θm=0ρ0=μ0,ρ1=μ1𝕊1eiθ𝑑ρt(θ)=0,for a.e t[0,1]}.\overline{W}_{2}(\mu_{0},\mu_{1})=\inf_{\nu=(\rho,m)}\left\{\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\nu}{d\lambda}\right)d\lambda\ |\ \text{subj. to }\left\{\begin{array}[]{lll}\partial_{t}\rho+\partial_{\theta}m=0\\ \rho_{0}=\mu_{0},\ \rho_{1}=\mu_{1}\\ \int_{\mathbb{S}^{1}}e^{i\theta}d\rho_{t}(\theta)=0,\ \text{for a.e }t\in[0,1]\end{array}\right.\right\}.

As we show next, W¯2\overline{W}_{2} defines a distance and we also recover the existence of geodesics.

Theorem 5.4.

For any two probability measures μ0\mu_{0} and μ1\mu_{1} in 0+\mathcal{M}^{+}_{0} such that W¯2(μ0,μ1)<+\overline{W}_{2}(\mu_{0},\mu_{1})<+\infty, there exist (ρ,m)(\rho,m) achieving the minimum in (14). Moreover, W¯2\overline{W}_{2} defines a distance on the space 𝒫(𝕊1)0+\mathcal{P}(\mathbb{S}^{1})\cap\mathcal{M}^{+}_{0}.

Proof.

The proof mainly requires adapting similar arguments developed for other variations of the optimal transport problem [11, 15]. For any ϕ𝒞1([0,1]×𝕊1)\phi\in\mathcal{C}^{1}([0,1]\times\mathbb{S}^{1}) and ψ𝒞([0,1])\psi\in\mathcal{C}([0,1]), let us define

J(ϕ,ψ1,ψ2)\displaystyle J(\phi,\psi_{1},\psi_{2}) =01𝕊1ιB(tϕ+cos(θ)ψ1(t)+sin(θ)ψ2(t),θϕ)𝑑θ𝑑t+𝕊1ϕ(0,)𝑑ρ0𝕊1ϕ(1,)𝑑ρ1\displaystyle=\int_{0}^{1}\int_{\mathbb{S}^{1}}\iota_{B}(\partial_{t}\phi+\cos(\theta)\psi_{1}(t)+\sin(\theta)\psi_{2}(t),\partial_{\theta}\phi)d\theta dt+\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}-\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}
=FA(ϕ,ψ1,ψ2)+G(ϕ,ψ1,ψ2)\displaystyle=F\circ A(\phi,\psi_{1},\psi_{2})+G(\phi,\psi_{1},\psi_{2})

where

F(α,β)=01𝕊1ιB(α(t,θ),β(t,θ))𝑑θ𝑑t\displaystyle F(\alpha,\beta)=\int_{0}^{1}\int_{\mathbb{S}^{1}}\iota_{B}(\alpha(t,\theta),\beta(t,\theta))d\theta dt
A(ϕ,ψ1,ψ2)=(tϕ+cos(θ)ψ1(t)+sin(θ)ψ2(t),θϕ)\displaystyle A(\phi,\psi_{1},\psi_{2})=(\partial_{t}\phi+\cos(\theta)\psi_{1}(t)+\sin(\theta)\psi_{2}(t),\partial_{\theta}\phi)
G(ϕ,ψ1,ψ2)=01𝕊1ϕ(0,)𝑑ρ001𝕊1ϕ(1,)𝑑ρ1\displaystyle G(\phi,\psi_{1},\psi_{2})=\int_{0}^{1}\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}-\int_{0}^{1}\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}

One can easily check that FF and GG are proper convex functions and are lower semi-continuous. Furthermore, taking for instance ϕ(t,θ)=at\phi(t,\theta)=at with a<0a<0 and ψ(t)=0\psi(t)=0, we see that FF is continuous at A(ϕ,ψ)=(a,0)A(\phi,\psi)=(a,0) since (a,0)(a,0) is in the interior of BB. We can thus apply the Fenchel-Rockafellar duality theorem, which leads to

(15) infϕ𝒞1([0,1]×𝕊1),ψ1,ψ2𝒞([0,1])J(ϕ,ψ)=supν([0,1]×𝕊1)×([0,1]×𝕊1){F(ν)G(A(ν))}.\underset{\phi\in\mathcal{C}^{1}([0,1]\times\mathbb{S}^{1}),\psi_{1},\psi_{2}\in\mathcal{C}([0,1])}{\inf}J(\phi,\psi)=\underset{\nu\in\mathcal{M}([0,1]\times\mathbb{S}^{1})\times\mathcal{M}([0,1]\times\mathbb{S}^{1})}{\sup}\left\{-F^{*}(\nu)-G^{*}(-A^{*}(\nu))\right\}.

We now compute for ν=(ρ,m)([0,1]×𝕊1)×([0,1]×𝕊1)\nu=(\rho,m)\in\mathcal{M}([0,1]\times\mathbb{S}^{1})\times\mathcal{M}([0,1]\times\mathbb{S}^{1}):

G(Aν)=supϕ,ψ1,ψ2{ν,A(ϕ,ψ1,ψ2)+𝕊1ϕ(1,)𝑑ρ1𝕊1ϕ(0,)𝑑ρ0}\displaystyle G^{*}(-A^{*}\nu)=\underset{\phi,\psi_{1},\psi_{2}}{\sup}\left\{\langle\nu,-A(\phi,\psi_{1},\psi_{2})\rangle+\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}-\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}\right\}
=supϕ,ψ1,ψ2{ν,(tϕ,θϕ)ν,(cos(θ)ψ1(t)+sin(θ)ψ2(t),0)+𝕊1ϕ(1,)𝑑ρ1𝕊1ϕ(0,)𝑑ρ0}\displaystyle=\underset{\phi,\psi_{1},\psi_{2}}{\sup}\left\{-\langle\nu,(\partial_{t}\phi,\partial_{\theta}\phi)\rangle-\langle\nu,(\cos(\theta)\psi_{1}(t)+\sin(\theta)\psi_{2}(t),0)\rangle+\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}-\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}\right\}
=supϕ,ψ1,ψ2{ν,(tϕ,θϕ)+𝕊1ϕ(1,)𝑑ρ1𝕊1ϕ(0,)𝑑ρ0ν,(cos(θ)ψ1(t)+sin(θ)ψ2(t),0)}\displaystyle=\underset{\phi,\psi_{1},\psi_{2}}{\sup}\left\{-\langle\nu,(\partial_{t}\phi,\partial_{\theta}\phi)\rangle+\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}-\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}-\langle\nu,(\cos(\theta)\psi_{1}(t)+\sin(\theta)\psi_{2}(t),0)\rangle\right\}
=supϕ{ν,(tϕ,ϕ)+𝕊1ϕ(1,)𝑑ρ1𝕊1ϕ(0,)𝑑ρ0}+supψ1,ψ2{ν,(cos(θ)ψ1(t)+sin(θ)ψ2(t),0)}\displaystyle=\underset{\phi}{\sup}\left\{-\langle\nu,(\partial_{t}\phi,\nabla\phi)\rangle+\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}-\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}\right\}+\underset{\psi_{1},\psi_{2}}{\sup}\left\{-\langle\nu,(\cos(\theta)\psi_{1}(t)+\sin(\theta)\psi_{2}(t),0)\rangle\right\}
=supϕ{ν,(tϕ,θϕ)+𝕊1ϕ(1,)𝑑ρ1𝕊1ϕ(0,)𝑑ρ0}+supψ1{[0,1]×𝕊1cos(θ)ψ1(t)𝑑ρ}\displaystyle=\underset{\phi}{\sup}\left\{-\langle\nu,(\partial_{t}\phi,\partial_{\theta}\phi)\rangle+\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}-\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}\right\}+\underset{\psi_{1}}{\sup}\left\{-\int_{[0,1]\times\mathbb{S}^{1}}\cos(\theta)\psi_{1}(t)d\rho\right\}
+supψ2{[0,1]×𝕊1sin(θ)ψ2(t)𝑑ρ}\displaystyle\phantom{\underset{\phi}{\sup}\left\{-\langle\nu,(\partial_{t}\phi,\partial_{\theta}\phi)\rangle+\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}-\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}\right\}+}+\underset{\psi_{2}}{\sup}\left\{-\int_{[0,1]\times\mathbb{S}^{1}}\sin(\theta)\psi_{2}(t)d\rho\right\}

and we find

supϕ{ν,(tϕ,θϕ)+𝕊1ϕ(1,)𝑑ρ1𝕊1ϕ(0,)𝑑ρ0}={0if tρ+θm=0,ρ(0,)=ρ0,ρ(1,)=ρ1+otherwise\underset{\phi}{\sup}\left\{-\langle\nu,(\partial_{t}\phi,\partial_{\theta}\phi)\rangle+\int_{\mathbb{S}^{1}}\phi(1,\cdot)d\rho_{1}-\int_{\mathbb{S}^{1}}\phi(0,\cdot)d\rho_{0}\right\}=\left\{\begin{array}[]{ll}0&\text{if }\partial_{t}\rho+\partial_{\theta}m=0,\rho(0,\cdot)=\rho_{0},\rho(1,\cdot)=\rho_{1}\\ +\infty&\text{otherwise}\end{array}\right.

from the definition of the weak continuity equation (12). It follows that the supremum on the right hand side of (15) can be restricted to measures ρ\rho and mm satisfying the continuity equation. Therefore, by Property 5.2, we obtain that

supψ1{[0,1]×𝕊1cos(θ)ψ1(t)𝑑ρ}\displaystyle\underset{\psi_{1}}{\sup}\left\{-\int_{[0,1]\times\mathbb{S}^{1}}\cos(\theta)\psi_{1}(t)d\rho\right\} =supψ1{01(𝕊1cos(θ)𝑑ρt(θ))ψ1(t)𝑑t}\displaystyle=\underset{\psi_{1}}{\sup}\left\{-\int_{0}^{1}\left(\int_{\mathbb{S}^{1}}\cos(\theta)d\rho_{t}(\theta)\right)\psi_{1}(t)dt\right\}
={0 if for a.e t[0,1],𝕊1cos(θ)𝑑ρt=0+ otherwise\displaystyle=\left\{\begin{array}[]{ll}0&\mbox{ if for a.e }t\in[0,1],\ \int_{\mathbb{S}^{1}}\cos(\theta)d\rho_{t}=0\\ +\infty&\mbox{ otherwise}\end{array}\right.

and similarly

supψ2{[0,1]×𝕊1sin(θ)ψ2(t)𝑑ρ}={0 if for a.e t[0,1],𝕊1sin(θ)𝑑ρt=0+ otherwise\displaystyle\underset{\psi_{2}}{\sup}\left\{-\int_{[0,1]\times\mathbb{S}^{1}}\sin(\theta)\psi_{2}(t)d\rho\right\}=\left\{\begin{array}[]{ll}0&\mbox{ if for a.e }t\in[0,1],\ \int_{\mathbb{S}^{1}}\sin(\theta)d\rho_{t}=0\\ +\infty&\mbox{ otherwise}\end{array}\right.

from which we deduce that

(16) sup𝜈{F(ν)G(A(ν))}=inf𝜈{F(ν)|subj. to {tρ+θm=0ρ0=μ0,ρ1=μ1𝕊1eiθ𝑑ρt(θ)=0,for a.e t[0,1]}\underset{\nu}{\sup}\left\{-F^{*}(\nu)-G^{*}(-A^{*}(\nu))\right\}=-\underset{\nu}{\inf}\left\{F^{*}(\nu)\ |\ \text{subj. to }\left\{\begin{array}[]{lll}\partial_{t}\rho+\partial_{\theta}m=0\\ \rho_{0}=\mu_{0},\ \rho_{1}=\mu_{1}\\ \int_{\mathbb{S}^{1}}e^{i\theta}d\rho_{t}(\theta)=0,\ \text{for a.e }t\in[0,1]\end{array}\right.\right\}

Finally the convex conjugate of FF is computed for instance in [15] based on Theorem 5 of [40] and can be shown to be:

F(ν)=[0,1]×𝕊1f(dνdλ)𝑑λ.F^{*}(\nu)=\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\nu}{d\lambda}\right)d\lambda.

This allows to conclude that (16) corresponds to W¯2(μ0,μ1)-\overline{W}_{2}(\mu_{0},\mu_{1}) and, since we assume that W¯2(μ0,μ1)<+\overline{W}_{2}(\mu_{0},\mu_{1})<+\infty, we obtain by Fenchel-Rockafellar theorem the existence of a minimizer (ρ,m)(\rho,m) to (14).

In addition, we have by construction that W¯2(μ0,μ1)W2(μ0,μ1)\overline{W}_{2}(\mu_{0},\mu_{1})\geq W_{2}(\mu_{0},\mu_{1}) and therefore W¯2(μ0,μ1)=0\overline{W}_{2}(\mu_{0},\mu_{1})=0 implies that μ0=μ1\mu_{0}=\mu_{1}. The symmetry of W¯2\overline{W}_{2} is also immediate from the fact that for any measure path ν=(ρ,m)\nu=(\rho,m) satisfying the conditions in (14), one can consider ν~=(ρ~t=ρ1t,m~t=m1t)\tilde{\nu}=(\tilde{\rho}_{t}=\rho_{1-t},\tilde{m}_{t}=m_{1-t}) the time-reversed paths which satisfy ρ~0=μ1\tilde{\rho}_{0}=\mu_{1}, ρ~1=μ0\tilde{\rho}_{1}=\mu_{0}, tρ~+θm~=0\partial_{t}\tilde{\rho}+\partial_{\theta}\tilde{m}=0 and 𝕊1eiθ𝑑ρ~t(θ)=0\int_{\mathbb{S}^{1}}e^{i\theta}d\tilde{\rho}_{t}(\theta)=0 while:

[0,1]×𝕊1f(dν~dλ)𝑑λ=[0,1]×𝕊1f(dνdλ)𝑑λ.\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\tilde{\nu}}{d\lambda}\right)d\lambda=\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\nu}{d\lambda}\right)d\lambda.

Finally, the triangular inequality can be shown in similar way as for the standard Wasserstein metric i.e. by concatenation. Let μ0,μ1,μ2𝒫(𝕊1)0+\mu_{0},\mu_{1},\mu_{2}\in\mathcal{P}(\mathbb{S}^{1})\cap\mathcal{M}^{+}_{0} and assume that W¯2(μ0,μ1)<+\overline{W}_{2}(\mu_{0},\mu_{1})<+\infty and W¯2(μ1,μ2)<+\overline{W}_{2}(\mu_{1},\mu_{2})<+\infty (otherwise the triangular inequality is trivially satisfied). From the above, we know that there exist ν=(ρ,m)\nu=(\rho,m) and ν=(ρ,m)\nu^{\prime}=(\rho^{\prime},m^{\prime}) that achieve the minimum in the distances W¯2(μ0,μ1)\overline{W}_{2}(\mu_{0},\mu_{1}) and W¯2(μ1,μ2)\overline{W}_{2}(\mu_{1},\mu_{2}) respectively. Letting τ0:(t,θ)(t/2,θ)\tau^{0}:(t,\theta)\mapsto(t/2,\theta) and τ1:(t,θ)((t+1)/2,θ)\tau^{1}:(t,\theta)\mapsto((t+1)/2,\theta), we define the following concatenated path ν¯=(ρ¯,m¯)\bar{\nu}=(\bar{\rho},\bar{m}):

ν¯={((τ0)ρ,2(τ0)m)for 0t<12((τ1)ρ,2(τ1)m)for 12t1\bar{\nu}=\left\{\begin{aligned} &((\tau^{0})_{\sharp}\rho,2(\tau^{0})_{\sharp}m)\ \ \text{for }0\leq t<\frac{1}{2}\\ &((\tau^{1})_{\sharp}\rho^{\prime},2(\tau^{1})_{\sharp}m^{\prime})\ \ \text{for }\frac{1}{2}\leq t\leq 1\end{aligned}\right.

for which one can easily verify that ρ¯0=μ0\bar{\rho}_{0}=\mu_{0}, ρ¯1=μ2\bar{\rho}_{1}=\mu_{2}, tρ¯+θm¯=0\partial_{t}\bar{\rho}+\partial_{\theta}\bar{m}=0 and 𝕊1eiθ𝑑ρ¯t(θ)=0\int_{\mathbb{S}^{1}}e^{i\theta}d\bar{\rho}_{t}(\theta)=0 for almost all t[0,1]t\in[0,1]. It follows that:

W¯2(μ0,μ2)[0,1]×𝕊1f(dν¯dλ)𝑑λ\displaystyle\overline{W}_{2}(\mu_{0},\mu_{2})\leq\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\bar{\nu}}{d\lambda}\right)d\lambda =[0,1]×𝕊1f(dνdλ)𝑑λ+[0,1]×𝕊1f(dνdλ)𝑑λ\displaystyle=\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\nu}{d\lambda}\right)d\lambda+\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\nu^{\prime}}{d\lambda}\right)d\lambda
=W¯2(μ0,μ1)+W¯2(μ1,μ2).\displaystyle=\overline{W}_{2}(\mu_{0},\mu_{1})+\overline{W}_{2}(\mu_{1},\mu_{2}).

In general, one cannot guarantee that the distance W¯2\overline{W}_{2} is finite between any two measures. Nevertheless, it holds in particular when:

Proposition 5.5.

Let μ0=ρ0(θ)dθ\mu_{0}=\rho_{0}(\theta)d\theta and μ1=ρ1(θ)dθ\mu_{1}=\rho_{1}(\theta)d\theta be two measures in 𝒫(𝕊1)0+\mathcal{P}(\mathbb{S}^{1})\cap\mathcal{M}^{+}_{0} with densities with respect to the Lebesgue measure on 𝕊1\mathbb{S}^{1} with ρ0,ρ1L1(𝕊1)\rho_{0},\rho_{1}\in L^{1}(\mathbb{S}^{1}), and assume that there exists c>0c>0 such that ρ0(θ)c\rho_{0}(\theta)\geq c, ρ1(θ)c\rho_{1}(\theta)\geq c for almost all θ𝕊1\theta\in\mathbb{S}^{1}. Then W¯2(μ0,μ1)<+\overline{W}_{2}(\mu_{0},\mu_{1})<+\infty.

Proof.

With the above assumptions, let us consider the linear interpolation path ρt(θ)=(1t)ρ0(θ)+tρ1(θ)\rho_{t}(\theta)=(1-t)\rho_{0}(\theta)+t\rho_{1}(\theta) and define mt=H(θ)dθm_{t}=H(\theta)d\theta where H(θ)=0θ(ρ1(α)ρ0(α))𝑑αH(\theta)=\int_{0}^{\theta}(\rho_{1}(\alpha)-\rho_{0}(\alpha))d\alpha. Then one easily checks that we have tρ+θm=0\partial_{t}\rho+\partial_{\theta}m=0 in the sense of distributions. Furthermore,

𝕊1eiθρt(θ)𝑑θ=(1t)𝕊1eiθρ0(θ)𝑑θ+t𝕊1eiθρ1(θ)𝑑θ=0.\int_{\mathbb{S}^{1}}e^{i\theta}\rho_{t}(\theta)d\theta=(1-t)\int_{\mathbb{S}^{1}}e^{i\theta}\rho_{0}(\theta)d\theta+t\int_{\mathbb{S}^{1}}e^{i\theta}\rho_{1}(\theta)d\theta=0.

Then, taking λ\lambda the Lebesgue measure on [0,1]×𝕊1[0,1]\times\mathbb{S}^{1} and since ρt(θ)c\rho_{t}(\theta)\geq c for almost all θ\theta, we find:

01𝕊1f(ρt(θ),mt(θ))𝑑θ𝑑t=01𝕊1H(θ)22ρt(θ)𝑑θ𝑑t12c𝕊1H(θ)2𝑑θ<+\int_{0}^{1}\int_{\mathbb{S}^{1}}f\left(\rho_{t}(\theta),m_{t}(\theta)\right)d\theta dt=\int_{0}^{1}\int_{\mathbb{S}^{1}}\frac{H(\theta)^{2}}{2\rho_{t}(\theta)}d\theta dt\leq\frac{1}{2c}\int_{\mathbb{S}^{1}}H(\theta)^{2}d\theta<+\infty

and thus W¯2(μ0,μ1)\overline{W}_{2}(\mu_{0},\mu_{1}) is also finite. ∎

Note that as a special case of this proposition, we deduce that the distance is finite between any two continuous strictly positive densities on 𝕊1\mathbb{S}^{1}.

Now, thanks to the bijection induced by MM between the set of convex curves of length one modulo translation which we will denote C~conv1\tilde{C}_{conv}^{1} and the set of probability measures of 0+\mathcal{M}^{+}_{0}, we can equip C~conv1\tilde{C}_{conv}^{1} with the structure of a geodesic space by setting:

(17) d(c0,c1)W¯2(M(c0),M(c1))d(c_{0},c_{1})\doteq\overline{W}_{2}(M(c_{0}),M(c_{1}))

for any c0,c1C~conv1c_{0},c_{1}\in\tilde{C}_{conv}^{1}. A geodesic between the two curves is then given by c(t)=M1(ρ(t))c(t)=M^{-1}(\rho(t)) where ρ(t)\rho(t) is a geodesic between the measures M(c0)M(c_{0}) and M(c1)M(c_{1}) for the metric W¯2\overline{W}_{2}.

Remark 5.6.

The above distance on C~conv1\tilde{C}_{conv}^{1} is in addition equivariant to the action of rotations of the plane in the sense that for any RSO(2)R\in SO(2), we have d(Rc0,Rc1)=d(c0,c1)d(R\cdot c_{0},R\cdot c_{1})=d(c_{0},c_{1}). This follows from the equivariance of the constrained Wasserstein distance W¯2\overline{W}_{2} with respect to rotations. Indeed, we see that for any probability measures μ0\mu_{0} and μ1\mu_{1} in 0+\mathcal{M}^{+}_{0} and any RSO(2)R\in SO(2), we have RM(c0)=M(Rc0)R_{\ast}M(c_{0})=M(R\cdot c_{0}) and RM(c1)=M(Rc1)R_{\ast}M(c_{1})=M(R\cdot c_{1}). Furthermore, if ν=(ρ,m)\nu=(\rho,m) is a pair of measures satisfying the continuity equation together with the boundary conditions ρ0=μ0\rho_{0}=\mu_{0}, ρ1=μ1\rho_{1}=\mu_{1} and the closure condition 𝕊1eiθ𝑑ρt(θ)=0\int_{\mathbb{S}^{1}}e^{i\theta}d\rho_{t}(\theta)=0 for almost all t[0,1]t\in[0,1] then one easily verifies that ν~=(ρ~,m~)\tilde{\nu}=(\tilde{\rho},\tilde{m}) defined by ρ~=(Rρt)dt\tilde{\rho}=(R_{\ast}\rho_{t})\otimes dt and m~=(Rmt)dt\tilde{m}=(R_{\ast}m_{t})\otimes dt still satisfy the continuity equation, the closure condition and that:

[0,1]×𝕊1f(dν~dλ)𝑑λ=[0,1]×𝕊1f(dνdλ)𝑑λ.\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\tilde{\nu}}{d\lambda}\right)d\lambda=\int_{[0,1]\times\mathbb{S}^{1}}f\left(\frac{d\nu}{d\lambda}\right)d\lambda.

Then taking the infimum leads to W¯2(M(Rc0),M(Rc1))=W¯2(M(c0),M(c1))\overline{W}_{2}(M(R\cdot c_{0}),M(R\cdot c_{1}))=\overline{W}_{2}(M(c_{0}),M(c_{1})). This equivariance property allows to further quotient out rotations and obtain a distance between unit length convex curves modulo rigid motions that writes:

d(c0,c1)minRSO(2)d(c0,Rc1).d^{\prime}(c_{0},c_{1})\doteq\min_{R\in SO(2)}d(c_{0},R\cdot c_{1}).

5.3.2. Numerical approach

We now derive a numerical method to estimate the constrained Wasserstein distance W¯2\overline{W}_{2} between two discretized densities of 𝕊1\mathbb{S}^{1} and thereby the distance dd between convex curves given by (17). Our proposed approach mainly follows and adapts the first-order proximal methods introduced in [37] for the computation of Wasserstein metrics, specifically the primal-dual algorithm which itself is a special instantiation of [13]. We briefly present the main elements of the algorithm in the following paragraphs.

We shall consider measures on [0,1]×𝕊1[0,1]\times\mathbb{S}^{1} discretized over two types of regular time/angle grids: centered and a staggered grids. We define the centered grid 𝒢c\mathcal{G}_{c} with P+1P+1 time samples and NN angular samples as:

𝒢c={(ti=iP,θj=2πjN)| 0iP, 0jN1}.\mathcal{G}_{c}=\left\{\left(t_{i}=\frac{i}{P},\theta_{j}=2\pi\frac{j}{N}\right)\ |\ 0\leq i\leq P,\ 0\leq j\leq N-1\right\}.

Discrete measure pairs sampled on the centered grid will be then represented as an element VV of the space c𝒢c×𝒢c\mathcal{E}_{c}\doteq\mathbb{R}^{\mathcal{G}_{c}}\times\mathbb{R}^{\mathcal{G}_{c}}. The staggered grids consist of time and angular samples shifted to the middle of the samples of 𝒢c\mathcal{G}_{c}. Specifically, we introduce a time and a space staggered grid defined as follows:

𝒢st={(ti=i+1/2P,θj=2πjN)|1iP, 0jN1}\displaystyle\mathcal{G}_{s}^{t}=\left\{\left(t_{i}=\frac{i+1/2}{P},\theta_{j}=2\pi\frac{j}{N}\right)\ |\ -1\leq i\leq P,\ 0\leq j\leq N-1\right\}
𝒢sθ={(ti=iP,θj=2πj+1/2N)| 0iP,1jN2}.\displaystyle\mathcal{G}_{s}^{\theta}=\left\{\left(t_{i}=\frac{i}{P},\theta_{j}=2\pi\frac{j+1/2}{N}\right)\ |\ 0\leq i\leq P,\ -1\leq j\leq N-2\right\}.

We shall then denote U=(ρ,m)U=(\rho,m) an element of s𝒢st×𝒢sθ\mathcal{E}_{s}\doteq\mathbb{R}^{\mathcal{G}_{s}^{t}}\times\mathbb{R}^{\mathcal{G}_{s}^{\theta}} which represents a pair of discretized measures of [0,1]×𝕊1[0,1]\times\mathbb{S}^{1} on the staggered grids and we introduce the interpolation operator :sc\mathcal{I}:\mathcal{E}_{s}\rightarrow\mathcal{E}_{c} defined by (U)=(ρ¯,m¯)\mathcal{I}(U)=(\bar{\rho},\bar{m}) with

ρ¯i,j=ρi1,j+ρi,j2,for 0iP, 0jN1.\displaystyle\bar{\rho}_{i,j}=\frac{\rho_{i-1,j}+\rho_{i,j}}{2},\ \text{for }0\leq i\leq P,\ 0\leq j\leq N-1.
m¯i,j={mi,j1+mi,j2,for 0iP, 0jN2mi,N2+mi,12,for 0iP,j=N1\displaystyle\bar{m}_{i,j}=\left\{\begin{aligned} &\frac{m_{i,j-1}+m_{i,j}}{2},\ \text{for }0\leq i\leq P,\ 0\leq j\leq N-2\\ &\frac{m_{i,N-2}+m_{i,-1}}{2},\ \text{for }0\leq i\leq P,\ j=N-1\end{aligned}\right.

Note that the slight difference between the time and angular components and with the approach of [37] is due to the periodic boundary condition on the samples θj𝕊1\theta_{j}\in\mathbb{S}^{1}. We also define a discrete divergence operator div:s𝒢c\text{div}:\mathcal{E}_{s}\rightarrow\mathbb{R}^{\mathcal{G}_{c}} as div(U)i,j=P(ρi,jρi1,j)+N(mi,jmi,j1)\text{div}(U)_{i,j}=P(\rho_{i,j}-\rho_{i-1,j})+N(m_{i,j}-m_{i,j-1}) for all 0iP, 0jN10\leq i\leq P,\ 0\leq j\leq N-1. This allows to write a discrete version of (14) as a convex minimization problem:

(18) minU=(ρ,m)s{i=0Pj=0N1f((U)i,j)+ι𝒞(ρ0,ρ1)(U)}.\min_{U=(\rho,m)\in\mathcal{E}_{s}}\left\{\sum_{i=0}^{P}\sum_{j=0}^{N-1}f\left(\mathcal{I}(U)_{i,j}\right)+\iota_{\mathcal{C}(\rho_{0},\rho_{1})}(U)\right\}.

where

𝒞(ρ0,ρ1){Uss.t {ρ0,j=(ρ0)j,ρP,j=(ρ1)j,for all 1jN2div(U)i,j=0,for all 0iP, 0jN1j=0N1cos(θj)ρi,j=0,j=0N1sin(θj)ρi,j=0for all 1iP.}\mathcal{C}(\rho_{0},\rho_{1})\doteq\left\{U\in\mathcal{E}_{s}\ \text{s.t }\left\{\begin{aligned} &\rho_{0,j}=(\rho_{0})_{j},\ \rho_{P,j}=(\rho_{1})_{j},\ \text{for all }-1\leq j\leq N-2\\ &\text{div}(U)_{i,j}=0,\ \text{for all }0\leq i\leq P,\ 0\leq j\leq N-1\\ &\sum_{j=0}^{N-1}\cos(\theta_{j})\rho_{i,j}=0,\ \sum_{j=0}^{N-1}\sin(\theta_{j})\rho_{i,j}=0\ \text{for all }-1\leq i\leq P.\end{aligned}\right.\right\}

and ι𝒞(ρ0,ρ1)(U)=0\iota_{\mathcal{C}(\rho_{0},\rho_{1})}(U)=0 if U𝒞(ρ0,ρ1)U\in\mathcal{C}(\rho_{0},\rho_{1}), ι𝒞(ρ0,ρ1)(U)=+\iota_{\mathcal{C}(\rho_{0},\rho_{1})}(U)=+\infty if U𝒞(ρ0,ρ1)U\notin\mathcal{C}(\rho_{0},\rho_{1}).

To solve the convex problem (18), we will use a primal-dual method adapted from the work of [13]. We first recall the definition of the proximal operator of a convex function h:kh:\mathbb{R}^{k}\rightarrow\mathbb{R}:

Proxh(x)=argminyk12xy2+h(y)\text{Prox}_{h}(x)=\underset{y\in\mathbb{R}^{k}}{\text{argmin}}\ \frac{1}{2}\|x-y\|^{2}+h(y)

as well as the convex dual of hh:

h(w)=maxyky,wh(y).h^{*}(w)=\underset{y\in\mathbb{R}^{k}}{\max}\ \langle y,w\rangle-h(y).

The function to minimize in (18) is of the form F((U))+G(U)F(\mathcal{I}(U))+G(U). The primal-dual algorithm we implement consists of the following iterative updates on the sequences (U(n),V(n),Γ(n))s×c×s(U^{(n)},V^{(n)},\Gamma^{(n)})\in\mathcal{E}_{s}\times\mathcal{E}_{c}\times\mathcal{E}_{s} starting from an initialization U(0)sU^{(0)}\in\mathcal{E}_{s}, V(0)cV^{(0)}\in\mathcal{E}_{c} and Γ(0)=U(0)\Gamma^{(0)}=U^{(0)}:

(19) V(n+1)=ProxσF(V(n)+σΓ(n))\displaystyle V^{(n+1)}=\text{Prox}_{\sigma F^{*}}(V^{(n)}+\sigma\mathcal{I}\Gamma^{(n)})
U(n+1)=ProxτG(U(n)τV(n+1))\displaystyle U^{(n+1)}=\text{Prox}_{\tau G}(U^{(n)}-\tau\mathcal{I}^{*}V^{(n+1)})
Γ(n+1)=U(n+1)+θ(U(n+1)U(n))\displaystyle\Gamma^{(n+1)}=U^{(n+1)}+\theta(U^{(n+1)}-U^{(n)})

where σ\sigma, τ\tau are positive step sizes and 0θ10\leq\theta\leq 1 an inertia parameter. It follows from the general result of [13] that the algorithm converges to a solution of (18) if τσ2<1\tau\sigma\|\mathcal{I}\|^{2}<1 where \|\mathcal{I}\| denotes the operator norm of \mathcal{I}. We typically have 1\|\mathcal{I}\|\approx 1 as \mathcal{I} is here an interpolation operator.

It only remains to specify the exact expressions of the proximal operators appearing in the first two equations of (19). First, by Moreau’s identity, one has that for all VcV\in\mathcal{E}_{c}:

ProxσF(V)=VσProxF/σ(V/σ).\text{Prox}_{\sigma F^{*}}(V)=V-\sigma\text{Prox}_{F/\sigma}(V/\sigma).

Now the proximal of FF is computed in [37] Proposition 1 from which we obtain for all V~c\tilde{V}\in\mathcal{E}_{c}:

ProxF/σ(V~)=(Proxf/σ(V~i,j))0iP0jN1\text{Prox}_{F/\sigma}(\tilde{V})=\left(\text{Prox}_{f/\sigma}(\tilde{V}_{i,j})\right)_{\begin{subarray}{c}0\leq i\leq P\phantom{-1}\\ 0\leq j\leq N-1\end{subarray}}

where

Proxf/σ(d,m)={(σdr(d,m)σd+1,r(d,m))if r(d,m)>0(0,0)otherwise\text{Prox}_{f/\sigma}(d,m)=\left\{\begin{aligned} &\left(\frac{\sigma dr^{\star}(d,m)}{\sigma d+1},r^{\star}(d,m)\right)\ \ &\text{if }r^{\star}(d,m)>0\\ &\left(0,0\right)&\text{otherwise}\end{aligned}\right.

with r(d,m)r^{\star}(d,m) being the largest real root of the polynomial equation (xd)(x+1/σ)212σm2=0(x-d)(x+1/\sigma)^{2}-\frac{1}{2\sigma}m^{2}=0.

To express the second update in (19), we first point out that :cs\mathcal{I}^{*}:\mathcal{E}_{c}\rightarrow\mathcal{E}_{s} is given by (ρ,m)=(ρ¯,m¯)(\rho,m)=\mathcal{I}^{*}(\bar{\rho},\bar{m}) with:

ρi,j={ρ¯i+1,j2,for i=1, 0jN2ρ¯i,j+ρ¯i+1,j2,for 0iP1, 0jN1ρ¯i,j2,for i=P, 0jN2\displaystyle\rho_{i,j}=\left\{\begin{aligned} &\frac{\bar{\rho}_{i+1,j}}{2},\ \text{for }i=-1,\ 0\leq j\leq N-2\\ &\frac{\bar{\rho}_{i,j}+\bar{\rho}_{i+1,j}}{2},\ \text{for }0\leq i\leq P-1,\ 0\leq j\leq N-1\\ &\frac{\bar{\rho}_{i,j}}{2},\ \text{for }i=P,\ 0\leq j\leq N-2\end{aligned}\right.
mi,j={m¯i,0+m¯i,N12,for 0iP,j=1m¯i,j+m¯i,j+12,for 0iP, 0jN2\displaystyle m_{i,j}=\left\{\begin{aligned} &\frac{\bar{m}_{i,0}+\bar{m}_{i,N-1}}{2},\ \text{for }0\leq i\leq P,\ j=-1\\ &\frac{\bar{m}_{i,j}+\bar{m}_{i,j+1}}{2},\ \text{for }0\leq i\leq P,\ 0\leq j\leq N-2\end{aligned}\right.

Moreover since GG is here the convex indicator function of the constraint set 𝒞(ρ0,ρ1)\mathcal{C}(\rho_{0},\rho_{1}), we have τG=G\tau G=G for any τ>0\tau>0 and ProxτG=ProxG=Proj𝒞(ρ0,ρ1)\text{Prox}_{\tau G}=\text{Prox}_{G}=\text{Proj}_{\mathcal{C}(\rho_{0},\rho_{1})} the orthogonal projector on the convex set 𝒞(ρ0,ρ1)\mathcal{C}(\rho_{0},\rho_{1}). In addition to the usual boundary and divergence constraints, we also have to deal with the two extra closure conditions. We introduce the operator Ξ:s𝒢c×(N×N)×(P+2×P+2)\Xi:\mathcal{E}_{s}\rightarrow\mathbb{R}^{\mathcal{G}_{c}}\times(\mathbb{R}^{N}\times\mathbb{R}^{N})\times(\mathbb{R}^{P+2}\times\mathbb{R}^{P+2}) defined for any U=(ρ,m)sU=(\rho,m)\in\mathcal{E}_{s} by:

Ξ(U)=(div(U),(ρ1,j,ρP,j)0jN1,(j=0N1cos(θj)ρi,j,j=0N1sin(θj)ρi,j)1iP)\Xi(U)=\left(\text{div}(U),(\rho_{-1,j},\rho_{P,j})_{0\leq j\leq N-1},\left(\sum_{j=0}^{N-1}\cos(\theta_{j})\rho_{i,j},\sum_{j=0}^{N-1}\sin(\theta_{j})\rho_{i,j}\right)_{-1\leq i\leq P}\right)

which allows us to write 𝒞(ρ0,ρ1)={Us|Ξ(U)=b(0,(ρ0,ρ1),(0,0))}\mathcal{C}(\rho_{0},\rho_{1})=\big{\{}U\in\mathcal{E}_{s}\ |\ \Xi(U)=b\doteq\left(0,(\rho_{0},\rho_{1}),(0,0)\right)\big{\}}. Thus, we can express the orthogonal projection on 𝒞(ρ0,ρ1)\mathcal{C}(\rho_{0},\rho_{1}) as:

Proj𝒞(ρ0,ρ1)(U)=UΞΔ1Ξ(U)+ΞΔ1b\text{Proj}_{\mathcal{C}(\rho_{0},\rho_{1})}(U)=U-\Xi^{*}\Delta^{-1}\Xi(U)+\Xi^{*}\Delta^{-1}b

where Δ=ΞΞ\Delta=\Xi\Xi^{*} can be interpreted as a modified Laplace operator on the spatial-angular domain, which in practice we can precompute together with its inverse before iterating (19) or alternatively implement in the Fourier domain.

5.3.3. Numerical results

Refer to caption Refer to caption Refer to caption Refer to caption
t=0t=0 t=1/3t=1/3 t=2/3t=2/3 t=1t=1
Refer to caption Refer to caption Refer to caption Refer to caption
Figure 9. Geodesic for the constrained Wasserstein metric computed with the proposed primal-dual algorithm. On top, the intermediate convex curves c(t)C~convc(t)\in\tilde{C}_{conv} and in the bottom row are shown the associated measures μ(t)\mu(t).

We present a few numerical results obtained with the above approach which we implemented in Python. In all experiments, we take N=150N=150 angular samples on 𝕊1=[0,2π)\mathbb{S}^{1}=[0,2\pi) and P=50P=50 time samples in [0,1][0,1]. As for the step parameters of the primal-dual algorithm, we picked σ=0.5\sigma=0.5, τ=0.25\tau=0.25 and θ=0.6\theta=0.6. We run the algorithm (19) for 1000 iterations although we observe in practice that convergence to a minimum is typically reached after around 250 iterations (cf the energy plot in Figure 10).

First, in Figure 9, is shown the geodesic between the same polygons as in Figures 7 and 8 for the new constrained Wasserstein metric. Observe that unlike kernel metrics, the geodesic involves mass transportation rather than pure interpolation between the measures but that unlike the standard Wasserstein distance, the closure constraint is satisfied along the path and the corresponding shapes all stay within the set of closed convex curves. We further illustrate those effects with another example of a geodesic between two density measures and the corresponding geodesic between the convex curves in Figure 10.

Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption
t=0t=0 t=1/3t=1/3 t=2/3t=2/3 t=1t=1
Refer to caption Refer to caption Refer to caption Refer to caption
Figure 10. Geodesic for the constrained Wasserstein metric between two density measures. The evolution of the energy throughout the iterations of the proposed primal-dual algorithm is plotted on the top row to illustrate the convergence. On the middle row, are shown the intermediate convex curves c(t)C~convc(t)\in\tilde{C}_{conv} along the estimated geodesic. On the bottom row we display the final density ρ1\rho_{1} in red and the density ρ(t)\rho(t) along the geodesic in blue.

6. Some open problems and future directions

As a conclusion to this paper, we discuss a few possible topics and tracks for future investigation that arise as natural follow-ups to this work and the results we presented.

Unbalanced optimal transport for convex shape analysis

In section 5.3, we introduced a distance based on a constrained version of the Wasserstein metric between length measures which turns the set of convex closed curves of length one into a length space. There are yet several questions which are left partly unaddressed when it comes to this distance and its properties. In particular, it remains to be studied whether W¯2\overline{W}_{2} is finite between any pair of probability measures of 0+\mathcal{M}^{+}_{0} or at least if the result of Proposition 5.5 could be extended to other families of measures. Besides, W¯2\overline{W}_{2} is only defined for probability measures which implies that the resulting distance dd in (17) requires the convex curves to be normalized to have unit length. This makes such a distance unadapted to retrieve and quantify e.g. global or local changes in scale.

We believe that both of the previous issues could be addressed by considering an unbalanced extension of the metric W¯2\overline{W}_{2} to all measures of 0+\mathcal{M}^{+}_{0}, along similar lines as the unbalanced frameworks for standard optimal transport proposed in works such as [38, 36, 15]. In the context of this paper, this could be done by adding a source term to the continuity equation, namely consider tρ+θm=ζ\partial_{t}\rho+\partial_{\theta}m=\zeta with ζ\zeta being a signed measure of [0,1]×𝕊1[0,1]\times\mathbb{S}^{1} that models an additional transformation of the measure ρ\rho in conjunction with mass transportation. Then, as in [36, 15], one could penalize this extra source term based on the Fischer-Rao metric of ζ\zeta, which would lead to define the following distance between any μ0,μ10+\mu_{0},\mu_{1}\in\mathcal{M}^{+}_{0}:

W¯FR(μ0,μ1)=infν=(ρ,m,ζ){[0,1]×𝕊1g(dνdλ)dλ|subj. to {tρ+θm=ζρ0=μ0,ρ1=μ1𝕊1eiθ𝑑ρt(θ)=0,for a.e t[0,1]}.\overline{W}_{FR}(\mu_{0},\mu_{1})=\inf_{\nu=(\rho,m,\zeta)}\left\{\int_{[0,1]\times\mathbb{S}^{1}}g\left(\frac{d\nu}{d\lambda}\right)\ d\lambda\ |\ \text{subj. to }\left\{\begin{array}[]{lll}\partial_{t}\rho+\partial_{\theta}m=\zeta\\ \rho_{0}=\mu_{0},\ \rho_{1}=\mu_{1}\\ \int_{\mathbb{S}^{1}}e^{i\theta}d\rho_{t}(\theta)=0,\ \text{for a.e }t\in[0,1]\end{array}\right.\right\}.

where λ\lambda is such that ρ,m,ζλ\rho,m,\zeta\ll\lambda and g:××+g:\mathbb{R}\times\mathbb{R}\times\mathbb{R}\rightarrow\mathbb{R}_{+} is given by:

g(d,m,ζ)={|m|2+δ2ζ22dif d>00if (d,m,ζ)=(0,0,0)+otherwiseg(d,m,\zeta)=\left\{\begin{aligned} &\frac{|m|^{2}+\delta^{2}\zeta^{2}}{2d}\ \ \text{if }d>0\\ &0\ \ \text{if }(d,m,\zeta)=(0,0,0)\\ &+\infty\ \ \text{otherwise}\end{aligned}\right.

with δ>0\delta>0 a weighting parameter between the two components of the cost. In this case, it becomes easy to show that W¯FR(μ0,μ1)\overline{W}_{FR}(\mu_{0},\mu_{1}) is finite for any measures μ0,μ10+\mu_{0},\mu_{1}\in\mathcal{M}^{+}_{0} by simply considering the special path ρt=(1t)ρ0+tρ1\rho_{t}=(1-t)\rho_{0}+t\rho_{1}, mt=0m_{t}=0 and ζt=ρ1ρ0\zeta_{t}=\rho_{1}-\rho_{0} for all t[0,1]t\in[0,1]. Although we leave it for future work, we expect the existence of geodesics and distance properties to hold for W¯FR\overline{W}_{FR} by extending the proof of Theorem 5.4. Moreover, the precise analysis of its topological properties and how the resulting metric between convex curves compare to other geometric distances remain to be examined.

Generalization to higher dimension

Although this work focused on the case of planar curves, the concept of length measure extends to closed hypersurfaces in any dimension which is known as area measures, as mentioned in the introduction. Area measures, particularly of surfaces in 3\mathbb{R}^{3}, is a central concept in the Brunn-Minkowski theory of mixed volumes [42] but also appear in some applications such as object recognition [29] or surface reconstruction from computerized tomography [39]. The area measure of a (n1)(n-1)-dimensional closed oriented submanifold (more generally rectifiable subset) SS of n\mathbb{R}^{n} can be defined, similarly to Definition 2.1, as the positive Radon measure on 𝕊n1\mathbb{S}^{n-1} given by σS(B)=Voln1({xS|nS(x)B})\sigma_{S}(B)=\text{Vol}^{n-1}(\{x\in S\ |\ \vec{n}_{S}(x)\in B\}) for all Borel subset B𝕊n1B\subset\mathbb{S}^{n-1} where nS(x)𝕊n1\vec{n}_{S}(x)\in\mathbb{S}^{n-1} denotes the unit normal vector to SS at xx and Voln1\text{Vol}^{n-1} is the (n1)(n-1) volume measure i.e. the Hausdorff measure of dimension (n1)(n-1). Importantly, the Minkowski-Fenchel-Jessen theorem still holds for general area measures, namely the area measure again characterizes a convex set up to translation. However, there are two significant differences when n3n\geq 3 compared to the situation of planar curves. First the Minkowski sum of two convex does not generally correspond to the sum of their area measures. Second, reconstructing the convex shape from its area measure, even for discrete measures and polyhedra, is no longer straightforward: it is in fact an active research topic and several different approaches and algorithms have been proposed, see e.g. [47, 33, 25].

In connection to the results presented in this paper, we expect for instance that an isoperimetric characterization of convex sets similar to Theorem 4.4 should hold for the positive volume enclosed by (n1)(n-1)-dimensional non self-intersecting rectifiable sets that share the same area measure under the adequate regularity assumptions. Such a property has been shown in particular for polyhedra in [8]. As for the construction of metrics and geodesics between convex shapes, the mathematical construction of the constrained Wasserstein metric presented in Section 5.3.1 can be a priori adapted to area measures in any dimension, by replacing the closure constraint to 𝕊n1x𝑑ρt(x)=0\int_{\mathbb{S}^{n-1}}xd\rho_{t}(x)=0 for almost all t[0,1]t\in[0,1]. However, the numerical implementation of such metrics along the lines of the approach of Section 5.3.2 would induce additional difficulties. Indeed, the computation of the operators on grids over higher-dimensional sphere becomes more involved and numerically intensive. Moreover, recovering the convex curves associated to a geodesic in the space of area measures would further require, as explained above, applying some reconstruction algorithm a posteriori. Finally, an important issue for future investigation is to derive an efficient implementation of the variation of the distance with respect to the discrete distributions, which could be then used for instance to estimate Kärcher means in the space of convex sets.

Acknowledgements

The authors would like to thank F-X Vialard for some useful insights in relation to the optimal transport model presented in this paper. This work was supported by the National Science Foundation (NSF) under the grant 1945224.

References

  • [1] H. Abdallah and Q. Mérigot, On the reconstruction of convex sets from random normal measurements, Proceedings of the thirtieth annual symposium on Computational geometry, 2014, pp. 300–307.
  • [2] A. Alexandrov, Zur theorie der gemischten volumina von konvexen koörpern. i. verallgemeinerung einiger begriffe der theorie von konvexen körpern, Matematicheskii Sbornik 44 (1937), no. 5, 947–972.
  • [3] W.K. Allard, On the first variation of a varifold, Annals of mathematics (1972), 417–491.
  • [4] L. Ambrosio, N. Fusco, and D. Pallara, Functions of bounded variation and free discontinuity problems, Oxford : Clarendon Press, 2000.
  • [5] N. Aronszajn, Theory of reproducing kernels, Trans. Amer. Math. Soc. 68 (1950), 337–404.
  • [6] J-D. Benamou and Y. Brenier, A computational fluid mechanics solution to the Monge-Kantorovich mass transfer problem, Numerische Mathematik 84 (2000), no. 3, 375–393.
  • [7] W. Blaschke, Kreis und kugel, Verlag von Veit &\& Comp, 1916.
  • [8] K. Böröczky, I. Bárány, E. Makai, and J. Pach, Maximal volume enclosed by plates and proof of the chessboard conjecture, Discrete Mathematics 60 (1986), 101 – 120.
  • [9] H. Brezis, Functional analysis, sobolev spaces and partial differential equations, Universitext, Springer New York, 2010.
  • [10] B. Buet, G.P. Leonardi, and S. Masnou, Discretization and approximation of surfaces using varifolds, Geometric Flows 3 (2018), no. 1, 28–56.
  • [11] P. Cardaliaguet, G. Carlier, and B. Nazaret, Geodesics for a class of distances in the space of probability measures, Calculus of Variations and Partial Differential Equations 48 (2013), no. 3-4, 395–420.
  • [12] G. Carlier, On a theorem of Alexandrov, Journal of nonlinear and convex analysis 5 (2004), no. 1, 49–58.
  • [13] A. Chambolle and T. Pock, A first-order primal-dual algorithm for convex problems with applications to imaging, Journal of mathematical imaging and vision 40 (2011), no. 1, 120–145.
  • [14] N. Charon and A. Trouvé, The varifold representation of nonoriented shapes for diffeomorphic registration, SIAM Journal on Imaging Sciences 6 (2013), no. 4, 2547–2580.
  • [15] L. Chizat, G. Peyré, B. Schmitzer, and F-X. Vialard, An Interpolating Distance Between Optimal Transport and Fisher-Rao Metrics, Foundations of Computational Mathematics 18 (2018), no. 1, 1–44.
  • [16] D. Cohen-Steiner and J.M Morvan, Restricted Delaunay triangulation and normal cycle, Comput. Geom (2003), 312–321.
  • [17] M. Cuturi, Sinkhorn Distances: Lightspeed Computation of Optimal Transportation Distances, Advances in Neural Information Processing Systems 26 (2013).
  • [18] J. Delon, J. Salomon, and A. Sobolevski, Fast transport optimization for Monge costs on the circle, SIAM Journal on Applied Mathematics 70 (2010), no. 7, 2239–2258.
  • [19] S. Durrleman, X. Pennec, A. Trouvé, and N. Ayache, Statistical models of sets of curves and surfaces based on currents, Medical image analysis 13 (2009), no. 5, 793–808.
  • [20] L. Evans and R. Gariepy, Measure theory and fine properties of functions, CRC press, 2015.
  • [21] I. Fáry and E. Makai, Isoperimetry in variable metric, Studiu Scientiarum Mathematicarum Hungarica 17 (1982), 143 – 158.
  • [22] H. Federer, Geometric measure theory, Springer, 1969.
  • [23] W. Fenchel and B. Jessen, Mengenfunktionen und konvexe körper, Matematisk-fysiske meddelelser, Levin & Munksgaard, 1938.
  • [24] R. Gardner, The Brunn-Minkowski inequality, Bulletin of the American Mathematical Society 39 (2002), no. 3, 355–405.
  • [25] R. Gardner, M. Kiderlen, and P. Milanfar, Convergence of Algorithms for Reconstructing Convex Bodies and Directional Measures, The Annals of Statistics 34 (2006), no. 3, 1331–1374.
  • [26] J. Glaunès, A. Trouvé, and L. Younes, Diffeomorphic matching of distributions: A new approach for unlabelled point-sets and sub-manifolds matching, Computer Vision and Pattern Recognition (CVPR) 2 (2004), 712–718.
  • [27] J. Glaunès and M. Vaillant, Surface matching via currents, Proceedings of Information Processing in Medical Imaging (IPMI), Lecture Notes in Computer Science 3565 (2006), no. 381-392.
  • [28] A. Gretton, K. Borgwardt, M. Rasch, B. Schölkopf, and A.J. Smola, A kernel method for the two-sample-problem, Advances in neural information processing systems, 2007, pp. 513–520.
  • [29] M. Hebert, K. Ikeuchi, and H. Delingette, A spherical representation for recognition of free-form surfaces, IEEE transactions on pattern analysis and machine intelligence 17 (1995), no. 7, 681–690.
  • [30] H. Heijmans and A. Tuzikov, Similarity and symmetry measures for convex shapes using Minkowski addition, IEEE Transactions on Pattern Analysis and Machine Intelligence 20 (1998), no. 9, 980–993.
  • [31] Y. Hu, M. Hudelson, B. Krishnamoorthy, A. Tumurbaatar, and K.R. Vixie, Median Shapes, Journal of Computational Geometry 10 (2019), 322–388.
  • [32] I. Kaltenmark, B. Charlier, and N. Charon, A general framework for curve and surface comparison and registration with oriented varifolds, Computer Vision and Pattern Recognition (CVPR) (2017).
  • [33] T. Lachand-Robert and E. Oudet, Minimizing within convex bodies using a convex hull method, SIAM Journal on Optimization 16 (2005), no. 2, 368–379.
  • [34] G. Letac, Mesures sur le cercle et convexes du plan, Annales scientifiques de l’Université de Clermont-Ferrand 2. Série Probabilités et applications 76 (1983), no. 1, 35–65.
  • [35] Y. Li, K. Swersky, and R. Zemel, Generative moment matching networks, International Conference on Machine Learning, 2015, pp. 1718–1727.
  • [36] M. Liero, A. Mielke, and G. Savaré, Optimal transport in competition with reaction: The Hellinger–Kantorovich distance and geodesic curves, SIAM Journal on Mathematical Analysis 48 (2016), no. 4, 2869–2911.
  • [37] N. Papadakis, G. Peyré, and E. Oudet, Optimal transport with proximal splitting, SIAM Journal on Imaging Sciences 7 (2014), no. 1, 212–238.
  • [38] B. Piccoli and F. Rossi, Generalized Wasserstein Distance and its Application to Transport Equations with Source, Archive for Rational Mechanics and Analysis 211 (2014), no. 1, 335–358.
  • [39] J.L. Prince and A. S. Willsky, Reconstructing convex sets from support line measurements, IEEE Transactions on Pattern Analysis and Machine Intelligence 12 (1990), no. 4, 377–389.
  • [40] R. Rockafellar, Integrals which are convex functionals. II, Pacific Journal of Mathematics 39 (1971), no. 2, 439–469.
  • [41] P. Roussillon and J. Glaunès, Kernel Metrics on Normal Cycles and Application to Curve Matching, SIAM Journal on Imaging Sciences 9 (2016), no. 4, 1991–2038.
  • [42] R. Schneider, Convex Bodies: The Brunn-Minkowski Theory, 2 ed., Encyclopedia of Mathematics and its Applications, Cambridge University Press, 2013.
  • [43] B. Sriperumbudur, K. Fukumizu, and G. Lanckriet, Universality, characteristic kernels and RKHS embedding of measures, Journal of Machine Learning Research 12 (2011), 2389–2410.
  • [44] M. Van Kreveld, O. Schwarzkopf, M. de Berg, and M. Overmars, Computational geometry algorithms and applications, Springer, 2000.
  • [45] C. Villani, Optimal transport: old and new, vol. 338, Springer Science & Business Media, 2008.
  • [46] L. Younes, Shapes and diffeomorphisms, Springer, 2019.
  • [47] H. Zouaki, Representation and geometric computation using the extended Gaussian image, Pattern recognition letters 24 (2003), no. 9-10, 1489–1501.