Frontiers journals are at the top of citation and impact metrics

Original Research ARTICLE

Front. Appl. Math. Stat., 12 December 2018 |

A Variational Model for Data Fitting on Manifolds by Minimizing the Acceleration of a Bézier Curve

  • 1Research Group Numerical Mathematics (Partial Differential Equations), Faculty of Mathematics, Technische Universität Chemnitz, Chemnitz, Germany
  • 2ICTEAM Institute, Université Catholique de Louvain, Louvain-la-Neuve, Belgium

We derive a variational model to fit a composite Bézier curve to a set of data points on a Riemannian manifold. The resulting curve is obtained in such a way that its mean squared acceleration is minimal in addition to remaining close the data points. We approximate the acceleration by discretizing the squared second order derivative along the curve. We derive a closed-form, numerically stable and efficient algorithm to compute the gradient of a Bézier curve on manifolds with respect to its control points, expressed as a concatenation of so-called adjoint Jacobi fields. Several examples illustrate the capabilities and validity of this approach both for interpolation and approximation. The examples also illustrate that the approach outperforms previous works tackling this problem.

AMS subject classification (2010). 65K10, 65D10, 65D25, 53C22, 49Q99.

1. Introduction

This papers addresses the problem of fitting a smooth curve to data points d0, …, dn lying on a Riemannian manifold M and associated with real-valued parameters t0, …, tn. The curve strikes a balance between a data proximity constraint and a smoothing regularization constraint.

Several applications motivate this problem in engineering and the sciences. For instance, curve fitting is of high interest in projection-based model order reduction of one-dimensional dynamical systems [1]. In that application, the dynamical system depends on the Reynolds number and the model reduction is obtained by computing suitable projectors as points on a Grassmann manifold. Finding a projector is however a time- and memory-consuming task. Based on projectors precomputed for given parameter values, fitting is used to approximate the projector associated with a new parameter value. In Gousenbourger et al. [2], the same strategy is used to approximate wind field orientations represented as points on the manifold of positive semidefinite covariance matrices of size p and rank r. Further applications are involving rigid-body motions on SE(3), like Cosserat rods applications [3], or orientation tracking [4]. Sphere-valued data are also of interest in many applications of data analysis, for storm tracking and prediction, or the study of bird migration [5].

There exists different approaches to tackle the curve fitting problem. Among others, we name here the subdivision schemes approach [6, 7] or the Lie-algebraic methods [8]. However, the most popular approach nowadays is probably to encapsulate the two above-mentioned constraints into an optimization problem

minγΓEλ(γ):=minγΓt0tnD2γ(t)dt2γ(t)2dt+λ2i=0nd2(γ(ti),di)),    (1)

where Γ is an admissible space of curves γ:[t0,tn]M, t↦γ(t), D2dt2 denotes the (Levi-Civita) second covariant derivative, and ∥·∥γ(t) denotes the Riemannian metric at γ(t) from which we can define a Riemannian distance d(·, ·). Finally, λ∈ℝ is a parameter that strikes the balance between the regularizer t0tnD2γ(t)dt2γ(t)2dt and the fitting term i=0nd2(γ(ti),di)). This approach leads to the remarkable property that, when the manifold M reduces to the Euclidean space and Γ is the Sobolev space H2(t0,tn), the solution to (1) is the natural cubic spline [9].

Optimization on manifolds has gained a lot of interest this last decade, starting with the textbook [10] that summarizes several optimization methods on matrix manifolds. Recently, toolboxes have emerged, providing easy access to such optimization methods, e.g., Manopt [11] and MVIRT [12]. The former received a very positive return in many different topics of research, with applications for example in low-rank modeling in image analysis [13], dimensionality reduction [14], phase retrieval [15] or even 5G-like MIMO systems [16]. The latter stems from recent interest in manifold-valued image and data processing, phrased as variational models on the product manifold MN, where N is the number of pixels. Starting with total variation (TV) regularization of phase-valued data [17, 18], different methods for TV on manifolds [19, 20] have been developed as well as second order methods [21, 22] up to infimal convolution [23] and total generalized variation [23, 24]. Furthermore, different algorithms have been generalized to manifold-valued data, besides the previous works using gradient descent or cyclic proximal point methods, a Douglas–Rachford splitting [25], iteratively reweighted least squares [26] and more general half-quadratic minimization [27] have been introduced.

The curve fitting problem (1), has been tackled differently the past few years. Samir et al. [28] considered the case where Γ is an infinite dimensional Sobolev space of curves, and used the Palais-metric to design a gradient descent algorithm for (1) (see e.g., [29] for an application of this approach). Another method consists in discretizing the curve γ in N points, and therefore considering Γ=MN (see e.g., [30] for a result on SO(3)). Finally, the limit case where λ → 0 is already well studied and known as the geodesic regression [3133].

A recent topic concerns curve fitting by means of Bézier curves. In that approach, the search space Γ is reduced to as set of composite Bézier curves. Those are a very versatile tool to model smooth curves and surfaces for real- and vector-valued discrete data points (see [34] for a comprehensive textbook), but they can also be used to model smooth curves and surfaces for manifold-valued data [35, 36]. The advantage to work with such objects, compared to classical approaches, are that (i) the search space is drastically reduced to the so-called control points of the Bézier curves (and this leads to better time and memory performances) and (ii) it is very simple to impose differentiability for the optimal curve, which is appreciated in several of the above-mentioned applications. However, while obtaining such an optimal curve reduces directly to solving a linear system of equations for data given on a Euclidean space, there is up to now no known closed form of the optimal Bézier curve for manifold valued data.

In this work, we derive a gradient descent algorithm to compute a differentiable composite Bézier curve B:[t0,tn]M that satisfies (1), i.e., such that B(t) has a minimal mean squared acceleration, and fits the set of n+1 manifold-valued data points at their associated time-parameters. We consider the manifold-valued generalization of Bézier curves [36] in the same setting as in Arnould et al. [37], or more recently in Gousenbourger et al. [38]. We employ the (squared) second order absolute difference introduced in Bačák et al. [22] to obtain a discrete approximation of the regularizer from (1). The quality of the approximation depends only on the number of sampling points. We exploit the recursive structure of the De Casteljau algorithm [36] to derive the gradient of the objective function with respect to the control points of B(t). The gradient is built as a recursion of Jacobi fields that, for numerical reasons, are implemented as a concatenation of so-called adjoint Jacobi fields. Furthermore, the corresponding variational model only depends on the number of control points of the composite Bézier curve, and not on the number of sampling points. We finally obtain an approximating model to (1) that we solve with an algorithm only based on three tools on the manifold: the exponential map, the logarithmic map, and a certain Jacobi field along geodesics.

The paper is organized as follows. We introduce the necessary preliminaries—Bézier curves, Riemannian manifolds and Riemannian second order finite differences— in section 2. In section 3 we derive the gradient of the discretized mean squared acceleration of the composite Bézier curve with respect to its control points, and thus of the regularizer of (1). In section 4, we present the corresponding gradient descent algorithm, as well as an efficient gradient evaluation method, to solve (1) for different values of λ. The limit case where λ → ∞ is studied as well. Finally, in section 5, we validate, analyze and illustrate the performance of the algorithm for several numerical examples on the sphere S2 and on the special orthogonal group SO(3). We also compare our solution to existing Bézier fitting methods. A conclusion is given in section 6.

2. Preliminaries

2.1. Bézier Functions and Composite Bézier Spline

Consider the Euclidean space ℝm. A Bézier curve of degree K∈ℕ is a function βK:[0,1]m parametrized by control points b0,,bKm and taking the form

βK(t;b0,,bK) :=j=0KbjBj,K(t),

where Bj,K(t)=(Kj)tj(1-t)K-j are the Bernstein basis polynomials [34]. For example the linear Bézier curve β1 is just the line segment (1−t)b0+tb1 connecting the two control points b0 and b1. The explicit formulae of the quadratic and cubic Bézier curves read

      β2(t;b0,b1,b2)=b0(1t)2+2b1(1t)t+b2t2,β3(t;b0,b1,b2,b3)=b0(1t)3+3b1(1t)2t+3b2(1t)t2                                      +b3t3,

for given control points b0,b1,b2m and an additional point b3m for the cubic case.

A composite Bézier curve is a function B:[0, n] → ℝm composed of n Bézier curves and defined as

B(t):={βK0(t;b00,,bK00)             if t[0,1]βKi(ti;b0i,,bKii)       if t(i,i+1],                                              i{1,,n1},    (2)

where Ki∈ℕ denotes the degree of the ith Bézier curve βKi of B and bji, j = 0, …, Ki, are its control points. Furthermore, B(t) is continuous if the last and first control points of two consecutive segments coincide [34]. We introduce pi as the point at the junction of two consecutive Bézier segments, i.e., pi:=bKi-1i-1=b0i. Differentiability is obtained if the control points of two consecutive segments are aligned. We introduce further bi+1-:=bKi-1i and bi+1+:=b1i+1 such that the differentiability condition reads pi+1=Kibi+1-+Ki+1b+i+1Ki+Ki+1.

Example 1 (C1 conditions for composite cubic Bézier curves). We consider the case where the Bézier segments are all cubic, as represented in Figure 1. The composite Bézier curve B(t) is C1 if pi=12(bi-+bi+), i = 1, …, 4, with pi=b3i-1=b0i, bi-=b2i-1, and bi+=b1i.


Figure 1. Schematic representation of the composite cubic Bézier curve B:[0,5]M, for M= (black), and its control points (green circles). Continuous differentiability is reached at the junction of the segments (the blue arrows draw the first derivative of B).

2.2. Riemannian Manifolds

We consider a complete m-dimensional Riemannian manifold M. We refer to O'Neill [39] and do Carmo [40] for an introduction to Riemannian manifolds and to [10] for optimization thereon. We will use the following terminology and notations.

We denote by TaM the (Euclidean) tangent space to M at aM; TM:=aMTaM is the tangent bundle to M; 〈·, ·〉a denotes the inner product in the tangent space TaM at a and from which we deduce the norm of vTaM denoted by va=v,va. For a (not necessarily unique) shortest geodesic between a and bM, we write g(·;a,b):M, tg(t; a, b), parametrized such that g(0;a, b) = a and g(1;a, b) = b. This choice of parametrization also means that the covariant derivative Ddt of g with respect to time satisfies Ddtg(t;a,b)g(t)=dM(a,b), for all t∈[0, 1], where dM(a,b) is the geodesic distance between a and bM. The Riemannian exponential reads expa:TaMM,vb=expa(v) and we denote by ra∈ℝ the maximal radius such that the exponential map is bijective on Da:={bM:dM(a,b)<ra}. Then loga:DaTaM,bv=loga(b) is called the Riemannian logarithm which is (locally) the inverse of the exponential. A Riemannian manifold is called symmetric in xM if the geodesic reflection sx at xM given by the mapping γ(t)↦γ(−t) is an isometry at least locally near x, for all geodesics through γ(0) = x. If M is symmetric in every xM, the manifold is called (Riemannian) symmetric space or symmetric manifold.

In the following we assume, that both the exponential and the logarithmic map are available for the manifold and that they are computationally not too expensive to evaluate. Furthermore, we assume that the manifold is symmetric.

2.3. Composite Bézier Curves on Manifolds

One well-known way to generalize Bézier curves to a Riemannian manifold M is via the De Casteljau algorithm [36, section 2]. This algorithm only requires the Riemannian exponential and logarithm and conserves the interpolation property of the first and last control points. Some examples on interpolation and fitting with Bézier curves or Bézier surfaces, i.e., generalizations of tensor product Bézier curves, can be found in Absil et al. [35].

Consider βK:[0,1]M,tβK(t;b0,b1,,bK), the manifold-valued Bézier curve of order K driven by K+1 control points b0,,bKM. We introduce the points xi[0]=bi and iterate the construction of further points. For i = 0, …, Kk, k = 1, …, K, we define

xi[k]:=βk(t;bi,,bi+k)=g(t;xi[k1],xi+1[k1])    (3)

as the ith point of the kth step of the De Casteljau algorithm, and obtain βK(t;b0,,bK)=x0[K].

The De Casteljau algorithm is illustrated on Figure 2 for a Euclidean cubic Bézier curve β3(t; b0, b1, b2, b3). The general cubic Bézier curve can be explicitly expressed on a manifold M as

β3(t;x0[0],x1[0],x2[0],x3[0])=g(t;g(t;g(t;x0[0],x1[0]),g(t;x1[0],x2[0])),                                                             g(t;g(t;x1[0],x2[0]),g(t;x2[0],x3[0])))                                                    =g(t;g(t;x0[1],x1[1]),g(t;x1[1],x2[1]))                                                    =g(t;x0[2],x1[2])                                                    =x0[3].

Figure 2. Construction of a cubic Bézier curve via the De Casteljau algorithm.

The conditions of continuity and differentiability are generalized to manifolds in Popiel and Noakes [36].

Lemma 2 (Differentiability conditions [36]). Consider the composite Bézier curve B:[0,2]M consisting of f:[0,1]M,tf(t)=βK(t;b00,b10,,bK0) and g:[1,2]M,tg(t)=βK̄(t-1;b01,b11,,bK̄1), i.e.,

B(t):={f(t) for t[0,1],g(t) for t(1,2].

The composite Bézier curve B(t) is continuous and continuously differentiable if the two following conditions hold:

bK0=b01 and bK0=g(s;bK10,b11),s=KK+K¯.    (4)

The composite Bézier curve B:[0,n]M is then defined completely analogously to the Euclidean case from Equation (2). The differentiability conditions (4) of Lemma 2 have to be satisfied at each junction point pi, i = 1, …, n − 1.

Example 3 (Composite cubic Bézier curves). The composite cubic Bézier curve B(t):[0,n]M is C1 if pi=g(12;bi-,bi+). See Figure 1 for an example on M= with n = 5 segments and Figure 3 for an example on M=S2 with n = 3 segments.


Figure 3. A composite cubic Bézier curve B:[0, 3] → S2. The end points pi, i = 0, …, 3, (cyan) and intermediate points bi± (green) determine its shape; continuous differentiability is illustrated by the logarithmic maps logpibi±, i∈{1, 2} (blue arrows).

2.4. Discrete Approximation of the Mean Squared Acceleration

We discretize the mean squared acceleration (MSA) of a curve γ:[0,1]M, by discretizing the corresponding integral

01D2γ(t)dt2γ(t)2dt,    (5)

i.e., the regularizer from (1). We approximate the squared norm of the second (covariant) derivative by the second order absolute finite difference introduced by Bačák et al. [22]. Consider three points x,y,zM and the set of mid-points of x and z


for all (not necessarily shortest) geodesics g connecting x and z. The manifold-valued second order absolute finite difference is defined by

d2[x,y,z]=mincCx,z2dM(c,y).    (6)

This definition is equivalent, on the Euclidean space, to x-2y+z=212(x+z)-y.

Using equispaced points t0, …, tN, N∈ℕ, with step size Δt=t1-t0=1N, we approximate D2γ(ti)dt2γ(ti)1Δt2d2[γ(ti-1),γ(ti),γ(ti+1)], i = 1, …, N−1, and obtain by the trapezoidal rule


For Bézier curves γ(t) = B(t) we obtain for the regularizer in (1) the discretized MSA A(b) that depends on the control points b and reads

A(b):=i=1N1d22[B(ti1),B(ti),B(ti+1)]Δt3.    (7)

3. The Gradient of the Discretized Mean Squared Acceleration

In order to minimize the discretized MSA A(b), we aim to employ a gradient descent algorithm on the product manifold MM, where M is the number of elements in b. In the following, we derive a closed form of the gradient ∇bA(b) of the discretized MSA (7). This gradient is obtained by means of a recursion and the chain rule. In fact, the derivative of (6) is already known [22], such that it only remains to compute the derivative of the composite Bézier curve.

We first introduce the following notation. We denote by Dxf[η](x0)Tf(x0)M the directional derivative of f:MM evaluated at x0, with respect to its argument x and in the direction ηTxM. We use the short hand Dxf[η] = Dxf[η](x) whenever this directional derivative is evaluated afterwards again at x.

We now state the two following definitions, which are crucial for the rest of this section.

Definition 4 (Gradient [10, Equation (3.31), p. 46]). Let f:M be a real-valued function on a manifold M, xM and ηTxM.

The gradient Mf(x)TxM of f at x is defined as the tangent vector that fulfills

f(x),ηx=Dxf[η]    for all    ηTx.    (8)

For multivariate functions f(x, y), we denote the gradient of f with respect to x at (x0, y0) by writing M,xf(x0,y0). We shorten this notation as M,xf=M,xf(x,y) when this gradient is seen as a function of x (and y).

Definition 5 (Chain rule on manifolds [10, p. 195]). Let f:MM, h:MM be two functions on a manifold M and F:MM, xF(x) = (f ° h)(x) = f(h(x)), their composition. Let xM and ηTxM. The directional derivative DxF[η] of F with respect to x in the direction η is given by

DxF[η]=Dh(x)f[Dxh[η]],    (9)

where Dxh[η]Th(x)M and DxF[η]TF(x)M.

The remainder of this section is organized in four parts. We first recall the theory on Jacobi fields in section 3.1 and their relation to the differential of geodesics (with respect to start and end point). In section 3.2, we apply the chain rule to the composition of two geodesics, which appears within the De Casteljau algorithm. We use this result to build an algorithmic derivation of the differential of a general Bézier curve on manifolds with respect to its control points (section 3.3). We extend the result to composite Bézier curves in section 3.4, including their constraints on junction points pi to enforce the C1 condition (4), and finally gather these results to state the gradient MA(b) of the discretized MSA (7) with respect to the control points.

3.1. Jacobi Fields as Derivative of a Geodesic

In the following, we introduce a closed form of the differential Dxg(t; ·, y) of a geodesic g(t; x, y), t∈[0, 1], with respect to its start point xM. The differential with respect to the end point yM can be obtained by taking the geodesic g(t, y, x) = g(1−t; x, y).

As represented in Figure 4, we denote by γx, ξ, the geodesic starting in γx, ξ(0) = x and with direction Ddtγx,ξ(0)=ξTxM. We introduce ζ(s): = logγx, ξ(s)y, the tangential vector in Tγx,ξ(s)M pointing toward y. Then, the geodesic variation Γg, ξ(s, t) of the geodesic g(·;x, y) with respect to the tangential direction ξTxM is given by

Γg,ξ(s,t):=expγx,ξ(s)(tζ(s)),    s(ε,ε), t[0,1],

where ε>0. The corresponding Jacobi field Jg, ξ along g is then given by the vector field


that represents the direction of the displacement of g if x is perturbed in a direction ξ.


Figure 4. Schematic representation of the variation Γg, ξ(s, t) of a geodesic g w.r.t. the direction ξTxM. The corresponding Jacobi field along g and in the direction ξ is the vector field Jg,ξ(t)=sΓg,ξ(s,t)|s=0.

We directly obtain Jg, ξ(0) = ξ, and Jg, ξ(1) = 0 as well as Jg,ξ(t)Tg(t;x,y)M. Since Γg, ξ(s, t) = g(t; γx, ξ(s), y) we obtain by the chain rule

Dxg(t,·,y)[ξ]=ddsg(t;γx,ξ(s),y)|s=0=ddsΓg,ξ(s,t)|s=0=Jg,ξ(t).    (10)

Remark. This relation between the derivative of geodesics and Jacobi fields is of high interest on symmetric spaces, where Jacobi fields can be computed in closed form, as summarized in the following Lemma.

Lemma 6. [22, Prop. 3.5] Let M be a m-dimensional symmetric Riemannian manifold. Let g(t; x, y), t ∈ [0, 1], be a geodesic between x,yM, ηTxM a tangent vector and {ξ1, …ξm} be an orthonormal basis (ONB) of TxM that diagonalizes the curvature operator of M with eigenvalues κ, ℓ = 1, …, m. For details, see Ch. 4.2 and 5 (Ex. 5) of [40]. Let further denote by {Ξ1(t), …, Ξm(t)} the parallel transported frame of {ξ1, …, ξm} along g. Decomposing η==1mηξTxM, the derivative Dxg[η] becomes


where the Jacobi field Jg,ξ:Tg(t;x,y)M along g and in the direction ξ is given by

Jg,ξ(t)={sinh(dg(1t)κ)sinh(dgκ)Ξ(t)if κ<0,sin(dg(1t)κ)sin(κdg)Ξ(t)if κ>0,(1t)Ξ(t)if κ=0,    (11)

with dg = d(x, y) denoting the length of the geodesic g(t; x, y), t∈[0, 1].

The Jacobi field of the reversed geodesic ḡ(t): = g(t; y, x) = g(1−t; x, y) is obtained using the same orthonormal basis and transported frame but evaluated at s = 1−t. We thus obtain Dyg(t; x, y)[ξ] = Dyg(1−t; y, x)[ξ] = Jḡ,ξ(1−t), where

Jg¯,ξ(1t)={sinh(dgtκ)sinh(dgκ)Ξ(t)if κ<0,sin(dgtκ)sin(dgκ)Ξ(t)if κ>0,tΞ(t)if κ=0.

Note that Tg(t)M=T(1-t)M. Therefore Ξ(t)Tg(t)M, ℓ = 1, …, m, is an orthonormal basis for this tangent space.

3.2. Derivative of Coupled Geodesics

Let M be a symmetric Riemannian manifold. We use the result of Lemma 6 to directly compute the derivative of coupled geodesics, i.e., a function composed of g1(t): = g(t; x, y) and g2(t): = g(t; g1(t), z). By Definition 5, we have


and by (10), we obtain


where the variation direction used in the Jacobi field is now the derivative of g1(t) in direction η. Similarly, we compute the derivative of a reversed coupled geodesic g3(t): = g(t; z, g1(t)) as


Note that the Jacobi field is reversed here, but that its variation direction is the same as the one of the Jacobi field introduced for g2(t). In a computational perspective, it means that we can use the same ONB for the derivatives of both g3 and ḡ3 Furthermore, in this case, the variation direction is also computed by a Jacobi field since Dxg1(t)[η] = Jg1, η(t).

Finally the derivative of g2 (resp. g3) on symmetric spaces is obtained as follows. Let {ξ1[1],,ξm[1]} be an ONB of TxM for the inner Jacobi field along g1, and {ξ1[2],,ξm[2]} be an ONB of Tg1(t)M for the outer Jacobi field along g2 (resp. g3). As η==1mηξ[1]TxM, and stating Jg1,ξ[1](t)=l=1mμlξl[2]Tg1(t)M, the derivative of g2 (resp. g3) with respect to x in the direction ηTxM reads

Dxg2(t)[η]=l=1m=1mJg2,ξl[2](t)μlη,    (12)

and accordingly for g3.

3.3. Derivative of a Bézier Curve

Sections 3.1 and 3.2 introduced the necessary concepts to compute the derivative of a general Bézier curve βK(t; b0, …, bK), as described in Equation (3), with respect to its control points bj. For readability of the recursive structure investigated in the following, we introduce a slightly simpler notation and the following setting.

Let K be the degree of a Bézier curve βK(t; b0, …, bK) with the control points b0,,bKM. We fix k∈{1, …, K}, i∈{0, …, Kk} and t∈[0, 1]. We introduce

gi[k](t):=g(t;gi[k1](t),gi+1[k1](t))=βi[k](t;bi,,bi+k),    (13)

for the ith Bézier curve of degree k in the De Casteljau algorithm, and gi[0](t)=bi.

Furthermore, given x∈{bi, …, bi+k}, we denote by

ηi[k]:=Dxgi[k](t)[η],    (14)

its derivative with respect to one of its control points x in the direction ηTxM.

Remark. Clearly any other derivative of gi[k] with respect to x = bj, j<i or j>i+k is zero. In addition we have ηi[0]=Dxgi[0][η]=η for x = bi and zero otherwise.

Theorem 7 (Derivative of a Bézier curve). Let k ∈ {1, …, K}and i ∈ {0, …, Kk}be given. The derivative ηi[k]=Dxgi[k](t)[η] of gi[k] with respect to its control point x: = bj, iji+k, and in the direction ηTxM is given by

ηi[k]:=Dxgi[k](t)[η]                  ={Jgi[k],ηi[k1](t) if j=i,Jgi[k]k,ηi[k1](t)+Jg¯i[k],ηi+1[k1](1t) if i<j<i+k,Jg¯i[k],ηi+1[k1](1t) if j=i+k.

Proof. Let fix t∈[0, 1] and x = bj, iji+k. For readability we set a:=gi[k-1](t), b:=gi+1[k-1](t), and f:=gi[k](t)=g(t;a,b). Note that while f depends on the control points bi, …, bi+k and is a Bézier curve of degree k, both a and b are Bézier curves of degree k−1. The former does not depend on bi+k, and the latter is independent of bi.

We prove the claim by induction. For k = 1 the function gi[1] is just a geodesic. The case i<j<i+1 does not occur and the remaining first and third cases follow by the notation introduced for k = 0 and Lemma 6.

For k>1 we apply the chain rule (9) to Dxf[η] and obtain


Consider the first term Daf[Dxa[η]] and j<i+k. By (10) and the notation from (14), one directly has


For j = i+k, clearly Dxa[η] = Daf[Dxa[η]] = 0, as a does not depend on bi+k.

We proof the second term similarly. For j>i, by applying the chain rule and using the reversed Jacobi field formulation of Lemma 6 for the derivative of a geodesic with respect to its end point, we obtain


Finally, as Dxb[η] = Dbf[Dxb[η]] = 0 for x = bi, the assumption follows.

Figure 5 represents one level of the schematic propagation tree to compute the derivative of a Bézier curve.


Figure 5. Schematic representation of the cases where elements compose the chained derivative of the ith Bézier curve of order k in the De Casteljau algorithm. The solid line represents a Jacobi field along gi[k], while the dashed one represents a reversed Jacobi field. (A) Represents the case x = bi, (B) represents the cases where x∈{bi+1, …, bi+k+1}, and (C) represents the case x = bi+k.

Example 8 (Quadratic Bézier curve). Consider the quadratic Bézier curve β2:[0,1]M defined as


Using the notations (13), we have

g0[1](t):=g(t;b0,b1),    g1[1](t):=g(t;b1,b2),g0[2](t):=g(t;g0[1],g1[1]).

The derivative of β2 at t with respect to b0 in the direction ηTboM, is given by

Db0β2[η]=Jg0[2],η0[1](t), with η0[1]:=Jg0[1]1,η(t).

The derivative of β2 at t with respect to b2 in the direction ηTb2M can be seen as deriving by the first point after inverting the Bezier curve, i.e., looking at β̄2(t)=β2(1-t), hence we have analogously to the first term

Db2β2[η]=Jg¯0[2],η1[1](1t), with η1[1]:=Jg¯1[1],η(1t).

The case Db1β2[η], ηTb1M, involves a chain rule, where b1 appears in both g0[1] (as its end point) and g1[1] (as its starting point). Using the two intermediate results (or Jacobi fields of geodesics)

η0[1]:=Jg¯0[1],η(1t)    and    η1[1]:=Jg1[1],η(t),

we obtain


Example 9 (Cubic Bézier curve). Consider the cubic Bézier curve β3:[0,1]M defined as


As in Example 8, we use the notations (13) and define

gj[1](t):=g(t;bj,bj+1),    j=0,1,2,gj[2](t):=g(t;gj[1],gj+1[1]),    j=0,1, and g0[3](t):=g(t;g0[2],g1[2]).

The derivation of β3 with respect to b0 or b3 follows the same structure as in Example 8. The case of Db1β3[η], however, requires two chain rules. The needed Jacobi fields follow the tree structure shown in Figure 6B: given ηTb1M, we define at the first recursion step

η0[1]:=Jg¯0[1],η(1t),    η1[1]:=Jg1[1],η(t),

and at the second recursion step

η0[2]:=Jg0[2],η0[1](t)+Jg¯0[2],η1[1](1t),    η1[2]:=Jg1[2],η1[1](t).

Note that both η0[2] and η1[2] are actually the derivative of β2(t; b0, b1, b2) and β2(t; b1, b2, b3), respectively, with respect to b1 and direction ηTb1M. Finally we have


Figure 6. Construction and derivation tree of a Bézier curve β3(t; b0, b1, b2, b3). The derivative with respect to a variable bi is obtained by a recursion of Jacobi fields added at each leaf of the tree. (A) Tree-representation of the construction of a cubic Bezier curve. The thick line tracks the propagation of b1 within the tree. (B) Tree-representation of the construction of η0[3]:=Db1β3[η]. The solid lines are Jacobi fields while dashed lines are reversed Jacobi fields.

The case of Db2β3[η] is obtained symmetrically, with arguments similar to b1.

Note that computing the Jacobi fields involved just follows the same decomposition tree as the De Casteljau algorithm (Figure 6A).

3.4. Joining Segments and Deriving the Gradient

In this subsection we derive the differential of a composite Bézier curve B(t) consisting of n segments and take the C1 conditions into account. We simplify the notations from section 2.1 and set the degree fixed to Ki = K for all segments, e.g., K = 3 for a cubic composite Bézier curve. Then the control points are bji, j = 0, …, K, i = 0, …, n−1. We further denote by pi=bKi-1=b0i, i = 1, …, n−1 the common junction point of the segments and p0 and pn the start and end points, respectively. For ease of notation we denote by bi-=bK-1i and bi+=b1i, i = 1, …, n−1, the two points needed for differentiability (C1) condition investigation, cf. Figure 1 for an illustration of the framework on M=, with K = 3.

One possibility to enforce the C1 condition (4) is to include it into the composite Bézier curve by replacing bi+ with

bi+=g(2;bi,pi),    i=1,,n1.    (15)

This way both the directional derivatives of B(t) with respect to bi- and pi change due to a further (most inner) chain rule.

Lemma 10 (Derivative of a composite Bézier curve with C1 condition). Let B be a composite Bézier curve and pi,bi+,bi-introduced as above. Replacing bi+=g(2;bi-,pi) eliminates that variable from the composite Bezier curve and keeps the remaining differentials unchanged, except the following which now read

DbiB(t)[η]={DbiβK(ti+1;    pi1,bi1+,,·,pi)[η]t(i1,i],Dbi+βK(ti;      pi,bi+,,·,pi+1)      [Dbig(2;·,pi)[η]]t(i,i+1],


DpiB(t)[η]={DpiβK(ti+1;pi1,    bi1+,,bi,·)[η]t(i1,i],DpiβK(ti;·,    bi+,,bi+1,pi+1)[η]+Dbi+βK(ti;pi,·,,     bi+1,pi+1)[Dpig(2;bi,·)[η]]t(i,i+1].

In both cases, the first interval includes i−1 = 0, when i = 1.

Proof. Both first cases are from the derivation of a Bézier curve as before, for both second cases replacing bi+=g(2;bi-,pi) yields one (for pi additional) term.

We now derive the gradient of the objective function (7). We introduce the abbreviation Bi=B(ti)M, and d2,i2=d22(Bi-1,Bi,Bi+1).

Theorem 11. Let M be a m-dimensional manifold, x be one of the control points of a composite Bézier curve B, and1, …, ξm} be a corresponding orthonormal basis (ONB) of TxM. The gradient M,xA(b) of the discretized mean squared acceleration A(b) of B, w.r.t. x, discretized at N+1 equispaced times t0, …, tN is given by


Proof: As M,xA(b)TxM, we seek for the coefficients a: = a(x) such that

Mf(x)==1maξ.    (16)

Therefore, for any tangential vector η:==1mηξTxM, we have,

M,xA(b),ηx==1maη.    (17)

By definition of A and Definition 4 this yields

M,xA(b),ηx= ​1Δt3i=1N1Md2,i2(x),ηx= ​1Δt3i=1N1Dxd2,i2[η].    (18)

We compute Dxd2,i2[η] using the chain rule (Definition 5) as

Dxd2,i2[η]=j=i1i+1DBjd2,i2[DxBj[η]],    (19)

which, by Definition 4, again becomes


The term on the left of the inner product is given in Bačák et al. [22, section 3] and the right term is given in section 3.3. While the former can be computed using Jacobi fields and a logarithmic map, the latter is the iteratively coupling of Jacobi fields. Furthermore, the differential DxBj[η] can be written as


Hence, we obtain


and by (17), (18), and (19), it follows


which yields the assertion (16).

4. Application to the Fitting Problem

The fitting problem has been tackled different ways this last decades. The approach with Bézier curves is more recent, and we refer to Absil et al. [35], Arnould et al. [37], and Gousenbourger et al. [38] for a detailed overview of these methods.

In this section, we present the numerical framework we use in order to fit a composite Bézier curve B to a set of data points d0,,dnM associated with time-parameters t0, …, tn, such that we meet (1). For the sake of simplicity, we limit the study to the case where ti = i, i = 0, …, n. Therefore, the fitting problem (1) becomes

minbΓBEλ(b):=minbΓBt0tnD2B(t)dt2B(t)2dt+λ2i=0nd2(pi,di),    (20)

where ΓBMM is the set of the M control points of B. Remark that, compared to (1), the optimization is now done on the product manifold MM. Furthermore, the fitting term now implies a distance between di and pi, as B(ti) = pi.

The section is divided in three parts: the product manifold MM is defined in section 4.1, where the contribution of the fitting term in the gradient of E is also presented. Then, we propose an efficient algorithm to compute the gradient of the discretized MSA, based on so-called adjoint Jacobi fields. We finally shortly mention the gradient descent algorithm we use as well as the involved Armijo rule.

4.1. Fitting and Interpolation

Let us clarify the set ΓBMM from (20). We will explicitly present the vector b and state its size M. The set ΓB is the set of the M remaining free control points to optimize, when the C1 continuity constraints are imposed. We distinguish two cases: (i) the fitting case, that corresponds to formulae presented in section 3, and (ii) the interpolation case (λ → ∞) where the constraint di = pi is imposed as well.

For a given composite Bézier curve B:[0,n]M consisting of a Bézier curve of degree K on each segment, and given the C1 conditions (4), the segments are determined by the points

b=(p0,b0+,b20,,bK20,b1,p1,,bn,pn)M.    (21)

We investigate the length of M. First, bi+ is given by pi and bi- via (4). Second, for the segments i = 2, …, n we can further omit pi−1 since the value also occurs in the segment i−1 as last entry. Finally, the first segment contains the additional value of b0+ that is not fixed by C1 constraints. The first segment is thus composed of K+1 control points, while the n−1 remaining segments are determined by K−1 points. In total we obtain M = n(K−1)+2 control points to optimize.

Minimizing A(b) alone leads to the trivial solution, for any set of control points b = (x, …, x), xM, and this is why the fitting term from (20) is important.

Fitting (0 < λ < ∞). If the segment start and end points are obstructed by noise or allowed to move, we employ a fitting scheme to balance the importance given to the data points d0, …, dn. Equation (20) reads

argminbΓBA˜(b),    A˜(b):=A(b)+λ2i=0ndM2(di,pi),    (22)

where λ∈ℝ+ sets the priority to either the data term (large λ) or the mean squared acceleration (small λ) within the minimization. The gradient of the data term is given in Karcher [41], and the gradient of à is given by

MM,xA˜(b)={MM,xA(b)λlogpidi​​ if x=pi, i=0,,n,MM,xA(b)​​ otherwise.

Interpolation (λ → ∞). For interpolation we assume that the start point pi−1 and end point pi of the segments i = 1, …, n are fixed to given data di−1 and diM, respectively. The optimization of the discrete mean squared acceleration A(b) reads

argminbΓBA(b)    s. t.pi=di, i=0,,n.    (23)

Since the pi are fixed by constraint, they can be omitted from the vector b. We obtain


Since there are n+1 additional constraints, the minimization is hence performed on the product manifold MM, M′ = M−(n+1) = n(K−2)+1.

4.2. Adjoint Jacobi Fields

In the Euclidean space ℝm, the adjoint operator T* of a linear bounded operator T:ℝm → ℝq is the operator fulfilling

T(x),yq=x,T*(y)m,     for all xm, yq.

The same can be defined for a linear operator S:TxMTyM, x,yM, on a m-dimensional Riemannian manifold M. The adjoint operator S*:TyMTxM satisfies

S(η),νy=η,S*(ν)x,     for all ηTxM, νTyM.

We are interested in the case where S is the differential operator Dx of a geodesic F(x) = g(t; x, y) for some fixed t∈ℝ and yM. The differential DxF:TxMTF(x)M can be written as


where α are the coefficients of the Jacobi field (11), and ξ, Ξ(t) are given as in Lemma 6. To derive the adjoint differential (DxF)*:TF(x)MTxM we observe that for any νTF(x)M we have

DxF[η],νF(x)==1mη,ξxαΞ(t),νF(x)                                 =η,=1mΞ(t),νF(x)αξx.

Hence the adjoint differential is given by

(DxF)*[ν]==1mν,Ξ(t)F(x)αξ,    νTF(x)M.

We introduce the adjoint Jacobi field JF,ν*:TxM, νTF(x)M as


Note that evaluating the adjoint Jacobi field J* involves the same transported frame {Ξ1(t), …Ξm(t)} and the same coefficients α as the Jacobi field J, which means that the evaluation of the adjoint is in no way inferior to the Jacobi field itself.

The adjoint D* of the differential is useful in particular, when computing the gradient M(h°F) of the composition of F:MM with h:M. Setting y:=F(x)M, we obtain for any ηTxM that

M,x(h°F)(x),ηx=Dx(h°F)[η]                                                =DF(x)h[DxF[η]]                                             =M,yh(y),DxF[η]y                                            =(DxF)*[M,yh(y)],ηx.

Especially for the evaluation of the gradient of the composite function h°F we obtain


The main advantage of this technique appears in the case of composite functions, i.e., of the form h°F1°F2 (the generalization to composition with K functions is straightforward). The gradient M,x(h°F1°F2)(x) now reads,

M,x(hF1F2)(x)=(DxF2)*[M,y2hF1(y2)]                                 =JF2,M,y2hF1(y2))*(t).

The recursive computation of η[3]=M,xh(x) is then given by the following algorithm


Example 12. For h=d22:M3 we know M3h by Lemma 3.1 and Lemma 3.2 from Bačák et al. [22]. Let t1, t2, t3∈[0, 1] be time points, and bMM be a given (sub)set of the control points of a (composite) Bézier curve B. We define F:MMM3,bF(b)=(B(t1),B(t2),B(t3)) as the evaluations of B at the three given time points. The composition h°F hence consists of (in order of evaluation) the geodesic evaluations of the De Casteljau algorithm, the mid point function for the first and third time point and a distance function.

The recursive evaluation of the gradient starts with the gradient of the distance function. Then, for the first and third arguments, a mid point Jacobi field is applied. The result is plugged into the last geodesic evaluated within the De Casteljau “tree” of geodesics. At each geodesic, a tangent vector at a point g(tj; a, b) is the input for two adjoint Jacobi fields, one mapping to TaM, the other to TbM. This information is available throughout the recursion steps anyways. After traversing this tree backwards, one obtains the required gradient of h°F.

Note also that even the differentiability constraint (4) yields only two further (most outer) adjoint Jacobi fields, namely Jg(2,bi-,pi),M,bi+B(tj)*(2) and Jg~(2,bi-,pi),M,bi+B(tj)*(2). They correspond to variation of the start point bi- and the end point pi, respectively as stated in (10).

4.3. A Gradient Descent Algorithm

To address (22) or (23), we use a gradient descent algorithm, as described in Absil et al. [10, Ch. 4]. For completeness the algorithm is given in Algorithm 1.


Algorithm 1. Gradient descent algorithm on a manifold N=MM

The step sizes are given by the Armijo line search condition presented in Absil et al. [10, Def. 4.2.2]. Let N be a Riemannian manifold, x=x(k)N be an iterate of the gradient descent, and β, σ∈(0, 1), α>0. Let m be the smallest positive integer such that

F(x)F(expx(βmαNF(x)))σβmαNF(x)x.    (24)

We set the step size to sk:=βmα in Algorithm 1.

As a stopping criterion we use a maximal number kmax of iterations or a minimal change per iteration dN(xk,xk+1)<ε. In practice, this last criterion is matched first.

The gradient descent algorithm converges to a critical point if the function F is convex [10, sections 4.3 and 4.4]. The mid-points model (6) posesses two advantages: (i) the complete discretized MSA (7) consists of (chained) evaluations of geodesics and a distance function, and (ii) it reduces to the classical second order differences on the Euclidean space. However, this model is not convex on general manifolds. An example is given in the arXiv preprint (version 3) of Bačák et al. [22]1, Remark 4.6. Another possibility (also reducing to classical second order differences in Euclidean space) is the so-called Log model (see e.g., [42])


The (merely technical) disadvantage of the Log model is that the computation of the gradient involves further Jacobi fields than the one presented above, namely to compute the differentials of the logarithmic map both with respect to its argument Dxlogyx as well as its base point Dylogyx. Still, these can be given in closed form for symmetric Riemannian manifolds [23, Th. 7.2][43, Lem. 2.3]. To the best of our knowledge, the joint convexity of the Log model in x, y and z is still an open question.

5. Examples

In this section, we provide several examples of our algorithm applied to the fitting problem (20).

We validate it first on the Euclidean space and verify that it retrieves the natural cubic smoothing spline. We then present examples on the sphere S2 and the special orthogonal group SO(3). We compare our results with the fast algorithm of Arnould et al. [37], generalized to fitting, and the so-called blended cubic splines from Gousenbourger et al. [38]. The control points of the former are obtained by generalizing the optimal Euclidean conditions of (20) (in ℝm, this is a linear system of equations) to the manifold setting; the curve is afterwards reconstructed by a classical De Casteljau algorithm. In the latter, the curve is obtained as a blending of solutions computed on carefully chosen tangent spaces, i.e., Euclidean spaces.

We will show in this section that the proposed method performs as good as the existing methods when the data is from the Eulidean space (section 5.1). This also means that all methods work equally well whenever the data points on the manifold are local enough. However, we will show by the other examples (sections 5.2 and 5.3) that our proposed method outperforms the others whenever the data points are spread out on the manifold.

The following examples were implemented in MVIRT [12]2, and the comparison implementations from Arnould et al. [37] and Gousenbourger et al. [38] use Manopt [11]3. Note that both toolboxes use a very similar matrix structure and are implemented in Matlab, such that the results can directly be compared.

5.1. Validation on the Euclidean Space

As a first example, we perform a minimization on the Euclidean space M=3. A classical result on ℝm is that the curve γ minimizing (1) is the natural (C2) cubic spline when Γ is a Sobolev space H2(t0,tn). We compare our approach to the tangential linear system approach derived in Arnould et al. [37] in order to validate our model. We use the following data points

d0=[001],  d1=[010],  d2=[100],  d3=182[019],    (25)

as well as the control points pi = di, and

b0+=expp0π82[110],    b1+=expp1π22[101],                                                  b1=g(2;b1+,p1),b2+=expp2π22[011],    b3=expp3π8[100],                                                  b2=g(2;b2+,p2),    (26)

where the exponential map and geodesic on ℝ3 are actually the addition and line segments, respectively. Note that, by construction, the initial curve is continuously differentiable but (obviously) does not minimize (1). The parameter λ is set to 50. The MSA of this initial curve Ã(b) is approximately 18.8828.

The data points are given to the algorithm of Arnould et al. [37] to reconstruct the optimal Bézier curve B(t), which is the natural C2 cubic spline. The result is shown in Figure 7A together with the first order differences along the curve in Figure 7B as a dotted curve.


Figure 7. The initial interpolating Bézier curve in ℝ3 (dashed, A) with an MSA of 18.8828 is optimized by the linear system method (LS) from [37] (dotted) and by the proposed gradient descent (GD, solid blue). As expected, both curve coincide, with an MSA of 4.981218. The correspondance is further illustrated with their first order differences in (B).

The set of control points (p0,b0+,,b3-,p3) is optimized with our proposed method. We discretize the second order difference using N = 1600 points. The resulting curve and the first order difference plots are also obtained using these sampling values and a first order forward difference. The gradient descent algorithm from Algorithm 1 employs the Armijo rule (24) setting β=12, σ = 10−4, and α = 1. The stopping criteria are i=11600dM(xi(k),xi(k+1))<ϵ=10-15 or MnÃ2<10-9, and the algorithm stops when one of the two is met. For this example, the first criterion stopped the algorithm, while the norm of the gradient was of magnitude 10−6.

Both methods improve the initial functional value of Ã(b)≈18.8828 to a value of Ã(bmin)≈4.981218. Both the linear system approach and the gradient descent perform equally. The difference of objective value is 2.4524 × 10−11 smaller for the gradient descent, and the maximal distance of any sampling point of the resulting curves is of size 4.3 × 10−7. Hence, in the Euclidean space, the proposed gradient descent yields the natural cubic spline, as one would expect.

5.2. Examples on the Sphere S2

Validation on a geodesic. As a second example with a known minimizer, we consider the manifold M=S2, i.e., the two-dimensional unit sphere embedded in ℝ3, where geodesics are great arcs. We use the data points

d0=[001],    d1=[010],    d2=[001]

aligned on the geodesic connecting the north pole p0 and the south pole p2, and running through a point p1 on the equator. We define the control points of the cubic Bézier curve as follows:

x0=16[112],    b0+=expp0(3logp0(x0)),                                    b1=16[121],x2=16[112],b1+=16[121],                                    b2=expp2(13logp0(x2)),

where x0 and x2 are temporary points and pi = di. We obtain two segments smoothly connected since logp1b1-=-logp1b1+. The original curve is shown in Figure 8A, where especially the tangent vectors illustrate the different speed at pi.


Figure 8. The initial curve (dashed, A) results in a geodesic (solid, B) when minimizing the discrete acceleration. This can also be seen on the first order differences (C).

The control points are optimized with our interpolating model, i.e., we fix the start and end points p0, p1, p2 and minimize A(b).

The curve, as well as the second and first order differences, is sampled with N = 201 equispaced points. The parameters of the Armijo rule are again set to β=12, σ = 10−4, and α = 1. The stopping criteria are slightly relaxed to 10−7 for the distance and 10−5 for the gradient, because of the sines and cosines involved in the exponential map.

The result is shown in Figure 8B. Since the minimizer is the geodesic running from p0 to p2 through p1, we measure the perfomance first by looking at the resulting first order difference, which is constant, as can be seen in Figure 8C. As a second validation, we observe that the maximal distance of the resulting curve to the geodesic is of 2.2357 × 10−6. These evaluations again validate the quality of the gradient descent.

Effect of the data term. As a third example we investigate the effect of λ in the fitting model. We consider the data points

d0=[001]T,    d1=[010]T,    d2=[100]T,                                                                                  d3=[001]T,

as well as the control points pi = di, and

b0+=expp0(π82[110]T),     ​b1+=expp1(π42[101]T),b2+=expp2(π42[011]T),​     b3=expp3(π82[110]T).

The remaining control points b1- and b2- are given by the C1 conditions (4). The corresponding curve B(t), t∈[0, 3], is shown in Figure 9 in dashed black. When computing a minimal MSA curve that interpolates B(t) at the points pi, i = 0, …, 3, the acceleration is not that much reduced, see the blue curve in Figure 9A.


Figure 9. The composite Bézier curves are composed of three segments. The initial curve (dashed, A) is optimized with the interpolating model (solid blue, A) as well as with the fitting model for a continuum of values of λ, from λ = 10 (violet, B) to λ = 0.01 (yellow, B). The limit case where λ = 0 yields an unconstrained geodesic (solid black, A).

For fitting, we consider different values of λ and the same parameters as for the last example. The optimized curve fits the data points closer and closer as λ grows, and the limit λ → ∞ yields the interpolation case. On the other hand smaller values of λ yield less fitting, but also a smaller value of the mean squared acceleration. In the limit case, i.e., λ = 0, the curve (more precisely the control points of the Bézier curve) just follows the gradient flow to a geodesic.

The results are collected in Figure 9. In Figure 9A, the original curve (dashed) is shown together with the solution of the interpolating model (solid blue) and the gradient flow result, i.e., the solution of the fitting model (solid black) with λ = 0. Figure 9B illustrates the effect of λ even further with a continuous variation of λ from a large value of λ = 10 to a small value of λ = 0.01. The image is generated by sampling this range with 1000 equidistant values colored in the colormap viridis. Furthermore, the control points are also shown in the same color.

Several corresponding functional values, see Table 1, further illustrate that with smaller values of λ the discretized MSA also reduces more and more to a geodesic. For λ = 0 there is no coupling to the data points and hence the algorithm does not restrict the position of the geodesic. In other words, any choice for the control points that yields a geodesic, is a solution. Note that the gradient flow still chooses a reasonably near geodesic to the initial data (Figure 9A, solid black).


Table 1. Functional values for the three-segment composite Bézier curve on the sphere and different values of λ.

Comparison with the tangential solution. In the Euclidean space, the method introduced in Arnould et al. [37] and Gousenbourger et al. [38] and the gradient descent proposed here yield the same curve, i.e., the natural cubic spline. On Riemannian manifolds, however, all approaches approximate the optimal solution. Indeed, their method provides a solution by working in different tangent spaces, and ours minimizes a discretization of the objective functional.

In this example we take the same data points di as in (25) now interpreted as points on M=S2. Note that the control points constructed in (26) still fit to the sphere since each bi± is built with a vector in TpiM and using the corresponding exponential map. The result is shown in Figure 10A. The norm of the first order differences is given in Figure 10B. The initial curve (dashed black) has an objective value of 10.9103. The tangent version (dotted) reduces this value to 7.3293. The proposed method (solid blue) yields a value of 2.7908, regardless of whether the starting curve is the initial one, or the one already computed by Arnould et al. [37]. Note that, when p3=[0,0,-1]T, the tangent version is even not able to compute any result, since the construction is performed in the tangent space of p0 and since logp0p3 is not defined.


Figure 10. The initial composite Bézier curve on S2 (dashed, A) has an MSA of 10.9103. The curve obtained with the linear system (LS, dotted) of [37] has an MSA of 7.3293. The solution returned by our proposed method (GD, solid blue) outperforms them all with an MSA of 2.7908. This is further illustrated by the first order derivative approximated by first order differences in (B).

5.3. An Example of Orientations

Finally we compare our method with the blended splines introduced in Gousenbourger et al. [38] for orientations, i.e., data given on SO(3). Let

Rxy(α)=[cosαsinα0sinαcosα0001],    Rxz(α)=[cosα0sinα010sinα0cosα],                                                                    Ryz(α)=[1000cosαsinα0sinαcosα]

denote the rotation matrices in the xy, xz, and yz planes, respectively. We introduce the three data points


These data points are shown in the first line of Figure 11 in cyan.


Figure 11. From top row to bottom row: (1) the initial data points (cyan), (2) the control points computed by the blended Bézier approach from Gousenbourger et al. [38], (3) 17 points on the corresponding curve, where the height represents the absolute first order difference, and the comparing curve is the first order differences from our model, (4) 17 points along the resulting curve from gradient descent, and (5) the resulting control points of the curve in (4).

We set λ = 10 and discretize (20) with N = 401 equispaced points. This sampling is also used to generate the first order finite differences. The parameters of the gradient descent algorithm are set to the same values as for the examples on the sphere.

We perform the blended spline fitting with two segments and cubic splines. The resulting control points are shown in the second row of Figure 11, in green. The objective value is Ã(b) = 0.6464. We improve this solution with the minimization problem (22) on the product manifold SO(3)6. We obtain the control points shown in the last line of Figure 11 and an objective value of Ã(b^)=0.2909 for the resulting minimizer b^.

We further compare both curves by looking at their absolute first order differences. In the third line of Figure 11, we display 17 orientations along the initial curve with its first order difference as height. The line without samples is the absolute first order differences of the minimizer b^, scaled to the same magnitude. The line is straightened especially for the first Bézier segment. Nevertheless, the line is still bent a little bit and hence not a geodesic. This can be seen in the fourth line which represents 17 samples of the minimizing curve compared to its control points in the last line, which are drawn on a straight line.

6. Conclusion and Future Work

In this paper, we introduced a method to solve the curve fitting problem to data points on a manifold, using composite Bézier curves. We approximate the mean squared acceleration of the curve by a suitable second order difference and a trapezoidal rule, and derive the corresponding gradient with respect to its control points. The gradient is computed in closed form by exploiting the recursive structure of the De Casteljau algorithm. Therefore, we obtain a formula that reduces to a concatenation of adjoint Jacobi fields.

The evaluation of Jacobi fields is the only additional requirement compared to previous methods, which are solely evaluating exponential and logarithmic maps. For these, closed forms are available on symmetric manifolds.

On the Euclidean space our solution reduces to the natural smoothing spline, the unique acceleration minimizing polynomial curve. The numerical experiments further confirm that the method presented in this paper outperforms the tangent space(s)-based approaches with respect to the functional value.

It is still an open question whether there exists a second order absolute finite difference on a manifold, that is jointly convex. Then convergence would follow by standard arguments of the gradient descent. For such a model, another interesting point for future work is to find out whether only an approximate evaluation of the Jacobi fields suffices for convergence. This would mean that, on manifolds where the Jacobi field can only be evaluated approximately by solving an ODE, the presented approach would still converge.

Author Contributions

Both authors listed have contributed equally to the content of this work and approved the paper for publication.

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.


This work was supported by the following fundings: RB gratefully acknowledges support by DFG grant BE 5888/2–1; P-YG gratefully acknowledges support by the Fonds de la Recherche Scientifique – FNRS and the Fonds Wetenschappelijk Onderzoek – Vlaanderen under EOS Project no 30468160, and Communauté française de Belgique - Actions de Recherche Concertées (contract ARC 14/19-060). Figure 9A is based on an idea of Rozan I. Rosandi.


1. ^See

2. ^Open source, available at

3. ^Open source, available at


1. Pyta L, Abel D. Interpolatory Galerkin models for Navier-Stokes-equations. IFAC-PapersOnLine (2016) 49:204–9. doi: 10.1016/j.ifacol.2016.07.442

CrossRef Full Text | Google Scholar

2. Gousenbourger PY, Massart E, Musolas A, Absil PA, Jacques L, Hendrickx JM, et al. Piecewise-Bézier C1 smoothing on manifolds with application to wind field estimation. In: ESANN2017. Bruges: Springer (2017). Available online at:

Google Scholar

3. Sander O. Geodesic finite elements for cosserat rods. Int J Numer Methods Eng. (2010) 82:1645–70. doi: 10.1002/nme.2814

CrossRef Full Text | Google Scholar

4. Park J. Interpolation and tracking of rigid body orientations. In: ICCAS, Gyeonggi-do (2010). p. 668–73. doi: 10.1109/ICCAS.2010.5670237

CrossRef Full Text | Google Scholar

5. Su J, Kurtek S, Klassen E, Srivastava A. Statistical analysis of trajectories on Riemannian manifolds: bird migration, Hurricane tracking and video surveillance. Ann Appl Stat. (2014) 8:530–52. doi: 10.1214/13-AOAS701

CrossRef Full Text | Google Scholar

6. Dyn N. Linear and nonlinear subdivision schemes in geometric modeling. In: Cucker F, Pinkus A, Todd MJ, editors. FoCum. Vol. 363 of London Math. Soc. Lecture Note Ser. Hong Kong: Cambridge University Press (2009). p. 68–92. doi: 10.1017/CBO9781139107068.004

CrossRef Full Text | Google Scholar

7. Wallner J, Nava Yazdani E, Grohs P. Smoothness properties of Lie group subdivision schemes. Multiscale Model Simulat. (2007) 6:493–505. doi: 10.1137/060668353

CrossRef Full Text | Google Scholar

8. Shingel T. Interpolation in special orthogonal groups. IMA J Numer Anal. (2008) 29:731–45. doi: 10.1093/imanum/drn033

CrossRef Full Text | Google Scholar

9. Green PJ, Silverman BW. Nonparametric Regression and Generalized Linear Models: A Roughness Penalty Approach. CRC Press (1993). Available online at:

Google Scholar

10. Absil PA, Mahony R, Sepulchre R. Optimization Algorithms on Matrix Manifolds. Princeton, NJ: Princeton University Press (2008). Available online at:

Google Scholar

11. Boumal N, Mishra B, Absil PA, Sepulchre R. Manopt, a matlab toolbox for optimization on manifolds. J Mach Learn Res. (2014) 15:1455–9. Available online at:

Google Scholar

12. Bergmann R. MVIRT a toolbox for manifold-valued image registration. In: IEEE International Conference on Image Processing, IEEE ICIP 2017, Beijing. (2017). Available online at:

Google Scholar

13. Zhou X, Yang C, Zhao H, Yu W. Low-rank modeling and its applications in image analysis. ACM Comput Surveys (2015) 47:36. doi: 10.1145/2674559

CrossRef Full Text | Google Scholar

14. Cunningham JP, Ghahramani Z. Linear dimensionality reduction: survey, insights, and generalizations. J Mach Learn Res. (2015) 16:2859–900. Available online at:

Google Scholar

15. Sun J, Qu Q, Wright J. A geometric analysis of phase retrieval. Found Comput Math. (2017) 8:1131–98. doi: 10.1007/s10208-017-9365-9

CrossRef Full Text | Google Scholar

16. Yu X, Shen JC, Zhang J, Letaief KB. Alternating minimization algorithms for hybrid precoding in millimeter wave MIMO systems. IEEE J Selected Top Signal Process. (2016) 10:485–500. doi: 10.1109/JSTSP.2016.2523903

CrossRef Full Text | Google Scholar

17. Strekalovskiy E, Cremers D. Total variation for cyclic structures: convex relaxation and efficient minimization. In: IEEE Conference on Computer Vision and Pattern Recognition (2011). p. 1905–11.

Google Scholar

18. Strekalovskiy E, Cremers D. Total cyclic variation and generalizations. J Mat Imaging Vis. (2013) 47:258–77. doi: 10.1007/s10851-012-0396-1

CrossRef Full Text | Google Scholar

19. Lellmann J, Strekalovskiy E, Koetter S, Cremers D. Total variation regularization for functions with values in a manifold. In: IEEE International Conference on Computer Vision. Sydney, NSW (2013). p. 2944–51. doi: 10.1109/ICCV.2013.366

CrossRef Full Text | Google Scholar

20. Weinmann A, Demaret L, Storath M. Total variation regularization for manifold-valued data. SIAM J Imaging Sci. (2014) 7:2226–57. doi: 10.1137/130951075

CrossRef Full Text | Google Scholar

21. Bergmann R, Laus F, Steidl G, Weinmann A. Second order differences of cyclic data and applications in variational denoising. SIAM J Imaging Sci. (2014) 7:2916–53. doi: 10.1137/140969993

CrossRef Full Text | Google Scholar

22. Bačák M, Bergmann R, Steidl G, Weinmann A. A second order nonsmooth variational model for restoring manifold-valued images. SIAM J Comput. (2016) 38:567–97. doi: 10.1137/15M101988X

CrossRef Full Text | Google Scholar

23. Bergmann R, Fitschen JH, Persch J, Steidl G. Priors with coupled first and second order differences for manifold-valued image processing. J Math Imaging Vis. (2017) 60:1459–81. doi: 10.1007/s10851-018-0840-y

CrossRef Full Text | Google Scholar

24. Bredies K, Holler M, Storath M, Weinmann A. Total generalized variation for manifold-valued data. SIAM J Imaging Sci. (2018) 11:1785–48. doi: 10.1137/17M1147597

CrossRef Full Text | Google Scholar

25. Bergmann R, Persch J, Steidl G. A parallel Douglas–Rachford algorithm for restoring images with values in symmetric Hadamard manifolds. SIAM J Imaging Sci. (2016) 9:901–37. doi: 10.1137/15M1052858

CrossRef Full Text | Google Scholar

26. Grohs P, Sprecher M. Total variation regularization on Riemannian manifolds by iteratively reweighted minimization. Inform Inference (2016) 5:353–78. doi: 10.1093/imaiai/iaw011

CrossRef Full Text | Google Scholar

27. Bergmann R, Chan RH, Hielscher R, Persch J, Steidl G. Restoration of manifold-valued images by half-quadratic minimization. Inverse Prob Imaging (2016) 10:281–304. doi: 10.3934/ipi.2016001

CrossRef Full Text | Google Scholar

28. Samir C, Absil PA, Srivastava A, Klassen E. A gradient-descent method for curve fitting on Riemannian manifolds. Found Comput Math. (2012) 12:49–73. doi: 10.1007/s10208-011-9091-7

CrossRef Full Text | Google Scholar

29. Su J, Dryden IL, Klassen E, Le H, Srivastava A. Fitting smoothing splines to time-indexed, noisy points on nonlinear manifolds. Image Vis Comput. (2012) 30:428–42. doi: 10.1016/j.imavis.2011.09.006

CrossRef Full Text | Google Scholar

30. Boumal N, Absil PA. A discrete regression method on manifolds and its application to data on SO(n). In: IFAC Proceedings Volumes (IFAC-PapersOnline). Vol. 18. Milano (2011). p. 2284–89. doi: 10.3182/20110828-6-IT-1002.00542

CrossRef Full Text | Google Scholar

31. Kim KR, Dryden IL, Le H. Smoothing splines on Riemannian manifolds, with applications to 3D shape space. (2018). arXiv[Preprint]:1801.04978.

Google Scholar

32. Fletcher PT. Geodesic regression and the theory of least squares on Riemannian manifolds. Int J Comput Vis. (2013) 105:171–85. doi: 10.1007/s11263-012-0591-y

CrossRef Full Text | Google Scholar

33. Rentmeesters Q. A gradient method for geodesic data fitting on some symmetric Riemannian manifolds. In: Decision and Control and European Control Conference (CDC-ECC), 2011 50th IEEE Conference on. Orlando, FL (2011). p. 7141–6. doi: 10.1109/CDC.2011.6161280

CrossRef Full Text | Google Scholar

34. Farin G. Curves Surfaces for CAGD. Morgan Kaufmann (2002). Available online at:

35. Absil PA, Gousenbourger PY, Striewski P, Wirth B. Differentiable Piecewise-Bézier Surfaces on Riemannian Manifolds. SIAM J Imaging Sci. (2016) 9:1788–828. doi: 10.1137/16M1057978

CrossRef Full Text | Google Scholar

36. Popiel T, Noakes L. Bézier curves and C2 interpolation in Riemannian manifolds. J Approx Theory (2007) 148:111–27. doi: 10.1016/j.jat.2007.03.002

CrossRef Full Text | Google Scholar

37. Arnould A, Gousenbourger PY, Samir C, Absil PA, Canis M. Fitting smooth paths on Riemannian manifolds: endometrial surface reconstruction and preoperative MRI-based navigation. In: Nielsen F, Barbaresco F, editors. GSI2015. Springer International Publishing (2015). p. 491–8.

Google Scholar

38. Gousenbourger PY, Massart E, Absil PA. Data fitting on manifolds with composite Bézier-like curves and blended cubic splines. J Math Imaging Vis. (2018). doi: 10.1007/s10851-018-0865-2

CrossRef Full Text | Google Scholar

39. O'Neill B. Elementary Differential Geometry. London: Academic Press Inc. (1966).

Google Scholar

40. do Carmo MP. Riemannian Geometry. Vol. 115. Basel: Birkhäuser; 1992. Tranl. by F. Flatherty.

Google Scholar

41. Karcher H. Riemannian center of mass and mollifier smoothing. Commun Pure Appl Math. (1977) 30:509–41. doi: 10.1002/cpa.3160300502

CrossRef Full Text | Google Scholar

42. Boumal N. Interpolation and regression of rotation matrices. In: Nielsen F, Barbaresco F, editors. Geometric Science of Information. Berlin; Heidelberg: Springer. (2013). p. 345–52.

Google Scholar

43. Persch J. Optimization Methods in Manifold-Valued Image Processing [Dissertation]. Technische Universität Kaiserslautern (2018).

Google Scholar

Keywords: Riemannian manifolds, curve fitting, composite Bézier curves, Jacobi fields, variational models

Citation: Bergmann R and Gousenbourger P-Y (2018) A Variational Model for Data Fitting on Manifolds by Minimizing the Acceleration of a Bézier Curve. Front. Appl. Math. Stat. 4:59. doi: 10.3389/fams.2018.00059

Received: 26 July 2018; Accepted: 13 November 2018;
Published: 12 December 2018.

Edited by:

Luca Marchetti, Microsoft Research-University of Trento Centre for Computational and Systems Biology(COSBI), Italy

Reviewed by:

Giovanni Mascali, Università della Calabria, Italy
Vincenzo Bonnici, Università degli Studi di Verona, Italy

Copyright © 2018 Bergmann and Gousenbourger. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Ronny Bergmann,