ORIGINAL RESEARCH article

Front. Appl. Math. Stat., 12 December 2018

Sec. Optimization

Volume 4 - 2018 | https://doi.org/10.3389/fams.2018.00059

A Variational Model for Data Fitting on Manifolds by Minimizing the Acceleration of a Bézier Curve

  • 1. Research Group Numerical Mathematics (Partial Differential Equations), Faculty of Mathematics, Technische Universität Chemnitz, Chemnitz, Germany

  • 2. ICTEAM Institute, Université Catholique de Louvain, Louvain-la-Neuve, Belgium

Article metrics

View details

27

Citations

4,4k

Views

1k

Downloads

Abstract

We derive a variational model to fit a composite Bézier curve to a set of data points on a Riemannian manifold. The resulting curve is obtained in such a way that its mean squared acceleration is minimal in addition to remaining close the data points. We approximate the acceleration by discretizing the squared second order derivative along the curve. We derive a closed-form, numerically stable and efficient algorithm to compute the gradient of a Bézier curve on manifolds with respect to its control points, expressed as a concatenation of so-called adjoint Jacobi fields. Several examples illustrate the capabilities and validity of this approach both for interpolation and approximation. The examples also illustrate that the approach outperforms previous works tackling this problem.

AMS subject classification (2010). 65K10, 65D10, 65D25, 53C22, 49Q99.

1. Introduction

This papers addresses the problem of fitting a smooth curve to data points d0, …, dn lying on a Riemannian manifold and associated with real-valued parameters t0, …, tn. The curve strikes a balance between a data proximity constraint and a smoothing regularization constraint.

Several applications motivate this problem in engineering and the sciences. For instance, curve fitting is of high interest in projection-based model order reduction of one-dimensional dynamical systems [1]. In that application, the dynamical system depends on the Reynolds number and the model reduction is obtained by computing suitable projectors as points on a Grassmann manifold. Finding a projector is however a time- and memory-consuming task. Based on projectors precomputed for given parameter values, fitting is used to approximate the projector associated with a new parameter value. In Gousenbourger et al. [2], the same strategy is used to approximate wind field orientations represented as points on the manifold of positive semidefinite covariance matrices of size p and rank r. Further applications are involving rigid-body motions on SE(3), like Cosserat rods applications [3], or orientation tracking [4]. Sphere-valued data are also of interest in many applications of data analysis, for storm tracking and prediction, or the study of bird migration [5].

There exists different approaches to tackle the curve fitting problem. Among others, we name here the subdivision schemes approach [6, 7] or the Lie-algebraic methods [8]. However, the most popular approach nowadays is probably to encapsulate the two above-mentioned constraints into an optimization problem

where Γ is an admissible space of curves , t↦γ(t), denotes the (Levi-Civita) second covariant derivative, and ∥·∥γ(t) denotes the Riemannian metric at γ(t) from which we can define a Riemannian distance d(·, ·). Finally, λ∈ℝ is a parameter that strikes the balance between the regularizer and the fitting term. This approach leads to the remarkable property that, when the manifold reduces to the Euclidean space and Γ is the Sobolev space , the solution to (1) is the natural cubic spline [9].

Optimization on manifolds has gained a lot of interest this last decade, starting with the textbook [10] that summarizes several optimization methods on matrix manifolds. Recently, toolboxes have emerged, providing easy access to such optimization methods, e.g., Manopt [11] and MVIRT [12]. The former received a very positive return in many different topics of research, with applications for example in low-rank modeling in image analysis [13], dimensionality reduction [14], phase retrieval [15] or even 5G-like MIMO systems [16]. The latter stems from recent interest in manifold-valued image and data processing, phrased as variational models on the product manifold , where N is the number of pixels. Starting with total variation (TV) regularization of phase-valued data [17, 18], different methods for TV on manifolds [19, 20] have been developed as well as second order methods [21, 22] up to infimal convolution [23] and total generalized variation [23, 24]. Furthermore, different algorithms have been generalized to manifold-valued data, besides the previous works using gradient descent or cyclic proximal point methods, a Douglas–Rachford splitting [25], iteratively reweighted least squares [26] and more general half-quadratic minimization [27] have been introduced.

The curve fitting problem (1), has been tackled differently the past few years. Samir et al. [28] considered the case where Γ is an infinite dimensional Sobolev space of curves, and used the Palais-metric to design a gradient descent algorithm for (1) (see e.g., [29] for an application of this approach). Another method consists in discretizing the curve γ in N points, and therefore considering (see e.g., [30] for a result on SO(3)). Finally, the limit case where λ → 0 is already well studied and known as the geodesic regression [3133].

A recent topic concerns curve fitting by means of Bézier curves. In that approach, the search space Γ is reduced to as set of composite Bézier curves. Those are a very versatile tool to model smooth curves and surfaces for real- and vector-valued discrete data points (see [34] for a comprehensive textbook), but they can also be used to model smooth curves and surfaces for manifold-valued data [35, 36]. The advantage to work with such objects, compared to classical approaches, are that (i) the search space is drastically reduced to the so-called control points of the Bézier curves (and this leads to better time and memory performances) and (ii) it is very simple to impose differentiability for the optimal curve, which is appreciated in several of the above-mentioned applications. However, while obtaining such an optimal curve reduces directly to solving a linear system of equations for data given on a Euclidean space, there is up to now no known closed form of the optimal Bézier curve for manifold valued data.

In this work, we derive a gradient descent algorithm to compute a differentiable composite Bézier curve that satisfies (1), i.e., such that B(t) has a minimal mean squared acceleration, and fits the set of n+1 manifold-valued data points at their associated time-parameters. We consider the manifold-valued generalization of Bézier curves [36] in the same setting as in Arnould et al. [37], or more recently in Gousenbourger et al. [38]. We employ the (squared) second order absolute difference introduced in Bačák et al. [22] to obtain a discrete approximation of the regularizer from (1). The quality of the approximation depends only on the number of sampling points. We exploit the recursive structure of the De Casteljau algorithm [36] to derive the gradient of the objective function with respect to the control points of B(t). The gradient is built as a recursion of Jacobi fields that, for numerical reasons, are implemented as a concatenation of so-called adjoint Jacobi fields. Furthermore, the corresponding variational model only depends on the number of control points of the composite Bézier curve, and not on the number of sampling points. We finally obtain an approximating model to (1) that we solve with an algorithm only based on three tools on the manifold: the exponential map, the logarithmic map, and a certain Jacobi field along geodesics.

The paper is organized as follows. We introduce the necessary preliminaries—Bézier curves, Riemannian manifolds and Riemannian second order finite differences— in section 2. In section 3 we derive the gradient of the discretized mean squared acceleration of the composite Bézier curve with respect to its control points, and thus of the regularizer of (1). In section 4, we present the corresponding gradient descent algorithm, as well as an efficient gradient evaluation method, to solve (1) for different values of λ. The limit case where λ → ∞ is studied as well. Finally, in section 5, we validate, analyze and illustrate the performance of the algorithm for several numerical examples on the sphere S2 and on the special orthogonal group SO(3). We also compare our solution to existing Bézier fitting methods. A conclusion is given in section 6.

2. Preliminaries

2.1. Bézier Functions and Composite Bézier Spline

Consider the Euclidean space ℝm. A Bézier curve of degree K∈ℕ is a function parametrized by control points and taking the form

where are the Bernstein basis polynomials [34]. For example the linear Bézier curve β1 is just the line segment (1−t)b0+tb1 connecting the two control points b0 and b1. The explicit formulae of the quadratic and cubic Bézier curves read

for given control points and an additional point for the cubic case.

A composite Bézier curve is a function B:[0, n] → ℝm composed of n Bézier curves and defined as

where Ki∈ℕ denotes the degree of the ith Bézier curve βKi of B and , j = 0, …, Ki, are its control points. Furthermore, B(t) is continuous if the last and first control points of two consecutive segments coincide [34]. We introduce pi as the point at the junction of two consecutive Bézier segments, i.e., . Differentiability is obtained if the control points of two consecutive segments are aligned. We introduce further and such that the differentiability condition reads .

Example 1 (C1 conditions for composite cubic Bézier curves). We consider the case where the Bézier segments are all cubic, as represented in Figure1. The composite Bézier curveB(t) is C1if, i = 1, …, 4, with, , and.

Figure 1

Figure 1

Schematic representation of the composite cubic Bézier curve , for (black), and its control points (green circles). Continuous differentiability is reached at the junction of the segments (the blue arrows draw the first derivative of B).

2.2. Riemannian Manifolds

We consider a complete m-dimensional Riemannian manifold . We refer to O'Neill [39] and do Carmo [40] for an introduction to Riemannian manifolds and to [10] for optimization thereon. We will use the following terminology and notations.

We denote by the (Euclidean) tangent space to at ; is the tangent bundle to ; 〈·, ·〉a denotes the inner product in the tangent space at a and from which we deduce the norm of denoted by . For a (not necessarily unique) shortest geodesic between a and , we write , tg(t; a, b), parametrized such that g(0;a, b) = a and g(1;a, b) = b. This choice of parametrization also means that the covariant derivative of g with respect to time satisfies , for all t∈[0, 1], where is the geodesic distance between a and . The Riemannian exponential reads and we denote by ra∈ℝ the maximal radius such that the exponential map is bijective on . Then is called the Riemannian logarithm which is (locally) the inverse of the exponential. A Riemannian manifold is called symmetric in if the geodesic reflection sx at given by the mapping γ(t)↦γ(−t) is an isometry at least locally near x, for all geodesics through γ(0) = x. If is symmetric in every , the manifold is called (Riemannian) symmetric space or symmetric manifold.

In the following we assume, that both the exponential and the logarithmic map are available for the manifold and that they are computationally not too expensive to evaluate. Furthermore, we assume that the manifold is symmetric.

2.3. Composite Bézier Curves on Manifolds

One well-known way to generalize Bézier curves to a Riemannian manifold is via the De Casteljau algorithm [36, section 2]. This algorithm only requires the Riemannian exponential and logarithm and conserves the interpolation property of the first and last control points. Some examples on interpolation and fitting with Bézier curves or Bézier surfaces, i.e., generalizations of tensor product Bézier curves, can be found in Absil et al. [35].

Consider , the manifold-valued Bézier curve of order K driven by K+1 control points . We introduce the points and iterate the construction of further points. For i = 0, …, Kk, k = 1, …, K, we define

as the ith point of the kth step of the De Casteljau algorithm, and obtain .

The De Casteljau algorithm is illustrated on Figure 2 for a Euclidean cubic Bézier curve β3(t; b0, b1, b2, b3). The general cubic Bézier curve can be explicitly expressed on a manifold as

Figure 2

Figure 2

Construction of a cubic Bézier curve via the De Casteljau algorithm.

The conditions of continuity and differentiability are generalized to manifolds in Popiel and Noakes [36].

Lemma 2 (Differentiability conditions [36]). Consider the composite Bézier curveconsisting ofand, i.e.,

The composite Bézier curveB(t) is continuous and continuously differentiable if the two following conditions hold:

The composite Bézier curve is then defined completely analogously to the Euclidean case from Equation (2). The differentiability conditions (4) of Lemma 2 have to be satisfied at each junction point pi, i = 1, …, n−1.

Example 3 (Composite cubic Bézier curves). The composite cubic Bézier curveis C1if. See Figure1for an example onwith n = 5 segments and Figure3for an example onwith n = 3 segments.

Figure 3

Figure 3

A composite cubic Bézier curve B:[0, 3] → S2. The end points pi, i = 0, …, 3, (cyan) and intermediate points (green) determine its shape; continuous differentiability is illustrated by the logarithmic maps , i∈{1, 2} (blue arrows).

2.4. Discrete Approximation of the Mean Squared Acceleration

We discretize the mean squared acceleration (MSA) of a curve , by discretizing the corresponding integral

i.e., the regularizer from (1). We approximate the squared norm of the second (covariant) derivative by the second order absolute finite difference introduced by Bačák et al. [22]. Consider three points and the set of mid-points of x and z

for all (not necessarily shortest) geodesics g connecting x and z. The manifold-valued second order absolute finite difference is defined by

This definition is equivalent, on the Euclidean space, to .

Using equispaced points t0, …, tN, N∈ℕ, with step size , we approximate , i = 1, …, N−1, and obtain by the trapezoidal rule

For Bézier curves γ(t) = B(t) we obtain for the regularizer in (1) the discretized MSA A(b) that depends on the control points b and reads

3. The Gradient of the Discretized Mean Squared Acceleration

In order to minimize the discretized MSA A(b), we aim to employ a gradient descent algorithm on the product manifold , where M is the number of elements in b. In the following, we derive a closed form of the gradient ∇bA(b) of the discretized MSA (7). This gradient is obtained by means of a recursion and the chain rule. In fact, the derivative of (6) is already known [22], such that it only remains to compute the derivative of the composite Bézier curve.

We first introduce the following notation. We denote by the directional derivative of evaluated at x0, with respect to its argument x and in the direction . We use the short hand Dxf[η] = Dxf[η](x) whenever this directional derivative is evaluated afterwards again at x.

We now state the two following definitions, which are crucial for the rest of this section.

Definition 4 (Gradient [10, Equation (3.31), p. 46]). Letbe a real-valued function on a manifold, and.

The gradientof f at x is defined as the tangent vector that fulfills

For multivariate functions f(x, y), we denote the gradient of f with respect to x at (x0, y0) by writing . We shorten this notation as when this gradient is seen as a function of x (and y).

Definition 5 (Chain rule on manifolds [10, p. 195]). Let, be two functions on a manifoldand, xF(x) = (f ° h)(x) = f(h(x)), their composition. Letand. The directional derivative DxF[η] of F with respect to x in the direction η is given by

whereand.

The remainder of this section is organized in four parts. We first recall the theory on Jacobi fields in section 3.1 and their relation to the differential of geodesics (with respect to start and end point). In section 3.2, we apply the chain rule to the composition of two geodesics, which appears within the De Casteljau algorithm. We use this result to build an algorithmic derivation of the differential of a general Bézier curve on manifolds with respect to its control points (section 3.3). We extend the result to composite Bézier curves in section 3.4, including their constraints on junction points pi to enforce the C1 condition (4), and finally gather these results to state the gradient of the discretized MSA (7) with respect to the control points.

3.1. Jacobi Fields as Derivative of a Geodesic

In the following, we introduce a closed form of the differential Dxg(t; ·, y) of a geodesic g(t; x, y), t∈[0, 1], with respect to its start point . The differential with respect to the end point can be obtained by taking the geodesic g(t, y, x) = g(1−t; x, y).

As represented in Figure 4, we denote by γx, ξ, the geodesic starting in γx, ξ(0) = x and with direction . We introduce ζ(s): = logγx, ξ(s)y, the tangential vector in pointing toward y. Then, the geodesic variation Γg, ξ(s, t) of the geodesic g(·;x, y) with respect to the tangential direction is given by

where ε>0. The corresponding Jacobi field Jg, ξ along g is then given by the vector field

that represents the direction of the displacement of g if x is perturbed in a direction ξ.

Figure 4

Figure 4

Schematic representation of the variation Γg, ξ(s, t) of a geodesic g w.r.t. the direction . The corresponding Jacobi field along g and in the direction ξ is the vector field .

We directly obtain Jg, ξ(0) = ξ, and Jg, ξ(1) = 0 as well as . Since Γg, ξ(s, t) = g(t; γx, ξ(s), y) we obtain by the chain rule

Remark. This relation between the derivative of geodesics and Jacobi fields is of high interest on symmetric spaces, where Jacobi fields can be computed in closed form, as summarized in the following Lemma.

Lemma 6. [22, Prop. 3.5] Let be a m-dimensional symmetric Riemannian manifold. Let g(t; x, y), t ∈ [0, 1], be a geodesic between , a tangent vector and {ξ1, …ξm} be an orthonormal basis (ONB) of that diagonalizes the curvature operator of with eigenvalues κ, ℓ = 1, …, m. For details, see Ch. 4.2 and 5 (Ex. 5) of [40]. Let further denote by {Ξ1(t), …, Ξm(t)} the parallel transported frame of {ξ1, …, ξm} along g. Decomposing , the derivative Dxg[η] becomes

where the Jacobi fieldalong g and in the direction ξis given by

with dg = d(x, y) denoting the length of the geodesic g(t; x, y), t∈[0, 1].

The Jacobi field of the reversed geodesic ḡ(t): = g(t; y, x) = g(1−t; x, y) is obtained using the same orthonormal basis and transported frame but evaluated at s = 1−t. We thus obtain Dyg(t; x, y)[ξ] = Dyg(1−t; y, x)[ξ] = Jḡ,ξ(1−t), where

Note that . Therefore , ℓ = 1, …, m, is an orthonormal basis for this tangent space.

3.2. Derivative of Coupled Geodesics

Let be a symmetric Riemannian manifold. We use the result of Lemma 6 to directly compute the derivative of coupled geodesics, i.e., a function composed of g1(t): = g(t; x, y) and g2(t): = g(t; g1(t), z). By Definition 5, we have

and by (10), we obtain

where the variation direction used in the Jacobi field is now the derivative of g1(t) in direction η. Similarly, we compute the derivative of a reversed coupled geodesic g3(t): = g(t; z, g1(t)) as

Note that the Jacobi field is reversed here, but that its variation direction is the same as the one of the Jacobi field introduced for g2(t). In a computational perspective, it means that we can use the same ONB for the derivatives of both g3 and ḡ3 Furthermore, in this case, the variation direction is also computed by a Jacobi field since Dxg1(t)[η] = Jg1, η(t).

Finally the derivative of g2 (resp. g3) on symmetric spaces is obtained as follows. Let be an ONB of for the inner Jacobi field along g1, and be an ONB of for the outer Jacobi field along g2 (resp. g3). As , and stating , the derivative of g2 (resp. g3) with respect to x in the direction reads

and accordingly for g3.

3.3. Derivative of a Bézier Curve

Sections 3.1 and 3.2 introduced the necessary concepts to compute the derivative of a general Bézier curve βK(t; b0, …, bK), as described in Equation (3), with respect to its control points bj. For readability of the recursive structure investigated in the following, we introduce a slightly simpler notation and the following setting.

Let K be the degree of a Bézier curve βK(t; b0, …, bK) with the control points . We fix k∈{1, …, K}, i∈{0, …, Kk} and t∈[0, 1]. We introduce

for the ith Bézier curve of degree k in the De Casteljau algorithm, and .

Furthermore, given x∈{bi, …, bi+k}, we denote by

its derivative with respect to one of its control points x in the direction .

Remark. Clearly any other derivative of with respect to x = bj, j<i or j>i+k is zero. In addition we have for x = bi and zero otherwise.

Theorem 7 (Derivative of a Bézier curve). Let k ∈ {1, …, K}and i ∈ {0, …, Kk}be given. The derivativeofwith respect to its control point x: = bj, iji+k, and in the directionis given by

Proof. Let fix t∈[0, 1] and x = bj, iji+k. For readability we set , , and . Note that while f depends on the control points bi, …, bi+k and is a Bézier curve of degree k, both a and b are Bézier curves of degree k−1. The former does not depend on bi+k, and the latter is independent of bi.

We prove the claim by induction. For k = 1 the function is just a geodesic. The case i<j<i+1 does not occur and the remaining first and third cases follow by the notation introduced for k = 0 and Lemma 6.

For k>1 we apply the chain rule (9) to Dxf[η] and obtain

Consider the first term Daf[Dxa[η]] and j<i+k. By (10) and the notation from (14), one directly has

For j = i+k, clearly Dxa[η] = Daf[Dxa[η]] = 0, as a does not depend on bi+k.

We proof the second term similarly. For j>i, by applying the chain rule and using the reversed Jacobi field formulation of Lemma 6 for the derivative of a geodesic with respect to its end point, we obtain

Finally, as Dxb[η] = Dbf[Dxb[η]] = 0 for x = bi, the assumption follows.

      □

Figure 5 represents one level of the schematic propagation tree to compute the derivative of a Bézier curve.

Figure 5

Figure 5

Schematic representation of the cases where elements compose the chained derivative of the ith Bézier curve of order k in the De Casteljau algorithm. The solid line represents a Jacobi field along , while the dashed one represents a reversed Jacobi field. (A) Represents the case x = bi, (B) represents the cases where x∈{bi+1, …, bi+k+1}, and (C) represents the case x = bi+k.

Example 8 (Quadratic Bézier curve). Consider the quadratic Bézier curvedefined as

Using the notations (13), we have

The derivative of β2at t with respect to b0in the direction, is given by

The derivative of β2at t with respect to b2in the directioncan be seen as deriving by the first point after inverting the Bezier curve, i.e., looking at, hence we have analogously to the first term

The case Db1β2[η], , involves a chain rule, where b1appears in both(as its end point) and(as its starting point). Using the two intermediate results (or Jacobi fields of geodesics)

we obtain

Example 9 (Cubic Bézier curve). Consider the cubic Bézier curvedefined as

As in Example 8, we use the notations (13) and define

The derivation of β3with respect to b0or b3follows the same structure as in Example 8. The case of Db1β3[η], however, requires two chain rules. The needed Jacobi fields follow the tree structure shown in Figure6B: given, we define at the first recursion step

and at the second recursion step

Note that bothandare actually the derivative of β2(t; b0, b1, b2) and β2(t; b1, b2, b3), respectively, with respect to b1and direction. Finally we have

Figure 6

Figure 6

Construction and derivation tree of a Bézier curve β3(t; b0, b1, b2, b3). The derivative with respect to a variable bi is obtained by a recursion of Jacobi fields added at each leaf of the tree. (A) Tree-representation of the construction of a cubic Bezier curve. The thick line tracks the propagation of b1 within the tree. (B) Tree-representation of the construction of . The solid lines are Jacobi fields while dashed lines are reversed Jacobi fields.

The case of Db2β3[η] is obtained symmetrically, with arguments similar to b1.

Note that computing the Jacobi fields involved just follows the same decomposition tree as the De Casteljau algorithm (Figure 6A).

3.4. Joining Segments and Deriving the Gradient

In this subsection we derive the differential of a composite Bézier curve B(t) consisting of n segments and take the C1 conditions into account. We simplify the notations from section 2.1 and set the degree fixed to Ki = K for all segments, e.g., K = 3 for a cubic composite Bézier curve. Then the control points are , j = 0, …, K, i = 0, …, n−1. We further denote by , i = 1, …, n−1 the common junction point of the segments and p0 and pn the start and end points, respectively. For ease of notation we denote by and , i = 1, …, n−1, the two points needed for differentiability (C1) condition investigation, cf. Figure 1 for an illustration of the framework on , with K = 3.

One possibility to enforce the C1 condition (4) is to include it into the composite Bézier curve by replacing with

This way both the directional derivatives of B(t) with respect to and pi change due to a further (most inner) chain rule.

Lemma 10 (Derivative of a composite Bézier curve with C1 condition). LetBbe a composite Bézier curve andintroduced as above. Replacingeliminates that variable from the composite Bezier curve and keeps the remaining differentials unchanged, except the following which now read

and

In both cases, the first interval includes i−1 = 0, when i = 1.

Proof. Both first cases are from the derivation of a Bézier curve as before, for both second cases replacing yields one (for pi additional) term.

We now derive the gradient of the objective function (7). We introduce the abbreviation , and .

Theorem 11. Letbe a m-dimensional manifold, x be one of the control points of a composite Bézier curveB, and1, …, ξm} be a corresponding orthonormal basis (ONB) of. The gradientof the discretized mean squared acceleration A(b) ofB, w.r.t. x, discretized at N+1 equispaced times t0, …, tNis given by

Proof: As , we seek for the coefficients a: = a(x) such that

Therefore, for any tangential vector , we have,

By definition of A and Definition 4 this yields

We compute using the chain rule (Definition 5) as

which, by Definition 4, again becomes

The term on the left of the inner product is given in Bačák et al. [22, section 3] and the right term is given in section 3.3. While the former can be computed using Jacobi fields and a logarithmic map, the latter is the iteratively coupling of Jacobi fields. Furthermore, the differential DxBj[η] can be written as

Hence, we obtain

and by (17), (18), and (19), it follows

which yields the assertion (16).

4. Application to the Fitting Problem

The fitting problem has been tackled different ways this last decades. The approach with Bézier curves is more recent, and we refer to Absil et al. [35], Arnould et al. [37], and Gousenbourger et al. [38] for a detailed overview of these methods.

In this section, we present the numerical framework we use in order to fit a composite Bézier curve B to a set of data points associated with time-parameters t0, …, tn, such that we meet (1). For the sake of simplicity, we limit the study to the case where ti = i, i = 0, …, n. Therefore, the fitting problem (1) becomes

where is the set of the M control points of B. Remark that, compared to (1), the optimization is now done on the product manifold . Furthermore, the fitting term now implies a distance between di and pi, as B(ti) = pi.

The section is divided in three parts: the product manifold is defined in section 4.1, where the contribution of the fitting term in the gradient of E is also presented. Then, we propose an efficient algorithm to compute the gradient of the discretized MSA, based on so-called adjoint Jacobi fields. We finally shortly mention the gradient descent algorithm we use as well as the involved Armijo rule.

4.1. Fitting and Interpolation

Let us clarify the set from (20). We will explicitly present the vector b and state its size M. The set ΓB is the set of the M remaining free control points to optimize, when the continuity constraints are imposed. We distinguish two cases: (i) the fitting case, that corresponds to formulae presented in section 3, and (ii) the interpolation case (λ → ∞) where the constraint di = pi is imposed as well.

For a given composite Bézier curve consisting of a Bézier curve of degree K on each segment, and given the C1 conditions (4), the segments are determined by the points

We investigate the length of M. First, is given by pi and via (4). Second, for the segments i = 2, …, n we can further omit pi−1 since the value also occurs in the segment i−1 as last entry. Finally, the first segment contains the additional value of that is not fixed by C1 constraints. The first segment is thus composed of K+1 control points, while the n−1 remaining segments are determined by K−1 points. In total we obtain M = n(K−1)+2 control points to optimize.

Minimizing A(b) alone leads to the trivial solution, for any set of control points b = (x, …, x), , and this is why the fitting term from (20) is important.

Fitting (0 < λ < ∞). If the segment start and end points are obstructed by noise or allowed to move, we employ a fitting scheme to balance the importance given to the data points d0, …, dn. Equation (20) reads

where λ∈ℝ+ sets the priority to either the data term (large λ) or the mean squared acceleration (small λ) within the minimization. The gradient of the data term is given in Karcher [41], and the gradient of à is given by

Interpolation (λ → ∞). For interpolation we assume that the start point pi−1 and end point pi of the segments i = 1, …, n are fixed to given data di−1 and , respectively. The optimization of the discrete mean squared acceleration A(b) reads

Since the pi are fixed by constraint, they can be omitted from the vector b. We obtain

Since there are n+1 additional constraints, the minimization is hence performed on the product manifold , M′ = M−(n+1) = n(K−2)+1.

4.2. Adjoint Jacobi Fields

In the Euclidean space ℝm, the adjoint operator T* of a linear bounded operator T:ℝm → ℝq is the operator fulfilling

The same can be defined for a linear operator , , on a m-dimensional Riemannian manifold . The adjoint operator satisfies

We are interested in the case where S is the differential operator Dx of a geodesic F(x) = g(t; x, y) for some fixed t∈ℝ and . The differential can be written as

where α are the coefficients of the Jacobi field (11), and ξ, Ξ(t) are given as in Lemma 6. To derive the adjoint differential we observe that for any we have

Hence the adjoint differential is given by

We introduce the adjoint Jacobi field, as

Note that evaluating the adjoint Jacobi field J* involves the same transported frame {Ξ1(t), …Ξm(t)} and the same coefficients α as the Jacobi field J, which means that the evaluation of the adjoint is in no way inferior to the Jacobi field itself.

The adjoint D* of the differential is useful in particular, when computing the gradient of the composition of with . Setting , we obtain for any that

Especially for the evaluation of the gradient of the composite function h°F we obtain

The main advantage of this technique appears in the case of composite functions, i.e., of the form h°F1°F2 (the generalization to composition with K functions is straightforward). The gradient now reads,

The recursive computation of is then given by the following algorithm

Example 12. Forwe knowby Lemma 3.1 and Lemma 3.2 from Bačák et al. [22]. Let t1, t2, t3∈[0, 1] be time points, andbe a given (sub)set of the control points of a (composite) Bézier curveB. We defineas the evaluations ofBat the three given time points. The composition h°F hence consists of (in order of evaluation) the geodesic evaluations of the De Casteljau algorithm, the mid point function for the first and third time point and a distance function.

The recursive evaluation of the gradient starts with the gradient of the distance function. Then, for the first and third arguments, a mid point Jacobi field is applied. The result is plugged into the last geodesic evaluated within the De Casteljau “tree” of geodesics. At each geodesic, a tangent vector at a point g(tj; a, b) is the input for two adjoint Jacobi fields, one mapping to, the other to. This information is available throughout the recursion steps anyways. After traversing this tree backwards, one obtains the required gradient of h°F.

Note also that even the differentiability constraint (4) yields only two further (most outer) adjoint Jacobi fields, namely and . They correspond to variation of the start point and the end point pi, respectively as stated in (10).

4.3. A Gradient Descent Algorithm

To address (22) or (23), we use a gradient descent algorithm, as described in Absil et al. [10, Ch. 4]. For completeness the algorithm is given in Algorithm 1.

Algorithm 1

Input. , its gradient , , step sizes sk>0, k∈ℕ.
Output:
k←0
repeat
Perform a gradient descent step
kk+1
until a stopping criterion is reached
return

Gradient descent algorithm on a manifold

The step sizes are given by the Armijo line search condition presented in Absil et al. [10, Def. 4.2.2]. Let be a Riemannian manifold, be an iterate of the gradient descent, and β, σ∈(0, 1), α>0. Let m be the smallest positive integer such that

We set the step size to in Algorithm 1.

As a stopping criterion we use a maximal number kmax of iterations or a minimal change per iteration . In practice, this last criterion is matched first.

The gradient descent algorithm converges to a critical point if the function F is convex [10, sections 4.3 and 4.4]. The mid-points model (6) posesses two advantages: (i) the complete discretized MSA (7) consists of (chained) evaluations of geodesics and a distance function, and (ii) it reduces to the classical second order differences on the Euclidean space. However, this model is not convex on general manifolds. An example is given in the arXiv preprint (version 3) of Bačák et al. [22]1, Remark 4.6. Another possibility (also reducing to classical second order differences in Euclidean space) is the so-called Log model (see e.g., [42])

The (merely technical) disadvantage of the Log model is that the computation of the gradient involves further Jacobi fields than the one presented above, namely to compute the differentials of the logarithmic map both with respect to its argument Dxlogyx as well as its base point Dylogyx. Still, these can be given in closed form for symmetric Riemannian manifolds [23, Th. 7.2][43, Lem. 2.3]. To the best of our knowledge, the joint convexity of the Log model in x, y and z is still an open question.

5. Examples

In this section, we provide several examples of our algorithm applied to the fitting problem (20).

We validate it first on the Euclidean space and verify that it retrieves the natural cubic smoothing spline. We then present examples on the sphere S2 and the special orthogonal group SO(3). We compare our results with the fast algorithm of Arnould et al. [37], generalized to fitting, and the so-called blended cubic splines from Gousenbourger et al. [38]. The control points of the former are obtained by generalizing the optimal Euclidean conditions of (20) (in ℝm, this is a linear system of equations) to the manifold setting; the curve is afterwards reconstructed by a classical De Casteljau algorithm. In the latter, the curve is obtained as a blending of solutions computed on carefully chosen tangent spaces, i.e., Euclidean spaces.

We will show in this section that the proposed method performs as good as the existing methods when the data is from the Eulidean space (section 5.1). This also means that all methods work equally well whenever the data points on the manifold are local enough. However, we will show by the other examples (sections 5.2 and 5.3) that our proposed method outperforms the others whenever the data points are spread out on the manifold.

The following examples were implemented in MVIRT [12]2, and the comparison implementations from Arnould et al. [37] and Gousenbourger et al. [38] use Manopt [11]3. Note that both toolboxes use a very similar matrix structure and are implemented in Matlab, such that the results can directly be compared.

5.1. Validation on the Euclidean Space

As a first example, we perform a minimization on the Euclidean space . A classical result on ℝm is that the curve γ minimizing (1) is the natural () cubic spline when Γ is a Sobolev space . We compare our approach to the tangential linear system approach derived in Arnould et al. [37] in order to validate our model. We use the following data points

as well as the control points pi = di, and

where the exponential map and geodesic on ℝ3 are actually the addition and line segments, respectively. Note that, by construction, the initial curve is continuously differentiable but (obviously) does not minimize (1). The parameter λ is set to 50. The MSA of this initial curve Ã(b) is approximately 18.8828.

The data points are given to the algorithm of Arnould et al. [37] to reconstruct the optimal Bézier curve B(t), which is the natural cubic spline. The result is shown in Figure 7A together with the first order differences along the curve in Figure 7B as a dotted curve.

Figure 7

Figure 7

The initial interpolating Bézier curve in ℝ3 (dashed, A) with an MSA of 18.8828 is optimized by the linear system method (LS) from [37] (dotted) and by the proposed gradient descent (GD, solid blue). As expected, both curve coincide, with an MSA of 4.981218. The correspondance is further illustrated with their first order differences in (B).

The set of control points is optimized with our proposed method. We discretize the second order difference using N = 1600 points. The resulting curve and the first order difference plots are also obtained using these sampling values and a first order forward difference. The gradient descent algorithm from Algorithm 1 employs the Armijo rule (24) setting , σ = 10−4, and α = 1. The stopping criteria are or , and the algorithm stops when one of the two is met. For this example, the first criterion stopped the algorithm, while the norm of the gradient was of magnitude 10−6.

Both methods improve the initial functional value of Ã(b)≈18.8828 to a value of Ã(bmin)≈4.981218. Both the linear system approach and the gradient descent perform equally. The difference of objective value is 2.4524 × 10−11 smaller for the gradient descent, and the maximal distance of any sampling point of the resulting curves is of size 4.3 × 10−7. Hence, in the Euclidean space, the proposed gradient descent yields the natural cubic spline, as one would expect.

5.2. Examples on the Sphere S2

Validation on a geodesic. As a second example with a known minimizer, we consider the manifold , i.e., the two-dimensional unit sphere embedded in ℝ3, where geodesics are great arcs. We use the data points

aligned on the geodesic connecting the north pole p0 and the south pole p2, and running through a point p1 on the equator. We define the control points of the cubic Bézier curve as follows:

where x0 and x2 are temporary points and pi = di. We obtain two segments smoothly connected since . The original curve is shown in Figure 8A, where especially the tangent vectors illustrate the different speed at pi.

Figure 8

Figure 8

The initial curve (dashed, A) results in a geodesic (solid, B) when minimizing the discrete acceleration. This can also be seen on the first order differences (C).

The control points are optimized with our interpolating model, i.e., we fix the start and end points p0, p1, p2 and minimize A(b).

The curve, as well as the second and first order differences, is sampled with N = 201 equispaced points. The parameters of the Armijo rule are again set to , σ = 10−4, and α = 1. The stopping criteria are slightly relaxed to 10−7 for the distance and 10−5 for the gradient, because of the sines and cosines involved in the exponential map.

The result is shown in Figure 8B. Since the minimizer is the geodesic running from p0 to p2 through p1, we measure the perfomance first by looking at the resulting first order difference, which is constant, as can be seen in Figure 8C. As a second validation, we observe that the maximal distance of the resulting curve to the geodesic is of 2.2357 × 10−6. These evaluations again validate the quality of the gradient descent.

Effect of the data term. As a third example we investigate the effect of λ in the fitting model. We consider the data points

as well as the control points pi = di, and

The remaining control points and are given by the conditions (4). The corresponding curve B(t), t∈[0, 3], is shown in Figure 9 in dashed black. When computing a minimal MSA curve that interpolates B(t) at the points pi, i = 0, …, 3, the acceleration is not that much reduced, see the blue curve in Figure 9A.

Figure 9

Figure 9

The composite Bézier curves are composed of three segments. The initial curve (dashed, A) is optimized with the interpolating model (solid blue, A) as well as with the fitting model for a continuum of values of λ, from λ = 10 (violet, B) to λ = 0.01 (yellow, B). The limit case where λ = 0 yields an unconstrained geodesic (solid black, A).

For fitting, we consider different values of λ and the same parameters as for the last example. The optimized curve fits the data points closer and closer as λ grows, and the limit λ → ∞ yields the interpolation case. On the other hand smaller values of λ yield less fitting, but also a smaller value of the mean squared acceleration. In the limit case, i.e., λ = 0, the curve (more precisely the control points of the Bézier curve) just follows the gradient flow to a geodesic.

The results are collected in Figure 9. In Figure 9A, the original curve (dashed) is shown together with the solution of the interpolating model (solid blue) and the gradient flow result, i.e., the solution of the fitting model (solid black) with λ = 0. Figure 9B illustrates the effect of λ even further with a continuous variation of λ from a large value of λ = 10 to a small value of λ = 0.01. The image is generated by sampling this range with 1000 equidistant values colored in the colormap viridis. Furthermore, the control points are also shown in the same color.

Several corresponding functional values, see Table 1, further illustrate that with smaller values of λ the discretized MSA also reduces more and more to a geodesic. For λ = 0 there is no coupling to the data points and hence the algorithm does not restrict the position of the geodesic. In other words, any choice for the control points that yields a geodesic, is a solution. Note that the gradient flow still chooses a reasonably near geodesic to the initial data (Figure 9A, solid black).

Table 1

λÃ(b)
original10.6122
4.1339
101.6592
10.0733
0.10.0010
0.011.0814e-5
0.0011.6240e-7
03.5988e-9

Functional values for the three-segment composite Bézier curve on the sphere and different values of λ.

Note that the case λ = ∞ corresponds to interpolation.

Comparison with the tangential solution. In the Euclidean space, the method introduced in Arnould et al. [37] and Gousenbourger et al. [38] and the gradient descent proposed here yield the same curve, i.e., the natural cubic spline. On Riemannian manifolds, however, all approaches approximate the optimal solution. Indeed, their method provides a solution by working in different tangent spaces, and ours minimizes a discretization of the objective functional.

In this example we take the same data points di as in (25) now interpreted as points on . Note that the control points constructed in (26) still fit to the sphere since each is built with a vector in and using the corresponding exponential map. The result is shown in Figure 10A. The norm of the first order differences is given in Figure 10B. The initial curve (dashed black) has an objective value of 10.9103. The tangent version (dotted) reduces this value to 7.3293. The proposed method (solid blue) yields a value of 2.7908, regardless of whether the starting curve is the initial one, or the one already computed by Arnould et al. [37]. Note that, when , the tangent version is even not able to compute any result, since the construction is performed in the tangent space of p0 and since logp0p3 is not defined.

Figure 10

Figure 10

The initial composite Bézier curve on S2 (dashed, A) has an MSA of 10.9103. The curve obtained with the linear system (LS, dotted) of [37] has an MSA of 7.3293. The solution returned by our proposed method (GD, solid blue) outperforms them all with an MSA of 2.7908. This is further illustrated by the first order derivative approximated by first order differences in (B).

5.3. An Example of Orientations

Finally we compare our method with the blended splines introduced in Gousenbourger et al. [38] for orientations, i.e., data given on SO(3). Let

denote the rotation matrices in the xy, xz, and yz planes, respectively. We introduce the three data points

These data points are shown in the first line of Figure 11 in cyan.

Figure 11

Figure 11

From top row to bottom row: (1) the initial data points (cyan), (2) the control points computed by the blended Bézier approach from Gousenbourger et al. [38], (3) 17 points on the corresponding curve, where the height represents the absolute first order difference, and the comparing curve is the first order differences from our model, (4) 17 points along the resulting curve from gradient descent, and (5) the resulting control points of the curve in (4).

We set λ = 10 and discretize (20) with N = 401 equispaced points. This sampling is also used to generate the first order finite differences. The parameters of the gradient descent algorithm are set to the same values as for the examples on the sphere.

We perform the blended spline fitting with two segments and cubic splines. The resulting control points are shown in the second row of Figure 11, in green. The objective value is Ã(b) = 0.6464. We improve this solution with the minimization problem (22) on the product manifold SO(3)6. We obtain the control points shown in the last line of Figure 11 and an objective value of for the resulting minimizer .

We further compare both curves by looking at their absolute first order differences. In the third line of Figure 11, we display 17 orientations along the initial curve with its first order difference as height. The line without samples is the absolute first order differences of the minimizer , scaled to the same magnitude. The line is straightened especially for the first Bézier segment. Nevertheless, the line is still bent a little bit and hence not a geodesic. This can be seen in the fourth line which represents 17 samples of the minimizing curve compared to its control points in the last line, which are drawn on a straight line.

6. Conclusion and Future Work

In this paper, we introduced a method to solve the curve fitting problem to data points on a manifold, using composite Bézier curves. We approximate the mean squared acceleration of the curve by a suitable second order difference and a trapezoidal rule, and derive the corresponding gradient with respect to its control points. The gradient is computed in closed form by exploiting the recursive structure of the De Casteljau algorithm. Therefore, we obtain a formula that reduces to a concatenation of adjoint Jacobi fields.

The evaluation of Jacobi fields is the only additional requirement compared to previous methods, which are solely evaluating exponential and logarithmic maps. For these, closed forms are available on symmetric manifolds.

On the Euclidean space our solution reduces to the natural smoothing spline, the unique acceleration minimizing polynomial curve. The numerical experiments further confirm that the method presented in this paper outperforms the tangent space(s)-based approaches with respect to the functional value.

It is still an open question whether there exists a second order absolute finite difference on a manifold, that is jointly convex. Then convergence would follow by standard arguments of the gradient descent. For such a model, another interesting point for future work is to find out whether only an approximate evaluation of the Jacobi fields suffices for convergence. This would mean that, on manifolds where the Jacobi field can only be evaluated approximately by solving an ODE, the presented approach would still converge.

Statements

Author contributions

Both authors listed have contributed equally to the content of this work and approved the paper for publication.

Acknowledgments

This work was supported by the following fundings: RB gratefully acknowledges support by DFG grant BE 5888/2–1; P-YG gratefully acknowledges support by the Fonds de la Recherche Scientifique – FNRS and the Fonds Wetenschappelijk Onderzoek – Vlaanderen under EOS Project no 30468160, and Communauté française de Belgique - Actions de Recherche Concertées (contract ARC 14/19-060). Figure 9A is based on an idea of Rozan I. Rosandi.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Footnotes

1.^See arxiv.org/abs/1506.02409v3

2.^Open source, available at http://ronnybergmann.net/mvirt/

3.^Open source, available at http://www.manopt.org

References

  • 1.

    PytaLAbelD. Interpolatory Galerkin models for Navier-Stokes-equations. IFAC-PapersOnLine (2016) 49:2049. 10.1016/j.ifacol.2016.07.442

  • 2.

    GousenbourgerPYMassartEMusolasAAbsilPAJacquesLHendrickxJMet al. Piecewise-Bézier C1 smoothing on manifolds with application to wind field estimation. In: ESANN2017. Bruges: Springer (2017). Available online at: https://www.elen.ucl.ac.be/Proceedings/esann/esannpdf/es2017-77.pdf

  • 3.

    SanderO. Geodesic finite elements for cosserat rods. Int J Numer Methods Eng. (2010) 82:164570. 10.1002/nme.2814

  • 4.

    ParkJ. Interpolation and tracking of rigid body orientations. In: ICCAS, Gyeonggi-do (2010). p. 66873. 10.1109/ICCAS.2010.5670237

  • 5.

    SuJKurtekSKlassenESrivastavaA. Statistical analysis of trajectories on Riemannian manifolds: bird migration, Hurricane tracking and video surveillance. Ann Appl Stat. (2014) 8:53052. 10.1214/13-AOAS701

  • 6.

    DynN. Linear and nonlinear subdivision schemes in geometric modeling. In: CuckerFPinkusAToddMJ editors. FoCum. Vol. 363 of London Math. Soc. Lecture Note Ser. Hong Kong: Cambridge University Press (2009). p. 6892. 10.1017/CBO9781139107068.004

  • 7.

    WallnerJNavaYazdani EGrohsP. Smoothness properties of Lie group subdivision schemes. Multiscale Model Simulat. (2007) 6:493505. 10.1137/060668353

  • 8.

    ShingelT. Interpolation in special orthogonal groups. IMA J Numer Anal. (2008) 29:73145. 10.1093/imanum/drn033

  • 9.

    GreenPJSilvermanBW. Nonparametric Regression and Generalized Linear Models: A Roughness Penalty Approach. CRC Press (1993). Available online at: https://www.crcpress.com/Nonparametric-Regression-and-Generalized-Linear-Models-A-roughness-penalty/Green-Silverman/p/book/9780412300400

  • 10.

    AbsilPAMahonyRSepulchreR. Optimization Algorithms on Matrix Manifolds. Princeton, NJ: Princeton University Press (2008). Available online at: https://press.princeton.edu/absil

  • 11.

    BoumalNMishraBAbsilPASepulchreR. Manopt, a matlab toolbox for optimization on manifolds. J Mach Learn Res. (2014) 15:14559. Available online at: http://jmlr.org/papers/v15/boumal14a.html

  • 12.

    BergmannR. MVIRT, a toolbox for manifold-valued image registration. In: IEEE International Conference on Image Processing, IEEE ICIP 2017, Beijing. (2017). Available online at: https://ronnybergmann.net/mvirt/

  • 13.

    ZhouXYangCZhaoHYuW. Low-rank modeling and its applications in image analysis. ACM Comput Surveys (2015) 47:36. 10.1145/2674559

  • 14.

    CunninghamJPGhahramaniZ. Linear dimensionality reduction: survey, insights, and generalizations. J Mach Learn Res. (2015) 16:2859900. Available online at: http://jmlr.org/papers/v16/cunningham15a.html

  • 15.

    SunJQuQWrightJ. A geometric analysis of phase retrieval. Found Comput Math. (2017) 8:113198. 10.1007/s10208-017-9365-9

  • 16.

    YuXShenJCZhangJLetaiefKB. Alternating minimization algorithms for hybrid precoding in millimeter wave MIMO systems. IEEE J Selected Top Signal Process. (2016) 10:485500. 10.1109/JSTSP.2016.2523903

  • 17.

    StrekalovskiyECremersD. Total variation for cyclic structures: convex relaxation and efficient minimization. In: IEEE Conference on Computer Vision and Pattern Recognition (2011). p. 190511.

  • 18.

    StrekalovskiyECremersD. Total cyclic variation and generalizations. J Mat Imaging Vis. (2013) 47:25877. 10.1007/s10851-012-0396-1

  • 19.

    LellmannJStrekalovskiyEKoetterSCremersD. Total variation regularization for functions with values in a manifold. In: IEEE International Conference on Computer Vision. Sydney, NSW (2013). p. 294451. 10.1109/ICCV.2013.366

  • 20.

    WeinmannADemaretLStorathM. Total variation regularization for manifold-valued data. SIAM J Imaging Sci. (2014) 7:222657. 10.1137/130951075

  • 21.

    BergmannRLausFSteidlGWeinmannA. Second order differences of cyclic data and applications in variational denoising. SIAM J Imaging Sci. (2014) 7:291653. 10.1137/140969993

  • 22.

    BačákMBergmannRSteidlGWeinmannA. A second order nonsmooth variational model for restoring manifold-valued images. SIAM J Comput. (2016) 38:56797. 10.1137/15M101988X

  • 23.

    BergmannRFitschenJHPerschJSteidlG. Priors with coupled first and second order differences for manifold-valued image processing. J Math Imaging Vis. (2017) 60:145981. 10.1007/s10851-018-0840-y

  • 24.

    BrediesKHollerMStorathMWeinmannA. Total generalized variation for manifold-valued data. SIAM J Imaging Sci. (2018) 11:178548. 10.1137/17M1147597

  • 25.

    BergmannRPerschJSteidlG. A parallel Douglas–Rachford algorithm for restoring images with values in symmetric Hadamard manifolds. SIAM J Imaging Sci. (2016) 9:90137. 10.1137/15M1052858

  • 26.

    GrohsPSprecherM. Total variation regularization on Riemannian manifolds by iteratively reweighted minimization. Inform Inference (2016) 5:35378. 10.1093/imaiai/iaw011

  • 27.

    BergmannRChanRHHielscherRPerschJSteidlG. Restoration of manifold-valued images by half-quadratic minimization. Inverse Prob Imaging (2016) 10:281304. 10.3934/ipi.2016001

  • 28.

    SamirCAbsilPASrivastavaAKlassenE. A gradient-descent method for curve fitting on Riemannian manifolds. Found Comput Math. (2012) 12:4973. 10.1007/s10208-011-9091-7

  • 29.

    SuJDrydenILKlassenELeHSrivastavaA. Fitting smoothing splines to time-indexed, noisy points on nonlinear manifolds. Image Vis Comput. (2012) 30:42842. 10.1016/j.imavis.2011.09.006

  • 30.

    BoumalNAbsilPA. A discrete regression method on manifolds and its application to data on SO(n). In: IFAC Proceedings Volumes (IFAC-PapersOnline). Vol. 18. Milano (2011). p. 228489. 10.3182/20110828-6-IT-1002.00542

  • 31.

    KimKRDrydenILLeH. Smoothing splines on Riemannian manifolds, with applications to 3D shape space. (2018). arXiv[Preprint]:1801.04978.

  • 32.

    FletcherPT. Geodesic regression and the theory of least squares on Riemannian manifolds. Int J Comput Vis. (2013) 105:17185. 10.1007/s11263-012-0591-y

  • 33.

    RentmeestersQ. A gradient method for geodesic data fitting on some symmetric Riemannian manifolds. In: Decision and Control and European Control Conference (CDC-ECC), 2011 50th IEEE Conference on. Orlando, FL (2011). p. 71416. 10.1109/CDC.2011.6161280

  • 34.

    FarinG,. Curves Surfaces for CAGD. Morgan Kaufmann (2002). Available online at: https://www.elsevier.com/books/curves-and-surfaces-for-cagd/farin/978-1-55860-737-8

  • 35.

    AbsilPAGousenbourgerPYStriewskiPWirthB. Differentiable Piecewise-Bézier Surfaces on Riemannian Manifolds. SIAM J Imaging Sci. (2016) 9:1788828. 10.1137/16M1057978

  • 36.

    PopielTNoakesL. Bézier curves and interpolation in Riemannian manifolds. J Approx Theory (2007) 148:11127. 10.1016/j.jat.2007.03.002

  • 37.

    ArnouldAGousenbourgerPYSamirCAbsilPACanisM. Fitting smooth paths on Riemannian manifolds: endometrial surface reconstruction and preoperative MRI-based navigation. In: NielsenFBarbarescoF editors. GSI2015. Springer International Publishing (2015). p. 4918.

  • 38.

    GousenbourgerPYMassartEAbsilPA. Data fitting on manifolds with composite Bézier-like curves and blended cubic splines. J Math Imaging Vis. (2018). 10.1007/s10851-018-0865-2

  • 39.

    O'NeillB. Elementary Differential Geometry. London: Academic Press Inc. (1966).

  • 40.

    doCarmo MP. Riemannian Geometry. Vol. 115. Basel: Birkhäuser; 1992. Tranl. by F. Flatherty.

  • 41.

    KarcherH. Riemannian center of mass and mollifier smoothing. Commun Pure Appl Math. (1977) 30:50941. 10.1002/cpa.3160300502

  • 42.

    BoumalN. Interpolation and regression of rotation matrices. In: NielsenFBarbarescoF editors. Geometric Science of Information. Berlin; Heidelberg: Springer. (2013). p. 34552.

  • 43.

    PerschJ. Optimization Methods in Manifold-Valued Image Processing [Dissertation]. Technische Universität Kaiserslautern (2018).

Summary

Keywords

Riemannian manifolds, curve fitting, composite Bézier curves, Jacobi fields, variational models

Citation

Bergmann R and Gousenbourger P-Y (2018) A Variational Model for Data Fitting on Manifolds by Minimizing the Acceleration of a Bézier Curve. Front. Appl. Math. Stat. 4:59. doi: 10.3389/fams.2018.00059

Received

26 July 2018

Accepted

13 November 2018

Published

12 December 2018

Volume

4 - 2018

Edited by

Luca Marchetti, Microsoft Research-University of Trento Centre for Computational and Systems Biology(COSBI), Italy

Reviewed by

Giovanni Mascali, Università della Calabria, Italy; Vincenzo Bonnici, Università degli Studi di Verona, Italy

Updates

Copyright

*Correspondence: Ronny Bergmann

This article was submitted to Optimization, a section of the journal Frontiers in Applied Mathematics and Statistics

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics