by specifying neighboring points and tangent vectors, which allows to differentiate smooth functions on the manifold. The topology also impacts the global structure as it determines if there exists a connected path between two points. However, we need an additional structure to quantify how far away two connected points are: a distance. By restricting to distances which are compatible with the differential structure, we enter into the realm of Riemannian geometry. A Riemannian metric is defined by a continuous collection of scalar products ) on each tangent space at point of the manifold. Thus, if we consider a curve on the manifold, we can compute at each point its instantaneous speed vector (this operation only involves the differential structure) and its norm to obtain the instantaneous speed (the Riemannian metric is needed for this operation). To compute the length of the curve, this value is integrated as usual along the curve. The distance between two points of a connected Riemannian manifold is the minimum length among the curves joining these points. The curves realizing this minimum are called geodesics. The calculus of variations shows that geodesics are the solutions of a system of second order differential equations depending on the Riemannian metric. In the following, we assume that the manifold is geodesically complete, i.e. that all geodesics can be indefinitely extended. This means that the manifold has neither boundary nor any singular point that we can reach in a finite time. As an important consequence, the Hopf-Rinow-De Rham theorem states that there always exists at least one minimizing geodesic between any two points of the manifold (i.e. whose length is the distance between the two points).
2.2 Exponential chart
Let
be a point of the manifold that we consider as a local reference and
a vector of the tangent space
at that point. From the theory of second order differential equations, we know that there exists one and only one geodesic
starting from that point with this tangent vector. This allows to wrap the tangent space onto the manifold, or equivalently to develop the manifold in the tangent space along the geodesics (think of rolling a sphere along its tangent plane at a given point), by mapping to each vector
the point
of the manifold that is reached after a unit time by the geodesic
. This mapping
is called the
exponential map at point
. Straight lines going through 0 in the tangent space are transformed into geodesics going through point
on the manifold and distances along these lines are conserved (Fig.
1).
The exponential map is defined in the whole tangent space
(since the manifold is geodesically complete) but it is generally one-to-one only locally around 0 in the tangent space (i.e. around
in the manifold). In the sequel, we denote by
the inverse of the exponential map: this is the smallest vector (in norm) such that
. If we look for the maximal definition domain, we find out that it is a star-shaped domain delimited by a continuous curve
called the
tangential cut-locus. The image of
by the exponential map is the cut locus
of point
. This is (the closure of) the set of points where several minimizing geodesics starting from
meet. On the sphere
for instance, the cut locus of a point
is its antipodal point and the tangential cut locus is the circle of radius
.
The exponential and log maps within this domain realize a chart (a local parameterization of the manifold) called
the exponential chart at point . It covers all the manifold except the cut locus of the reference point
, which has a null measure. In this chart, geodesics starting from
are straight lines, and the distance from the reference point are conserved. This chart is somehow the “most linear” chart of the manifold with respect to the reference point
. The set of all the exponential charts at each point of the manifold realize an atlas which allows working very easily on the manifold, as we will see in the following.
2.3 Practical implementation
The exponential and logarithmic maps (from now on Exp and Log maps) are obviously different for each manifold and for each metric. Thus they have to be determined and implemented on a case by case basis. Example for rotations, rigid body transformations can be found for the left invariant metric in [
60], and examples for tensors in [
4,
58]. Exponential charts constitute very powerful atomic functions in terms of implementation on which we will be able to express practically all the geometric operations: the implementation of
and
is the basis of programming on Riemannian manifolds, as we will see in the following.
In a Euclidean space, the exponential charts are nothing but one orthonormal coordinates system translated at each point: we have in this case
and
. This example is more than a simple coincidence. In fact, most of the usual operations using additions and subtractions may be reinterpreted in a Riemannian framework using the notion of
bi-point, an antecedent of vector introduced during the 19th Century. Indeed, vectors are defined as equivalent classes of bi-points in a Euclidean space. This is possible because we have a canonical way (the translation) to compare what happens at two different points. In a Riemannian manifold, we can still compare things locally (by parallel transportation), but not any more globally. This means that each “vector” has to remember at which point of the manifold it is attached, which comes back to a bi-point.
A second way to see the vector
is as a vector of the tangent space at point
. Such a vector may be identified to a point on the manifold using the exponential map
. Conversely, the logarithmic map may be used to map almost any bi-point
into a vector
of
. This reinterpretation of addition and subtraction using logarithmic and exponential maps is very powerful to generalize algorithms working on vector spaces to algorithms on Riemannian manifolds, as illustrated in Table
1 and the in following sections.
Table 1
Re-interpretation of standard operations in a Riemannian manifold
Subtraction |
|
|
Addition |
|
|
Distance |
|
|
Mean value (implicit) |
|
|
Gradient descent |
|
|
Geodesic interpolation |
|
|
2.4 Example of Metrics on Covariance matrices
Let us take an example with positive definite symmetric matrices, called tensors in medical image analysis. They are used for instance to encode the covariance matrix of the Brownian motion (diffusion) of water in Diffusion Tensor Imaging (DTI) [
8,
40] or to encode the joint variability at different places (Green function) in shape analysis (see [
29–
31] and Sect.
4). They are also widely used in image analysis to guide the segmentation, grouping and motion analysis [
16,
47,
73,
74].
The main problem is that the tensor space is a manifold that is not a vector space with the usual additive structure. Indeed, the positive definiteness constraint delimits a convex half-cone in the vector space of symmetric matrices. Thus, convex operations (like the mean) are stable in this space but problems arise with more complex operations. For instance, there is inevitably a point in the image where the time step is not small enough when smoothing fields of tensors with gradient descents, and this results into negative eigenvalues.
To answer that problem, we proposed in [
58] to endow the space of tensors with a Riemannian metric invariant by any change of the underlying space coordinates, i.e. invariant under the action of affine transformations on covariance matrices. A few mathematical developments showed that the Exp, Log and distance maps were given with quite simple formulas involving the matrix logarithm exp and log:
This metric leads to a very regular Hadamard manifold structure, a kind of hyperbolic space without cut-locus, which simplifies the computations. Tensors with null and infinite eigenvalues are both at an infinite distance of any positive definite symmetric matrix: the cone of positive definite symmetric matrices is changed into a space of “constant” (homogeneous) non-scalar curvature without boundaries. Moreover, there is one and only one geodesic joining any two tensors, the mean of a set of tensors is uniquely defined, and we can even define globally consistent orthonormal coordinate systems of tangent spaces. Thus, the structure we obtain is very close to a vector space, except that the space is curved.
This affine-invariant Riemannian metric derives from affine invariant connections on homogeneous spaces [
54]. It has been introduced in statistics to model the geometry of the multivariate normal family (the Fisher information metric) [
18,
19,
64] and in simultaneously by many teams in medical image analysis to deal with DTI [
9,
33,
42,
52,
58]. In [
58], we showed that this metric could be used not only to compute distances between tensors, but also as the basis of a complete computational framework on manifold-valued images as will be detailed in Sect.
3.
By trying to put a Lie group structure on the space of tensors, Vincent Arsigny observed later that the matrix exponential was a diffeomorphism from the space of symmetric matrices to the tensor space. Thus, one can seamlessly transport all the operations defined in the vector space of symmetric matrices to the tensor space [
4,
5]. This gives a commutative Lie group structure to the tensors, and the Euclidean metric on symmetric matrices is transformed into a bi-invariant Riemannian metric on the tensor manifold. As geodesics are straight lines in the space of symmetric matrices, the expression of the Exp, Log and distance maps for the Log-Euclidean metric is easily determined:
These formulas look more complex than for the affine invariant metric because they involve the differential of the matrix exponential and logarithm in order to transport tangent vectors from one space to another [
59]. However, they are in fact nothing but the transport of the addition and subtraction through the exponential of symmetric matrices. In practice, the log-Euclidean framework consist in taking the logarithm of the tensor data, computing like usual in the Euclidean space of symmetric matrices, and coming back at the end to the tensor space using the exponential [
3,
5].
From a theoretical point of view, geodesics through the identity are the same for both log-Euclidean and affine-invariant metrics, but this is not true any more in general at other points of the tensor manifold [
4]. A careful comparison of both metrics in practical applications [
3,
5] showed that there was very few differences on the results (of the order of 1%) on real DTI images, but that the log-Euclidean computations where 4 to 10 times faster. For other types of applications, like adaptive re-meshing [
53], the anisotropy of the tensors can be much larger, which may lead to larger differences. In any case, initializing the iterative optimizations of affine-invariant algorithms with the log-Euclidean result drastically speeds-up the convergence. Important application example of this tensor computing framework were provided in [
27,
28] with a statistically grounded estimation and regularization of DTI images. The white matter tractography that was allowed by these methods in clinical DTI images with very poor signal to noise ratios could lead to new clinical indications of DTI, for instance in the spinal chord [
23].