with the tensor D being the unknown variable and b a value that depends on the acquisition settings. The estimation of the tensors in the volume domain Ω can be done through direct inference (6 acquisitions are at least available), which is equivalent to minimizing:
3.1 Measuring Similarities from diffusion weighted images
We aim at simultaneously estimating and smoothing the tensor field, therefore the weights w(x, y) in E smooth should be precalculated using the raw data. The most straightforward estimation of the distances can be done through the algebraic distance between the log(S k /S 0) for two neighborhood voxels in any direction
One can easily show that such an expression does not reflect similarity between tensors according to the norm ∥ ⋅ ∥ F . In fact, this leads to
such that (each new vector of the new basis is a linear combination of the vectors of the initial basis). This procedure allows us to have an approximation of ||D(x) − D(y)|| F directly from the raw data S k and S 0 as follows
Get Clinical Tree app for offline access
3.2 Semi-Definite Positive Gradient Descent
One now can seek the lowest potential of the cost function towards recovering the optimal solution on the tensor space. The present framework consists of a convex energy with a unique minimum which can be reached using a projected gradient descent on the space of semi-definitive positive matrices. The projection from S 3 onto S +3 denoted by is well defined and has an explicit expression. Indeed, projecting M amounts to replacing the negative eigenvalues in its spectral decomposition by 0 [8, 12]. Note that we minimize over the set of semi-definite positive matrices because it is topologically closed, as opposed to the set of definite positive matrices. In the current setting, the problem is well posed and the projected gradient descent algorithm is convergent for a suitable choice of the time step dt. Using a weighting factor λ between the data attachment term and the regularization energy, the gradient descent can be expressed as the following equation
where
Let us define the norm ∥ ⋅ ∥ TF over the whole tensor field D as ∥ D ∥ TF = ∫ Ω ∥ D(x) ∥ F d x. Considering two tensor fields D 1 and D 2, we show in the following that the gradient of our energy functional is L-Lipschitz. The constant L will allow us to choose automatically a time step that insures the convergence of the algorithm.
where is the number of the considered neighbors. Thus the gradient of the objective function is L-Lipschitz with . Choosing makes the projected gradient descent convergent [3].
We can give an interpretation of our regularization energy in terms of diffusion-weighted images smoothing. It can be easily verified that for each direction k
We can see that minimizing E smooth has a direct implication on the normalized diffusion weighted images . Reconstructing the tensors using a linear combination of the tensors in its neighborhood leads to the reconstruction of the normalized signals using a weighted geometric mean of the neighboring signals where the weights are not calculated only with a single volume S k but also with the volumes obtained from the other magnetic gradient directions.
4 SVM classification and kernels on tensors
4.1 Two-class Support Vector Machines
We briefly review the principles of two class SVMs [26]. Given N points x i with known class information y i (either +1 or −1), SVM training consists in finding the optimal separating hyperplane described by the equation w t x + b = 0 with the maximum distance to the training examples. It amounts to solving a dual convex quadratic optimization problem and each data point x is classified using the SVM output function f(x) = (Σ i N α i y i xx i ) + b. The algorithm is extended to achieve non linear separation using a kernel function K(x,y)(symmetric, positive definite) that is used instead of the standard inner product.
4.2 Information diffusion kernel
In order to define a kernel on the set of definite positive matrices, we can propagate class and structure information using its geometry as a Riemannian manifold [22]. Intuitively, we can see the construction of this kernel as diffusing the labels of the training set to the whole set of definite positive matrices. Therefore, similarly to heat diffusion on a Euclidean space, where the solution is given by the convolution of the initial condition by a Gaussian kernel, heat diffusion on a Riemannian manifold is driven by a kernel function K t and given by the following asymptotic series expansion [17]:
where d corresponds to the geodesic distance induced by the Riemannian metric, a n are the coefficients of the series expansion and t is the diffusion time, which is a parameter of the kernel. We use a first order approximation in t of the previous expression that yields
In our case, d has an explicit expression given by where λ i are the generalized eigenvalues of D 1 and D 2 [22].