Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
10 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
3 tokens/sec
DeepSeek R1 via Azure Pro
51 tokens/sec
2000 character limit reached

Symmetric Matrix Eigenvalue Decomposition

Updated 14 July 2025
  • Eigenvalue decomposition for symmetric matrices is the process of orthogonally diagonalizing a real symmetric matrix into a basis of real eigenvalues and orthonormal eigenvectors.
  • The method leverages univariate Taylor polynomial arithmetic to compute higher-order derivatives, facilitating precise sensitivity analysis and structured handling of repeated eigenvalues.
  • Reverse mode algorithmic differentiation applies structured lifting and adjoint propagation to ensure stable, scalable computations in advanced numerical simulations and optimization.

Eigenvalue decomposition for symmetric matrices is a central concept in numerical linear algebra, theoretical physics, and applied mathematics. For a real symmetric matrix ARN×NA \in \mathbb{R}^{N \times N}, the spectral theorem guarantees that AA can be orthogonally diagonalized: there exists an orthogonal matrix QQ and a diagonal matrix Λ\Lambda such that A=QΛQA = Q\Lambda Q^\top. The eigenvalues are real and the eigenvectors can be chosen to form an orthonormal basis. This decomposition is fundamental not only for matrix functions, stability analysis, and dimensionality reduction, but also for the algorithmic differentiation of programs involving matrix factorizations.

1. Implicit Definition and Characterization

The symmetric eigenvalue decomposition can be cast as the solution of an implicit system of matrix equations. For a symmetric AA, one seeks QO(N)Q \in O(N) (orthogonal) and Λ\Lambda (diagonal) so that: QAQ=Λ,QQ=I,(PL+PR)Λ=0,Q^\top A Q = \Lambda, \qquad Q^\top Q = I,\qquad (P_L + P_R)\circ \Lambda = 0, where PLP_L and PRP_R are strictly lower and upper triangular projectors (with (PL)ij=δj<i(P_L)_{ij} = \delta_{j<i}, (PR)ij=δi<j(P_R)_{ij} = \delta_{i<j}), and \circ denotes element-wise (Hadamard) product. The "sifting" condition ensures off-diagonal entries of Λ\Lambda are zero in a structured way. These equations capture the eigenvalue-eigenvector mapping as an implicit function of AA, providing a foundation for algorithmic differentiation and higher-order analysis.

2. Higher-Order Derivatives: Univariate Taylor Polynomial Arithmetic

To differentiate the eigenvalue decomposition with respect to AA, higher-order derivatives are computed by "lifting" the implicit equations to operate on truncated Taylor series, known as univariate Taylor polynomial (UTP) arithmetic. One represents the input as: [A]D=A0+A1T++AD1TD1,[A]_D = A_0 + A_1 T + \cdots + A_{D-1} T^{D-1}, and seeks Taylor polynomial expansions for [Q]D[Q]_D, [Λ]D[\Lambda]_D satisfying the implicit constraints up to order DD. This is achieved by Newton-Hensel lifting:

  1. Compute the decomposition at zero order (A0A_0).
  2. Sequentially solve for corrections in higher orders, refining the Taylor expansions such that all terms up to TD1T^{D-1} satisfy the system to the desired truncation.
  3. When eigenvalues are repeated (block-diagonal structure), refinements are performed within each block, ensuring that the expansion respects eigenvalue multiplicities.

This approach enables systematic computation of all higher-order derivatives of the outputs (Q,Λ)(Q,\,\Lambda) with respect to perturbations of AA.

3. Reverse Mode Algorithmic Differentiation and Sensitivity Analysis

In reverse (adjoint) mode, sensitivities with respect to AA are "pulled back" from linear forms defined on the outputs (Q,Λ)(Q,\,\Lambda). Differentiating the implicit system yields, for distinct eigenvalues: A=Q[Λ+H(QQ)]Q\overline{A} = Q \left[\overline{\Lambda} + H \circ (Q^\top \overline{Q}) \right] Q^\top where Λ\overline{\Lambda} and Q\overline{Q} are the output adjoints and HH is the off-diagonal matrix with

Hij={1λjλiij 0i=jH_{ij} = \begin{cases} \dfrac{1}{\lambda_j - \lambda_i} & i \neq j\ 0 & i = j \end{cases}

This formula, derived from matrix calculus by differentiating the conditions QAQ=ΛQ^\top A Q = \Lambda and QQ=IQ^\top Q = I, provides a numerically stable way to propagate adjoints/sensitivities back to AA provided the eigenvalues are simple. When eigenvalues are repeated, more intricate block-diagonal arguments and higher-order conditions are enforced.

4. Structured Lifting and Block-Diagonalization

A major algorithmic feature is recognizing and efficiently handling eigenvalue multiplicity. At each order of the Taylor expansion, the solution separates into blocks corresponding to repeated eigenvalues. The algorithm computes small block-diagonal eigenproblems on each active block, "lifts" the solutions (denoted [Λ^s,s]Dd[\hat{\Lambda}_{s,s}]_{D-d}, [Q^s,s]Dd[\hat{Q}_{s,s}]_{D-d} for block ss) and assembles the global expansion: [Λd+1]D=[Λd]d+[Λ^]DdTd,[Qd+1]D=[Qd]D[Q]D,[\Lambda^{d+1}]_D = [\Lambda^d]_d + [\hat{\Lambda}]_{D-d} T^d,\quad [Q^{d+1}]_D = [Q^d]_D [Q]_D, where [Q]D[Q]_D is re-orthogonalized to ensure the global QQ remains orthogonal to order DD. This blockwise lifting is essential for structured sparsity and numerical efficiency.

5. Unified Forward and Reverse AD with UTP Arithmetic

Combining UTP arithmetic and matrix calculus, both forward and reverse modes operate natively over truncated Taylor polynomials:

  • Each variable (matrix or vector component) is replaced by a Taylor series.
  • Basic matrix functions (addition, multiplication, inversion) are extended to Taylor polynomials.
  • In forward mode, this propagates high-order derivatives; in reverse mode, it propagates adjoints through the chain rule at the Taylor series level.
  • The pullbacks are organized so the chain rule holds on bundles of Taylor coefficients, with explicit computation of higher-order derivatives via coefficient recursion.

This machinery allows all algorithms "containing" symmetric EVP routines (for example, optimal experimental design, index calculation for DAEs) to be differentiated to arbitrary order, within a unified automatic differentiation framework.

6. Numerical Examples, Applications, and Special Cases

Applications and examples in the framework include:

  • Matrices parameterized as A(T)=A0+A1T+A2T2+A(T) = A_0 + A_1 T + A_2 T^2 + \cdots, where the higher-order behavior of eigenvalues and eigenvectors is analyzed via the Taylor expansion.
  • When A0A_0 has repeated eigenvalues, higher-order terms often reveal splitting (eigenvalue unmixing), and the block-diagonalization framework applies, guaranteeing that at each relaxation level, the problem is well-posed within active blocks.
  • The approach handles both forward (pushforward) and reverse (pullback) differentiation and recovers correct derivatives/sensitivities even in the presence of near-degeneracy of eigenvalues.

Explicit algorithms (notably Algorithm 3 and Algorithm 4 "eigh" in the paper) provide stepwise details for sequential lifting and block-structured orthogonalization, addressing technicalities at all levels of eigenvalue multiplicity.

7. Impact and Broader Significance

The combination of UTP arithmetic with matrix calculus for eigenvalue decomposition systems enables:

  • Efficient, scalable, and accurate computation of higher-order derivatives for programs involving symmetric EVP.
  • The preservation of structural properties (orthogonality, block-diagonality) throughout the differentiation process.
  • Applicability to a wide range of domains including optimum experimental design, control theory, numerical simulation with sensitivity requirements, and automatic differentiation frameworks in modern machine learning and computational science.

This framework is a cornerstone for differentiable linear algebraic routines, addressing both theoretical and practical aspects of differentiation through symmetric matrix decompositions, and enabling new computational tools for sensitivity and optimization in large-scale scientific computing.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this topic yet.