Two-Sided Preconditioner Updates
- The paper introduces a two-sided preconditioning technique that splits a constant-coefficient approximation into left and right factors, resulting in a condition number independent of system size.
- The method exploits the diagonalizability of the spatial stiffness matrix via fast sine transforms and the Toeplitz structure of the temporal convolution for efficient inversion.
- The approach achieves near-optimal O(JN log(JN)) computational cost and scalable parallel-in-time updates for variable-coefficient non-local evolutionary equations.
Two-sided preconditioner updates refer to a class of preconditioning strategies for large, structured linear systems, especially those arising from the discretization of non-local evolutionary equations involving variable-coefficient elliptic operators and weakly singular temporal convolution kernels. In the context of parallel-in-time (PinT) algorithms for all-at-once systems, such preconditioners are constructed by splitting a suitably chosen constant-coefficient approximation of the system matrix into left and right factors, each of which can be efficiently inverted by exploiting the block-diagonal structure in space (via diagonalization) and the Toeplitz structure in time. The resulting two-sided preconditioning yields a preconditioned system with condition number provably independent of the system size, enables near-optimal parallel implementation, and admits efficient iterative and direct solvers, particularly for variable-coefficient problems where standard approaches are often suboptimal.
1. Construction of Two-Sided Preconditioners
Given the all-at-once matrix arising from the discretization of a variable-coefficient Laplacian (in space) and a Toeplitz temporal convolution matrix , the two-sided preconditioning technique begins by selecting a constant such that and . The constant-coefficient approximation is then defined as
where is the stiffness matrix for with unit diffusion.
The split is performed as follows:
- Right preconditioner (acts last): .
- Left preconditioner (acts first): .
The two-sided preconditioned system to solve is
where is the discretized right-hand side.
2. Derivation and Splitting Strategy
Starting from , the requirement that contains only a fractional power of and no leads uniquely to the given split. Algebraically,
by factoring with .
This split isolates the main spatial stiffness (constant-coefficient Laplacian) on the right, facilitating diagonalization, and collects the remaining convolution and compensation for variable coefficients on the left.
3. Explicit Inversion Using Diagonalization and Toeplitz Structure
For efficient inversion and matrix-vector multiplication, the following structure is exploited:
- The stiffness matrix can be diagonalized by a fast sine transform (FST);
with ( the sine-transform matrix) and .
- The right preconditioner inverse is
- The left preconditioner, under , block-diagonalizes into independent blocks,
each being a sum of diagonal and lower-triangular Toeplitz matrices, and thus invertible (with also lower-triangular Toeplitz). The inverse is efficiently given by
4. Krylov Iterative Algorithm and Parallel-in-Time Updates
To solve the preconditioned system , any standard Krylov subspace iterative algorithm (e.g., GMRES, normalized CG for symmetric systems) may be used. The core matrix-vector product is implemented in three phases:
- Pre-apply right preconditioner: Transform by using joint FST, scaling, and inverse FST: cost .
- Apply system matrix : Multiply by and efficiently. Cost for spatial multiplication is (if is sparse); temporal multiplication via Toeplitz structure is .
- Pre-apply left preconditioner: Transform by FST, invert the block Toeplitz systems in parallel for each spatial index (cost ), and transform back. All temporal solves are independent and can be distributed for ideal PinT scaling.
The total cost per matrix-vector product is , with storage dominated by the all-at-once vector and entries for first columns of . Only storage for is also needed. FFT-based routines enable in-place application of .
5. Theoretical Properties and Robustness
Under the assumption
the two-sided preconditioned matrix
has condition number
independent of and . This establishes robustness with respect to both problem size and discretization. The proof proceeds via bounding Rayleigh quotients using block-Kronecker inequalities and the positive-definiteness of the temporal matrices.
6. Implementation Complexity and Practical Considerations
Each application of the two-sided preconditioners has computational cost . If processors are available, distributing the independent temporal solves in Phase 3 achieves near-ideal PinT scaling. Storage requirements are minimal since only the diagonals of the block systems and the first columns of the Toeplitz inverses are needed.
If the underlying spatial coefficient varies slowly, it suffices to recompute scaled blocks as changes, requiring at most per update. In the special case , the matrix coincides with , and the two-sided preconditioner becomes a direct solver with complexity .
The method exploits the precise algebraic structure of the discretized system (Toeplitz in time, diagonalizable in space) without further approximation and is matrix-size robust in terms of conditioning. These properties distinguish it from classical one-sided and approximate preconditioners for similar equations.
7. Applications and Extensions
Two-sided preconditioners in the manner of Lin–Ng–Zhi (2020) are applicable in the fast iterative or direct solution of discretized non-local evolutionary equations with variable coefficients and weakly singular temporal kernels. They enable efficient PinT solvers for large-scale time-dependent PDEs with guaranteed convergence rates and parallel scalability. The approach is well suited to cases where temporal non-locality and spatially varying diffusion preclude straightforward block-diagonalization or classical (block) circulant preconditioners. A plausible implication is that similar splitting strategies may generalize to other hybrid-structured systems where distinct algebraic properties in different dimensions can be simultaneously exploited for parallelism and conditioning control (Lin et al., 2021).