Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 173 tok/s
Gemini 2.5 Pro 54 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 30 tok/s Pro
GPT-4o 77 tok/s Pro
Kimi K2 187 tok/s Pro
GPT OSS 120B 440 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Two-Sided Preconditioner Updates

Updated 10 November 2025
  • The paper introduces a two-sided preconditioning technique that splits a constant-coefficient approximation into left and right factors, resulting in a condition number independent of system size.
  • The method exploits the diagonalizability of the spatial stiffness matrix via fast sine transforms and the Toeplitz structure of the temporal convolution for efficient inversion.
  • The approach achieves near-optimal O(JN log(JN)) computational cost and scalable parallel-in-time updates for variable-coefficient non-local evolutionary equations.

Two-sided preconditioner updates refer to a class of preconditioning strategies for large, structured linear systems, especially those arising from the discretization of non-local evolutionary equations involving variable-coefficient elliptic operators and weakly singular temporal convolution kernels. In the context of parallel-in-time (PinT) algorithms for all-at-once systems, such preconditioners are constructed by splitting a suitably chosen constant-coefficient approximation of the system matrix into left and right factors, each of which can be efficiently inverted by exploiting the block-diagonal structure in space (via diagonalization) and the Toeplitz structure in time. The resulting two-sided preconditioning yields a preconditioned system with condition number provably independent of the system size, enables near-optimal parallel implementation, and admits efficient iterative and direct solvers, particularly for variable-coefficient problems where standard approaches are often suboptimal.

1. Construction of Two-Sided Preconditioners

Given the all-at-once matrix ARJN×JNA\in \mathbb{R}^{JN\times JN} arising from the discretization of a variable-coefficient Laplacian LaRJ×JL_a \in \mathbb{R}^{J\times J} (in space) and a Toeplitz temporal convolution matrix TRN×NT\in \mathbb{R}^{N\times N}, the two-sided preconditioning technique begins by selecting a constant β\beta such that 0<amina(x)amax<0< a_{\min} \leq a(x) \leq a_{\max} < \infty and β=aminamax\beta = \sqrt{a_{\min} a_{\max}}. The constant-coefficient approximation is then defined as

P:=βL1IN+IJT,P := \beta L_1 \otimes I_N + I_J \otimes T,

where L1L_1 is the stiffness matrix for Δ-\Delta with unit diffusion.

The split is performed as follows:

  • Right preconditioner (acts last): PR:=(βL1)1/2INP_R := (\beta L_1)^{1/2}\otimes I_N.
  • Left preconditioner (acts first): PL:=PPR1=(βL1)1/2IN+(βL1)1/2TP_L := P P_R^{-1} = (\beta L_1)^{1/2}\otimes I_N + (\beta L_1)^{-1/2}\otimes T.

The two-sided preconditioned system to solve is

PL1APR1U^=PL1F,U=PR1U^,P_L^{-1} A P_R^{-1} \widehat U = P_L^{-1} F, \quad U = P_R^{-1} \widehat U,

where FF is the discretized right-hand side.

2. Derivation and Splitting Strategy

Starting from P=βL1IN+IJTP = \beta L_1 \otimes I_N + I_J \otimes T, the requirement that PRP_R contains only a fractional power of L1L_1 and no TT leads uniquely to the given split. Algebraically,

PL=(βL1)1/2IN+(βL1)1/2T,P_L = (\beta L_1)^{1/2}\otimes I_N + (\beta L_1)^{-1/2} \otimes T,

by factoring P=PLPRP = P_L P_R with PR=(βL1)1/2INP_R = (\beta L_1)^{1/2}\otimes I_N.

This split isolates the main spatial stiffness (constant-coefficient Laplacian) on the right, facilitating diagonalization, and collects the remaining convolution and compensation for variable coefficients on the left.

3. Explicit Inversion Using Diagonalization and Toeplitz Structure

For efficient inversion and matrix-vector multiplication, the following structure is exploited:

  • The stiffness matrix L1L_1 can be diagonalized by a fast sine transform (FST);

L1=QΛQT,L_1 = Q \Lambda Q^T,

with Q=Sm1...SmdQ = S_{m_1} \otimes ... \otimes S_{m_d} (SmS_m the m×mm\times m sine-transform matrix) and Λ=diag(λ1,...,λJ)\Lambda = \operatorname{diag}(\lambda_1, ..., \lambda_J).

  • The right preconditioner inverse is

PR1=(QIN)[(βΛ)1/2IN](QIN).P_R^{-1} = (Q\otimes I_N)\left[(\beta \Lambda)^{-1/2} \otimes I_N\right](Q\otimes I_N).

  • The left preconditioner, under QINQ\otimes I_N, block-diagonalizes into JJ independent N×NN\times N blocks,

Ti:=(βλi)1/2IN+(βλi)1/2T,i=1,...,J,T_i := (\beta \lambda_i)^{1/2} I_N + (\beta \lambda_i)^{-1/2} T, \quad i=1, ..., J,

each being a sum of diagonal and lower-triangular Toeplitz matrices, and thus invertible (with Ti1T_i^{-1} also lower-triangular Toeplitz). The inverse is efficiently given by

PL1=(QIN)diag(T11,...,TJ1)(QIN).P_L^{-1} = (Q\otimes I_N) \operatorname{diag}(T_1^{-1}, ..., T_J^{-1}) (Q \otimes I_N).

4. Krylov Iterative Algorithm and Parallel-in-Time Updates

To solve the preconditioned system PL1APR1U^=PL1FP_L^{-1} A P_R^{-1} \widehat{U} = P_L^{-1} F, any standard Krylov subspace iterative algorithm (e.g., GMRES, normalized CG for symmetric systems) may be used. The core matrix-vector product w=(PL1APR1)vw = (P_L^{-1}AP_R^{-1})v is implemented in three phases:

  1. Pre-apply right preconditioner: Transform vv by PR1P_R^{-1} using joint FST, scaling, and inverse FST: cost O(JNlogJ)O(JN\log J).
  2. Apply system matrix AA: Multiply by LaINL_a\otimes I_N and IJTI_J\otimes T efficiently. Cost for spatial multiplication is O(Nnnz(La))=O(NJ)O(N\,\mathrm{nnz}(L_a))=O(NJ) (if LaL_a is sparse); temporal multiplication via Toeplitz structure is O(JNlogN)O(JN\log N).
  3. Pre-apply left preconditioner: Transform by FST, invert the block Toeplitz systems TiT_i in parallel for each spatial index (cost O(JNlogN)O(JN\log N)), and transform back. All JJ temporal solves are independent and can be distributed for ideal PinT scaling.

The total cost per matrix-vector product is O(JN(logJ+logN))O(JN(\log J + \log N)), with storage dominated by the JNJN all-at-once vector and JNJN entries for first columns of Ti1T_i^{-1}. Only O(J)O(J) storage for Λ\Lambda is also needed. FFT-based routines enable in-place application of QQ.

5. Theoretical Properties and Robustness

Under the assumption

aminIJL11/2LaL11/2amaxIJ,a_{\min} I_J \preceq L_1^{-1/2} L_a L_1^{-1/2} \preceq a_{\max} I_J,

the two-sided preconditioned matrix

M:=PL1APR1M := P_L^{-1} A P_R^{-1}

has condition number

κ2(M)amaxamin,\kappa_2(M) \leq \frac{a_{\max}}{a_{\min}},

independent of JJ and NN. This establishes robustness with respect to both problem size and discretization. The proof proceeds via bounding Rayleigh quotients using block-Kronecker inequalities and the positive-definiteness of the temporal matrices.

6. Implementation Complexity and Practical Considerations

Each application of the two-sided preconditioners has computational cost O(JNlog(JN))O(JN\log (JN)). If PP processors are available, distributing the JJ independent temporal solves in Phase 3 achieves near-ideal PinT scaling. Storage requirements are minimal since only the diagonals of the block systems and the first columns of the Toeplitz inverses are needed.

If the underlying spatial coefficient a(x)a(x) varies slowly, it suffices to recompute scaled TiT_i blocks as λi\lambda_i changes, requiring at most O(JNlogN)O(JN\log N) per update. In the special case a(x)consta(x) \equiv \mathrm{const}, the matrix AA coincides with PP, and the two-sided preconditioner becomes a direct solver with complexity O(JNlog(JN))O(JN\log(JN)).

The method exploits the precise algebraic structure of the discretized system (Toeplitz in time, diagonalizable in space) without further approximation and is matrix-size robust in terms of conditioning. These properties distinguish it from classical one-sided and approximate preconditioners for similar equations.

7. Applications and Extensions

Two-sided preconditioners in the manner of Lin–Ng–Zhi (2020) are applicable in the fast iterative or direct solution of discretized non-local evolutionary equations with variable coefficients and weakly singular temporal kernels. They enable efficient PinT solvers for large-scale time-dependent PDEs with guaranteed convergence rates and parallel scalability. The approach is well suited to cases where temporal non-locality and spatially varying diffusion preclude straightforward block-diagonalization or classical (block) circulant preconditioners. A plausible implication is that similar splitting strategies may generalize to other hybrid-structured systems where distinct algebraic properties in different dimensions can be simultaneously exploited for parallelism and conditioning control (Lin et al., 2021).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Two-Sided Preconditioner Updates.