Bi-Fidelity Interpolative Decomposition
- Bi-Fidelity Interpolative Decomposition is a non-intrusive, sample-based reduction technique that approximates high-fidelity outputs using inexpensive low-fidelity simulations with carefully selected high-fidelity samples.
- The method employs a greedy, pivoted QR strategy to extract a low-rank representation, enabling quick construction of surrogates for uncertainty quantification and design space exploration.
- Rigorous error analysis and convergence studies ensure that the surrogate maintains near high-fidelity accuracy while significantly reducing computational cost.
Bi-fidelity interpolative decomposition (ID) is a non-intrusive, sample-based model reduction technique for accurately approximating the outputs of high-fidelity models using many inexpensive low-fidelity (LF) simulations and a small set of high-fidelity (HF) runs. This decomposition exploits the empirical low-rank structure often observed in parametric or stochastic numerical simulations, enabling construction of surrogates for uncertainty quantification, design space exploration, and related many-query tasks. The resulting bi-fidelity model is characterized by an accuracy close to that of the HF model and a computational cost comparable to the LF model, assuming the rank structure is sufficiently sharp (Hampton et al., 2017).
1. Bi-Fidelity Interpolative Decomposition: Formulation and Construction
The central assumption of the bi-fidelity ID approach is that the mapping from parameters (or uncertain inputs) to the quantity of interest (QoI), denoted , is well-approximated by a low-rank linear combination of simulation outputs at a small set of parameter configurations. Specifically, the representation reads
where the basis vectors are simulations at carefully chosen parameters. The LF data is aggregated into a matrix , and a rank- interpolative decomposition is performed: Here, contains the selected skeleton columns, and is the coefficient (interpolation) matrix. The selection of the basis indices is performed by a greedy, pivoted QR strategy, maximizing the representativeness of .
To "lift" this construction to the HF space, is replaced by —the high-fidelity solutions at the same parameter values—to yield the bi-fidelity approximation: This means that only expensive HF runs are needed (at the selected parameters), while the interpolation rule derived from the cheap LF model applies throughout the domain.
2. Low-Rank Structure and Basis Selection
The bi-fidelity framework is predicated on the existence of rapid decay in the singular values of the LF data matrix , indicating approximate low-rankness. This property ensures:
- Compact representation: The majority of the variability in (and, crucially, in ) is captured by the selected basis.
- Efficiency: The cost of identifying the skeleton columns and building via QR with column pivoting is for and typically .
- Transferability: Once the skeleton indices are determined from , only HF runs (columns of ) must be computed, conferring significant computational savings.
The effectiveness of this decomposition hinges on the low-rank similarity between LF and HF data; if the LF basis fails to approximate the HF solution subspace, the bi-fidelity approach is suboptimal.
3. Rigorous Error Analysis
A key result is the derivation of a computable, pragmatic error bound for the bi-fidelity surrogate. The approach models the HF data as , with a lifting operator and an error term. Introducing the Gramian mismatch,
for , the main bound is: where
and denotes the th singular value of . The first term in the bound accounts for approximation error in the LF low-rank model, while the second depends on the energy mismatch between and . The computational cost of this error estimation is low, requiring Gramian approximations at only slightly more than HF points.
The theoretical analysis rigorously quantifies the tradeoff between the quality of the LF basis and the representational error in the HF space; the error decays provided the basis is sufficiently robust and the models are closely coupled. If the LF mesh is refined or the LF model is improved to better emulate the HF system, both and the bi-fidelity error decrease, with numerical evidence showing as optimal with improved LF models.
4. Convergence and Robustness Properties
The convergence properties of the bi-fidelity ID are governed by:
- The decay rate of the singular values of : Fast decay yields rapid convergence.
- The lifting operator : Its norm, along with the Gramian proximity parameter , controls the bound tightness.
- The number of HF samples: Empirically, a number slightly exceeding the interpolation rank is sufficient for robust estimation of error bounds and constructing the HF surrogate.
- Model hierarchy: As the LF model more accurately resolves the physics, the optimal parameter approaches unity, and the Gramian mismatch becomes negligible.
In all practical settings considered, only a small set of representative HF samples is needed for surrogate construction and error certification, indicating scalability as the problem dimension grows or the parameter space expands.
5. Applications in Uncertainty Quantification and Design
Two principal numerical examples demonstrate the efficacy of the method for uncertainty quantification (UQ):
- Heat-driven cavity flow: The quantity of interest is the steady-state heat flux along the cavity's hot wall, with uncertainties modeled via a Karhunen–Loève expansion and parameterized viscosity. High-fidelity solutions (e.g., mesh) are "lifted" from low-fidelity outputs (down to meshes). The bi-fidelity surrogate with basis vectors achieves error levels much lower than standalone LF, closely tracking the HF outputs for both mean and variance estimation, as verified via error histograms.
- Composite beam: The maximal vertical displacement under uncertain load is predicted using a bi-fidelity model that combines classical beam theory (LF) with a finite-element high-fidelity simulation. Even with only basis vector, the bi-fidelity surrogate surpasses LF performance by an order of magnitude in RMSE.
In both cases, effectiveness is demonstrated not just for mean predictions but also for higher moments (variance), with practical error bounds predicting real errors accurately—with empirical robustness to the number and selection of HF samples.
6. Comparison to Classical High- and Low-Fidelity Models
The essence of the bi-fidelity ID is to interpolate the accuracy gap between HF and LF models:
- HF models are computationally infeasible for many-query applications due to high simulation cost.
- LF models enable large sampling but at the price of lower accuracy and limited quantitative predictive power.
- Bi-fidelity surrogates blend the advantages: only a small set of HF runs is needed, and the LF sampling is used everywhere else, inheriting both speed and a quantifiable degree of HF accuracy. The error estimate ensures that the model does not underperform reliably given a sufficiently good LF representation and close proximity in the lifted subspaces.
Quantitatively, the computational savings can be dramatic: error levels an order of magnitude lower than LF-only estimates attained at costs well below those of HF-only approaches, and comparable to the cost of the LF evaluations.
7. Practical Workflow and Implementation Considerations
Implementation proceeds as follows:
- Sampling: Run the LF model on a large sample of parameter values; assemble the data matrix .
- Decomposition: Apply rank-revealing QR to obtain and .
- HF acquisition: Run the HF model at the selected parameter values (those corresponding to ).
- Surrogate construction: Compute to reconstruct the HF solution for any parameter via the LF-derived interpolation rule.
- Error estimation: Use a small set of additional HF runs to estimate Gramian mismatches and certify output accuracy using the derived bounds.
This workflow is strictly non-intrusive: no modifications to existing simulators are required, and the approach generalizes across physics and system types, provided the low-rank hypothesis holds.
Bi-fidelity interpolative decomposition thus constitutes a rigorous, low-cost, and widely applicable approach to surrogate construction, underpinned by provable error certificates and demonstrated effectiveness in uncertainty quantification tasks involving both linear and nonlinear PDE models (Hampton et al., 2017).