Sparse-TDA via QR Pivoting
- The paper presents a sparse TDA approach that leverages pivoted QR factorization to select informative pixels and reduce feature dimensions efficiently.
- It combines persistent homology, persistence images, and truncated SVD to convert complex topological data into actionable features for machine learning.
- Empirical results show that Sparse-TDA significantly lowers computation time while maintaining competitive accuracy compared to kernel-based TDA methods.
Sparse-TDA via QR Pivoting is a method for sparse realization of topological data analysis (TDA) in supervised multi-way classification settings. It combines persistent topological feature extraction with principled column selection via pivoted QR factorization, enabling efficient, information-preserving dimension reduction for subsequent machine learning tasks. Key components include the vectorization of persistence diagrams (PDs) into persistence images (PIs), low-rank approximation via truncated SVD, and sparse pixel selection using QR pivoting over the PI domain. Benchmark evaluations indicate that Sparse-TDA maintains competitive accuracy with significant computational savings compared to kernel-based TDA approaches (Guo et al., 2017).
1. Persistent Topological Feature Vectorization
Persistent homology summarizes topological features of data through PDs, representing birthādeath pairs corresponding to the lifetime of topological structures. To enable statistical learning, these diagrams are transformed into vector-valued features:
- Linear Coordinate Transformation: Each PD undergoes the map , so coordinates represent birth time and persistence.
- Persistence Surfaces: Define
where is a weighting (e.g., linear or nonlinear ), and is a Gaussian with variance centered at .
- Persistence Images (PI): is discretized on a grid; each cellās integral forms a component of the vector with .
- Data Matrix Assembly: The PI representations for samples form .
This process yields a standardized, high-dimensional encoding of persistent topological information suitable for downstream processing.
2. Low-Rank Approximation via SVD
Due to the invariance and redundancy in the PI representations, the data matrix admits a low-rank structure:
- Economy-Size SVD: Compute , where , , and holds the top-r singular values.
- Rank Selection: The optimal is determined by the GavishāDonoho rule: retain all exceeding times the median (), providing a near-minimax approximation in the presence of Gaussian noise.
This step isolates the dominant variance directions, yielding a subspace with maximal topological signal representation.
3. QR Pivoting for Sparse Pixel Selection
Sparse-TDA achieves drastic reduction in feature dimensionality via column subset selection:
- Pivoted QR Algorithm
- Input and target sparsity .
- Compute the pivoted QR of : , with the permutation matrix determined by BusingerāGolub norm-maximal pivoting.
- The first pivots identify columns (pixels) most informative for the dominant subspace.
- For each sample, PI vectors are restricted to these elements: .
This mechanism provides a principled, data-driven way to select a subset of PI grid locations most critical for information retention.
4. Computational Complexity
The overall efficiency of Sparse-TDA is determined by:
| Stage | Dominant Complexity | Remarks |
|---|---|---|
| Feature Extraction (PDāPI) | : simplices per data item | |
| Truncated SVD | Efficient via randomized methods | |
| Pivoted QR | BusingerāGolub algorithm | |
| SVM Training | Kernel method, feature dim. |
The SVD and QR stages dominate the pipeline. Sparse-TDA delivers computational savings primarily by reducing the SVM input dimensionality, typically setting .
5. Multi-Way Classification Procedure
Once the sparse features are extracted, classification proceeds as follows:
- Sparse Feature Assembly: All training PIs are projected to via QR-selected indices.
- RBF-kernel SVM Training: Use to train a one-versus-rest SVM; the RBF kernel is .
- Parameter Selection: Grid search and cross-validation are performed for SVM and kernel width .
- Decision Function:
where are the learned dual and bias parameters.
This pipeline integrates topological feature selection and statistical classification within a unified, sparse framework.
6. Empirical Evaluation and Benchmarking
Sparse-TDA was validated on human posture and texture image classification problems:
- SHRECā14 Synthetic: 15 classes, 300 meshes; Sparse-TDA (NW) accuracy: 94.0%, SVM runtime: 25.6s.
- SHRECā14 Real: 400 meshes; Sparse-TDA (LW) accuracy: 68.8%, SVM runtime: 82.2s.
- Outex Texture: 24 classes, 480 images; Sparse-TDA (NW) accuracy: 66.0%, SVM runtime: 120s.
Key findings:
- Sparse-TDA achieves runtime reductions of one to two orders of magnitude compared to kernel-based TDA, with a modest loss in accuracy.
- Sparse-TDA consistently outperforms L1-SVM in accuracy, often with equal or better training times.
| Dataset | L1-SVM Accuracy (%) | Sparse-TDA (LW) | Sparse-TDA (NW) | Kernel-TDA |
|---|---|---|---|---|
| SHREC Synthetic | 89.6 | 91.5 | 94.0 | 97.8 |
| SHREC Real | 63.9 | 68.8 | 67.8 | 65.3 |
| Outex Texture | 55.1 | 62.6 | 66.0 | 69.2 |
| Dataset | L1-SVM Time (s) | Sparse-TDA | Kernel-TDA |
|---|---|---|---|
| SHREC Synthetic | 35.4 | 25.6 | 1182 |
| SHREC Real | 305 | 82.2 | 92.3 |
| Outex Texture | 106 | 120 | 5457 |
Performance reflects mean ± std over multiple data splits; PI grid, SVD rank , and QR sample were dataset-specific (; ) (Guo et al., 2017).
7. Theoretical Properties and Stability
- Persistence Image Stability: The PD-to-PI transform is 1-Wasserstein stable: small input perturbations induce bounded changes in the resulting PI (Adams et al., 2017).
- Optimal SVD Truncation: The GavishāDonoho median threshold guarantees near-optimal rank selection for signal-plus-Gaussian-noise matrices.
- QR Pivot Conditioning: The BusingerāGolub strategy ensures the selected columns form a numerically well-conditioned basis. The associated interpolation error is bounded by the th pivot growth factor (DrmaÄ & Gugercin, 2015).
These theoretical properties guarantee robust, stable, and near-minimax realization of topological feature sampling, ensuring generalization and numerical reliability across diverse datasets (Guo et al., 2017).