Polynomial Coefficient Decomposition
- PCD is a framework that represents polynomial maps as structured sums over lower-complexity components, such as rank-one terms or powers of linear forms.
- It provides efficient methods for decomposition, identifiability analysis, and rank bounding, leveraging tools like catalecticants and secant variety theory.
- PCD underpins applications in signal processing, machine learning, and neural architectures by enabling scalable recovery of polynomial and tensor structures.
Polynomial Coefficient Decomposition (PCD) is a unifying framework for expressing polynomials and polynomial maps as structured sums over lower-complexity building blocks, typically rank-one terms or powers of linear forms. PCD generalizes classical tensor decompositions (e.g., CP, symmetric/Waring), algorithmic approaches in polynomial algebra, and recent applications in signal processing, machine learning, algebraic geometry, and graph neural networks. It rigorously addresses identifiability, generic and maximal rank, algorithmic recovery, and parameter efficiency for a broad class of models, with particular relevance to the decomposition of polynomial maps, the analysis of secant varieties, and multilinear architectures.
1. Formal Definition and Decoupled PCD Model
A polynomial map (where or , ) admits an -term decoupled Polynomial Coefficient Decomposition if there exist vectors , and univariate polynomials such that
Equivalently, in matrix form: where and (Comon et al., 2016).
Specializations of the PCD model encompass:
- Matrix factorization (),
- Ridge polynomial sums for scalar output (),
- Simultaneous Waring decompositions for homogeneous of degree .
2. Algebraic Geometry: X-Rank and Veronese Scroll
PCD extends tensor decompositions by interpreting the decomposed structures via X-rank on projective varieties. Let , and be a scale-invariant, irreducible, nondegenerate algebraic cone. Given , its X-rank is
The minimal such expression is an X-rank decomposition. Critical to PCD is the Veronese scroll variety , defined as the image of
typically with . For the full PCD model, one considers the Segre product , and studies secant varieties to analyze rank and identifiability (Comon et al., 2016).
3. Rank Bounds and Identifiability
The generic and maximal rank for PCD admits tight, dimension-dependent bounds. For , ,
with maximal rank
For scalar-output ridge-sum (), this improves the classical symmetric tensor bound .
Identifiability is established as follows: is -identifiable if, for general , the decomposition is unique up to permutation. Sufficient identifiability holds for
with the minimum of the dimension-based weak-defectivity bound and the scroll dimension (Comon et al., 2016).
Partial identifiability is guaranteed for coefficients of degree , provided does not exceed .
4. Computational Methods for Decomposition
Recovery of the PCD involves explicit algebraic constructions:
- Linearization via stacked catalecticant matrices whose minors characterize the Veronese scroll,
- Solving for by projection onto top-degree slices and simultaneous Waring decomposition leveraging classic identifiability results for Segre-Veronese varieties,
- Subsequent recovery of lower-degree coefficients by linear systems conditioned on full-rank catalecticant blocks (Comon et al., 2016).
For homogeneous polynomials or symmetric tensors, classical algorithms include Hankel moment matrices, apolar ideals, quotient algebras, multiplication operator simultaneous diagonalization, and solving for weights by coefficient matching (Laface et al., 2021). The Sylvester/catalecticant, Koszul flattening, and bundle methods provide practical procedures, with computable bounds on rank and uniqueness (Oeding et al., 2011).
For non-homogeneous power-sums of degree- polynomials , span extraction via partial derivatives and "desymmetrization" enables polynomial complexity algorithms that are stable to inverse-polynomial noise and work with generic up to when (Bafna et al., 2022).
5. Applications in Learning and Signal Processing
PCD underpins architectural design in spectral graph convolutional networks (SGCNs), where spectral filters are polynomials of a graph Laplacian or adjacency. The coefficient tensor admits classical CP and Tucker decompositions:
yielding SGCN layers with significant parameter reductions and controlled expressivity. Empirically, CoDeSGC-CP and -Tucker outperform baseline polynomial convolutions on classification tasks, particularly in settings with heterophilic graphs and medium-sized datasets (Huang et al., 2024).
6. Integer Decomposition and Bitwise Factorization
The "binary method of integer decomposition" formulates factorization of odd integers as a search for polynomials with binary coefficients, such that . Matching coefficients leads to systems of quadratic (and after reduction, linear) equations, solved deterministically via bitwise carry analysis. The worst-case cost is , and while not competitive with subexponential factoring, this method illustrates the algebraic generality of PCD in discrete settings (Gao, 2018).
7. Algebraic-Geometric Tools and Theoretical Techniques
Key concepts in the theoretical analysis of PCD include:
- Secant varieties and their dimension/defectivity properties,
- The Alexander-Hirschowitz theorem for the generic rank of symmetric tensors,
- Weak-defectivity and Terracini’s lemma to establish identifiability bounds,
- Koszul flattenings, catalecticants, and inheritance properties of ideals in projective geometry,
- Eigenvector methods for rank-one extraction in symmetric tensors (Comon et al., 2016, Oeding et al., 2011).
These techniques equip PCD with sharp criteria for uniqueness, robust dimension-dependent rank bounds, and effective algebraic/numerical recovery mechanisms.
A plausible implication is that PCD serves as a common language for structured decompositions across polynomial algebra, tensor methods, algebraic geometry, numerical algorithms, and neural architectures. It continues to motivate both foundational research (e.g., in secant variety theory, identifiability conditions, and stability analysis) and algorithmic innovation targeting scalable and interpretable models in data science.