- The paper introduces the functional variety concept to measure network expressiveness by quantifying the dimension of representable polynomial functions.
- It employs algebraic geometry tools and backpropagation over polynomial rings and finite fields to rigorously analyze and compute network properties.
- The study identifies a bottleneck effect with narrow intermediate layers, linking network architecture to optimization challenges and tensor decompositions.
This paper investigates the expressive power of deep neural networks that use polynomial activation functions. Unlike general non-linear activations, using a polynomial activation (specifically, raising inputs to the r-th power) makes the function computed by the network an algebraic mapping from weights to polynomials. This algebraic structure allows the authors to leverage tools from algebraic geometry to analyze the set of functions a network can represent.
The core concept introduced is the "functional variety" (Vd,r). For a fixed network architecture d=(d0,…,dh) and activation degree r, the set of all functions achievable by varying the network weights forms a set of polynomials, which is a subset of a finite-dimensional vector space of polynomials (specifically, dh output polynomials, each being a homogeneous polynomial of degree rh−1 in d0 input variables). The Zariski closure of this set is the functional variety Vd,r. The dimension of this variety is proposed as a precise measure of the network's expressive power, quantifying the degrees of freedom in the functions it can represent. While the actual functional space Fd,r might be smaller than its Zariski closure Vd,r in the standard Euclidean topology, their dimensions are equal.
The paper establishes a connection between the functional variety's properties and the network's optimization landscape. If the functional variety Vd,r "fills" the ambient space of polynomials (meaning its dimension equals the dimension of the polynomial space Symrh−1d0dh), the actual functional space Fd,r is "thick" (has positive Lebesgue measure). The authors prove that if the functional space is not thick, a convex loss function can exist with arbitrarily bad local minima (Proposition 3.1, 3.2). Conversely, architectures with thick/filling functional spaces are expected to have more favorable optimization properties.
The structure of polynomial networks is also related to tensor and polynomial decompositions. Shallow networks (h=2) are shown to compute partially symmetric tensors that can be expressed as a sum of d1 rank-1 terms, connecting to CP tensor decomposition. Deep polynomial networks correspond to an iterated form of such decompositions. The paper links the ability of a network to "fill" the ambient polynomial space to properties of these tensor/polynomial decompositions.
Key theoretical results focus on determining the dimension of Vd,r:
- Upper Bound: The dimension is bounded by the minimum of the number of parameters in the network, accounting for symmetries (∑didi−1−∑di=dh+i=1∑h(di−1−1)di), and the dimension of the ambient polynomial space dh(rh−1d0+rh−1−1) (Theorem 4.1).
dimVd,r≤min(dh+i=1∑h(di−1−1)di,dh(rh−1d0+rh−1−1))
- High Degree Limit: For a fixed architecture, the dimension achieves this upper bound related to the number of parameters when the activation degree r is sufficiently high (Theorem 4.1).
- Recursive Bound: The dimension of a deep network can be bounded based on the dimensions of shallower networks forming its parts (Proposition 4.3):
dimV(d0,…,dh),r≤dimV(d0,…,dk),r+dimV(dk,…,dh),r−dk
- Bottleneck Property: The paper proves that an intermediate layer width di=2d0−2 acts as an "asymptotic bottleneck". This means that regardless of how wide other layers or how deep the network becomes, an intermediate layer with this width (or smaller) prevents the functional variety from filling the ambient space (Theorem 4.4).
To paper these dimensions computationally, the authors developed methods based on the fact that the dimension of the functional variety equals the generic rank of the Jacobian matrix of the parameter-to-function map Φd,r (Lemma 3.3). They implemented this in SageMath, leveraging automatic differentiation (backpropagation).
- Backpropagation over Polynomial Ring: Define the network operations using polynomials in the input variables x1,…,xd0. Backpropagating the output polynomials with respect to network weights yields polynomials whose coefficients form the Jacobian entries.
- Backpropagation over Finite Field: Perform backpropagation over a finite field Z/pZ at several random input points x. The Jacobian entries can be recovered by solving a linear system. This approach is generally faster.
These computational experiments (examples shown in Tables 1 and 2) provide empirical insights:
- The dimension of the functional variety stabilizes as the activation degree r increases, consistent with the theoretical result.
- Minimal architectures that fill the ambient space tend to have intermediate layer widths that are unimodal (increase then decrease) as depth increases. This computational finding leads to a conjecture that minimal filling architectures are unimodal.
The paper proposes polynomial networks as a testbed for understanding general deep nonlinear networks, arguing that insights gained here could transfer due to polynomial approximation properties (Stone-Weierstrass theorem). Future work could involve studying the geometry of functional varieties beyond dimension (e.g., degree), exploring other algebraic architectures (convolutions, different polynomial activations), and developing algorithmic methods for learning based on tensor decomposition principles.
In summary, the paper provides a rigorous algebraic framework for studying the expressiveness of polynomial neural networks, defines expressiveness via the dimension of the functional variety, connects this dimension to optimization properties and tensor decompositions, provides theoretical results on dimension and bottleneck widths, and demonstrates practical computational methods for dimension calculation using standard ML techniques like backpropagation adapted to algebraic settings. The findings offer insights into how network architecture, particularly layer widths and activation degrees, impacts the space of functions a network can represent.