Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 67 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 15 tok/s Pro
GPT-5 High 15 tok/s Pro
GPT-4o 96 tok/s Pro
Kimi K2 121 tok/s Pro
GPT OSS 120B 440 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

ReLU Fan: Polyhedral Geometry in Neural Networks

Updated 10 September 2025
  • ReLU Fan is a polyhedral complex that partitions the input space into affine regions based on constant ReLU activation patterns.
  • It translates neural network function realization into toric and tropical geometry, linking combinatorial structures with algebraic invariants.
  • The geometric framework provides practical criteria for exact function realizability and guides analyses of network expressivity and identifiability.

A ReLU fan is the canonical polyhedral complex—termed a "fan" in toric geometry—arising from the piecewise linear structure of a feedforward ReLU neural network acting on Rn0\mathbb{R}^{n_0}. The ReLU fan is constructed from the regions of the input space on which the composite function of the network behaves affinely due to the gating effect of the ReLU activations. By recasting neural network function realization in terms of the combinatorics and geometry of the associated fan, the theory exploits algebraic and tropical geometry to provide necessary and sufficient criteria for the exact representability of piecewise linear functions by ReLU networks of a fixed architecture (Fu, 7 Sep 2025).

1. Definition and Geometry of the ReLU Fan

The ReLU fan Σf\Sigma_f associated with a network fθf_\theta is a rational polyhedral complex in Rn0\mathbb{R}^{n_0} corresponding to the partitioning of the input space into regions where the network is affine linear. Formally, each region (cell, cone) of Σf\Sigma_f is defined by the set of input points where the activation pattern of all ReLU neurons is constant. In each such cell, the network’s output admits the representation: fθ(x)=mσ,x+bσ,xσf_\theta(x) = \langle m_\sigma, x \rangle + b_\sigma, \quad x \in \sigma where mσm_\sigma is the slope vector and bσb_\sigma the offset associated with the region σ\sigma.

This structure encodes all non-linearities induced by the ReLU activation: each bent hyperplane corresponding to a neuron is reflected as a wall (codimension-1 face) of the fan, and the intersection patterns of these hyperplanes determine the combinatorics of the fan.

2. Toric Variety and the ReLU Cartier Divisor

Each ReLU fan Σf\Sigma_f defines a toric variety XΣfX_{\Sigma_f} (complete and proper, as the fan covers all of Rn0\mathbb{R}^{n_0}). The output function ff of the network naturally serves as a support function for a (rational) Cartier divisor DfD_f on XΣfX_{\Sigma_f}, with slope data (mσ)σΣf(m_\sigma)_{\sigma \in \Sigma_f}, uniquely determined up to addition of a global affine function.

The divisor DfD_f encodes the bending at the walls of Σf\Sigma_f and is central to the expressivity and realization problem. Specifically, the divisor’s intersection numbers with torus-invariant curves (corresponding to walls between regions) quantify how ff “bends” as one transitions across hyperplane boundaries.

3. Tropical Geometry Connection

ReLU functions are max-affine and, thus, are tropical polynomials when expressed as: f(x)=maxi{ci+ui,x}f(x) = \max_{i}\{c_i + \langle u_i, x \rangle\} in tropical arithmetic (=max\oplus = \max, =+\odot = +). The Newton polytope Newt(f)\operatorname{Newt}(f) is the convex hull of the exponent vectors {ui}\{u_i\}.

A central result establishes that the polytope PDP_{-D} associated to the negative of the ReLU Cartier divisor equals minus the Newton polytope: PD=Newt(f)P_{-D} = -\operatorname{Newt}(f) Thus, tropical convexity of ff aligns with convexity conditions on the divisor, allowing algebraic and combinatorial tools to analyze network expressivity.

4. Exact Function Realization Problem

The criterion for a piecewise linear function ff to be realizable exactly by an unbiased shallow ReLU network (one hidden layer, rational weights) can be stated in terms of intersection numbers:

  • Extend the walls (codimension-1 loci of nonlinearity of ff) to full hyperplanes, yielding a fan Σ\Sigma.
  • The output function’s divisor DfD_f must satisfy: DfV(τ1)=DfV(τ2)D_f \cdot V(\tau_1) = D_f \cdot V(\tau_2) for any two walls τ1,τ2\tau_1, \tau_2 lying on the same full hyperplane, where V(τ)V(\tau) denotes the torus-invariant curve crossing the wall τ\tau.

If these conditions are met, then ff is realizable (up to an affine shift) by the network.

5. Intersection Theory and Algebraic Invariants

Intersection numbers mσmσ,u\langle m_\sigma - m_{\sigma'}, u \rangle—where uu is the generator of the shared face between adjacent cones σ,σ\sigma, \sigma'—measure the “bending intensity” across each wall. The criterion above requires uniform intersection numbers for all walls arising from the same neuron (hyperplane), reflecting consistent activation behavior. The volume of the associated line bundle OX(D)\mathcal{O}_X(-D) is equal to the mixed volume of the Newton polytope: Vol(Newt(f))=Vol(OX(D))\operatorname{Vol}(\operatorname{Newt}(f)) = \operatorname{Vol}(\mathcal{O}_X(-D)) giving a direct geometric measure of network capacity.

6. Algorithmic and Combinatorial Structure

From a combinatorial geometry perspective, the ReLU fan provides complete information about the partitioning of input space, the arrangement of bent hyperplanes, and the possible regions of affine behavior. Enumerating activation patterns (as studied in learning-theoretic analyses of “ReLU fan” combinatorics (Bakshi et al., 2018)) allows algebraic geometric reasoning to be directly applied to understanding network identifiability, expressivity, and learning hardness.

In analysis and learning of network weights, the correspondence between fan combinatorics, divisor data, and activation patterns underpins both theoretical results (e.g., hardness criteria, exact recovery) and practical algorithms for recovery, clustering, and interpretability.

7. Significance for Neural Network Theory

The toric geometry perspective via the ReLU fan synthesizes algebraic geometry, tropical geometry, and deep learning. It translates the function realization problem to geometric invariants, facilitates classification of realizable functions, and emphasizes the foundational role of polyhedral combinatorics in neural network expressivity. Furthermore, it unites tropical and algebraic viewpoints for piecewise linear models, allowing both rigorous mathematical analysis and practical application in architecture certification, clustering-based interpretation, and function recovery.

In summary, the ReLU fan is the polyhedral complex governing the affine regions of a ReLU neural network, serving as the bridge between network architecture and toric geometry. It encodes all necessary combinatorial, algebraic, and geometric data to determine exact function realizability, supports the translation of tropical polynomials to toric Cartier divisors, and underlies both theoretical and algorithmic results on the expressive capacity and identifiability of ReLU networks (Fu, 7 Sep 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to ReLU Fan.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube