Papers
Topics
Authors
Recent
Search
2000 character limit reached

HIENet: Hybrid Invariant–Equivariant Networks

Updated 21 February 2026
  • HIENet is a neural architecture that explicitly decomposes data representations into invariant and equivariant components under group actions.
  • It leverages invariant encoders, equivariant encoders, group predictors, and decoders to separate and reassemble latent codes effectively.
  • This design achieves state-of-the-art performance across diverse applications such as 3D vision, materials science, and graph learning while maintaining computational efficiency.

A Hybrid Invariant–Equivariant Network (HIENet) is a neural architecture that explicitly decomposes data representations into invariant and equivariant components under a group action. By jointly leveraging both invariance (features unchanged under group actions) and equivariance (features that transform predictably), HIENets optimize the balance between inductive bias, computational efficiency, and representational expressivity. This design paradigm has been instantiated in unsupervised learning, generative models, self-supervised learning, geometric deep learning, and large-scale materials science models.

1. Mathematical Foundations and Latent Code Decomposition

The foundational principle of HIENet is to factor data representations relative to a group GG acting on the input space XX. Given a function f:XVf: X \rightarrow V and a group representation ρ:GGL(V)\rho: G \rightarrow \operatorname{GL}(V), equivariance requires

f(gx)=ρ(g)f(x),gG,xX.f(g \cdot x) = \rho(g) f(x), \quad \forall\, g \in G,\, x \in X.

HIENet structures the latent code as a direct product: x(η,ψ)(zinv,g)Zinv×G,x \overset{(\eta, \psi)}{\longmapsto} (z_\text{inv},\, g) \in Z_\text{inv} \times G, with η:XZinv\eta: X \rightarrow Z_\text{inv} enforcing

η(gx)=η(x),\eta(g \cdot x) = \eta(x),

and ψ:XG\psi: X \rightarrow G predicting the group action. The architecture further introduces an equivariant intermediate map μ:XY\mu: X \to Y with

μ(gx)=ρY(g)μ(x),\mu(g \cdot x) = \rho_Y(g)\, \mu(x),

and a canonicalization step ξ:YG\xi: Y \rightarrow G yielding g=ξ(μ(x))g = \xi(\mu(x)). The total latent space is thus Z=Zinv×GZ = Z_\text{inv} \times G. This structure enables universal expressivity for GG-equivariant maps via a provable correspondence between equivariant and (subgroup-)invariant functions (Sannai et al., 2024).

2. Network Architectures and Module Interactions

A typical HIENet instantiates the following modules (Winter et al., 2022, Yan et al., 25 Feb 2025, Garrido et al., 2023):

  • Invariant Encoder η\eta: Implements η(gx)=η(x)\eta(g \cdot x) = \eta(x); realized via invariant pooling, DeepSets, or scalar-valued steerable CNN/GNN blocks.
  • Equivariant Encoder μ\mu: Extracts features transforming by ρY(g)\rho_Y(g); typically vector/tensor channels in steerable or spherical harmonics GNNs/CNNs.
  • Group Predictor ψ=ξμ\psi = \xi \circ \mu: Determines the group element aligning μ(x)\mu(x) to a standard reference.
  • Decoder δ:ZinvX\delta:Z_\text{inv}\rightarrow X: Reconstructs a canonical element, to which the pose is reapplied via gg.
  • Final Alignment: The decoded object is transformed using the predicted group element: x~=ρ(ψ(x))δ(η(x))\tilde{x} = \rho(\psi(x))\,\delta(\eta(x)).

More complex hybridizations interleave pure invariant and equivariant blocks (e.g., scalar-only transformer layers followed by O(3)O(3)-equivariant message-passing), interspersed with invariant pooling, equivariant broadcasts, or group-canonicalization layers (Sannai et al., 2024, Yan et al., 25 Feb 2025). All components are fully differentiable.

3. Learning Objectives and Theoretical Guarantees

Training in HIENet frameworks is generally unsupervised and enforced via a single reconstruction-driven objective: Lrec=ExD[d(x,ρ(ψ(x))δ(η(x)))],L_\text{rec} = \mathbb{E}_{x \sim D}\left[d\left(x, \rho(\psi(x))\, \delta(\eta(x))\right)\right], where dd is a task-appropriate metric (e.g., 2\ell_2 for images/point clouds). Group-alignment terms can be included for parameterized groups if weak labels or proxies for gg are available, but HIENet is designed for situations with no group supervision.

The invariance of η\eta and equivariance of μ\mu ensure decomposition and disentanglement of group-invariant and group-variant factors. For permutation, rotation, translation, and Euclidean groups, explicit constructions of ξ\xi and ρ\rho are provided to guarantee theoretical correctness up to the stabilizer subgroup GxG_x.

The universal approximation theorem (Sannai et al., 2024) states that HIENets, built from universal invariant subnetworks for each HiH_i (stabilizer subgroup), are themselves universal for GG-equivariant maps. The depth is preserved up to one additional layer for the reassembly, and parameter complexity is controlled by the number and size of GG-orbits and their stabilizers.

4. Explicit Group-Specific Implementations and Instantiations

HIENet construction is group-agnostic and can be tailored as follows (Winter et al., 2022, Garrido et al., 2023, Yan et al., 25 Feb 2025):

  • SO(2)/SO(3): Utilize steerable CNNs/GNNs with vector and tensor capsules. μ(x)\mu(x) outputs (e.g., S1/S2 on the sphere) are orthonormalized, and group action is recovered via argmax/atan2 or SVD.
  • SE(n): Separate translation (vector) and rotation (matrix) channels; pose reapplication via rigid transformation.
  • S_N (permutations): DeepSets for η\eta, permutation-invariant multihead pooling, and group inference via soft-argsort to build doubly stochastic mappings.
  • Materials science models: A single invariant message-passing block (G-Transformer) precedes multiple O(3)O(3)-equivariant message-passing layers; readout preserves energy invariance and force/stress equivariance (Yan et al., 25 Feb 2025).
  • Self-supervised 3D vision: Latent spaces are split into yinvy_\text{inv} and yequiy_\text{equi}, with invariant and equivariant projection heads and a hypernetwork-based group action on the equivariant branch (Garrido et al., 2023).

5. Experimental Results, Benchmarks, and Efficiency

HIENets have demonstrated state-of-the-art or near-optimal performance across domains:

  • Vision (Rotated MNIST, ShapeNet): HIENets with steerable CNN backbones achieve RMSE $0.016$ on upright-to-rotated transfer, with latent-invariant codes yielding 90%90\% digit accuracy vs 68%68\% for classical autoencoders (Winter et al., 2022).
  • Sets and Permutations: On large-permutation sets of digits, perfect reconstruction is achieved with invariant latent dimension equal to class count, while equivariant-only models require dependence on NN (Winter et al., 2022).
  • 3D molecules/materials: Interatomic potentials computed with HIENet architectures produce MAE(E)=6.77 meV/atom\mathrm{MAE}(E) = 6.77\ \mathrm{meV/atom}, MAE(F)=24.82 meV/A˚\mathrm{MAE}(F) = 24.82\ \mathrm{meV/\AA}, outperforming pure-invariant and pure-equivariant models and achieving >2×>2\times speedup relative to equivariant-only models (Yan et al., 25 Feb 2025).
  • NLP (SCAN): A hybrid network with a group-invariant alignment module and a group-equivariant translator achieves 100%100\% accuracy on "Simple", "Add Jump", and "Around Right" splits—surpassing purely equivariant sequence-to-sequence models (White et al., 2022).
  • 3D self-supervised learning: A Split Invariant-Equivariant network achieves rotation R2=0.73R^2 = 0.73 and classification top-1 accuracy of 82.94%82.94\% on 3DIEBench (Garrido et al., 2023).

Parameter efficiency is generally superior to deep equivariant architectures by exploiting the lower cost of scalar-only operations in invariant blocks and limiting higher-order tensor operations to essential equivariant modules.

6. Applications, Design Considerations, and Future Directions

HIENets have been applied or proposed for:

  • Molecular property prediction, where permutation- and rotation-equivariance are essential for chemical validity.
  • Fluid simulation, combining scalar (invariant) and vector/tensor (equivariant) fields in graph neural networks for physically accurate long-term dynamics (Shankar et al., 2023).
  • Large-scale materials property prediction and discovery, satisfying physical constraints with analytic gradients due to symmetry-aware architectural design (Yan et al., 25 Feb 2025).
  • Graph learning and PDE-based solvers, interleaving local equivariant operators with global invariant pooling (Sannai et al., 2024).
  • Self-supervised representation learning with VICReg- or BYOL-style losses, enforcing invariance/equivariance splits at the latent level (Garrido et al., 2023).

Open questions and avenues for future research include:

  • Automated design of interleaving strategies (number and arrangement of invariant and equivariant blocks).
  • Extension to higher group complex symmetries (space groups, point groups, non-Euclidean symmetry actions).
  • Exploitation of data-driven group/lexicon discovery (NLP).
  • Theoretical trade-offs between invariance "hardness", model expressivity, and sample complexity.
  • Efficient implementation of high-order invariants, especially for large datasets or feature maps.

HIENet's modularity and mathematical guarantees make it a foundational approach for robust, interpretable, and efficient modeling in domains where explicit symmetry disentanglement is vital (Winter et al., 2022, Yan et al., 25 Feb 2025, Garrido et al., 2023, Sannai et al., 2024, White et al., 2022).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Hybrid Invariant–Equivariant Networks (HIENet).