Super-Affine Transformations
- Super-affine transformations are a generalization of affine transformations that use enriched polynomial bases to capture complex, non-affine symmetries in data.
- The framework employs null-space learning with manifold constraints to efficiently recover symmetry generators and compute invariant features.
- Empirical evaluations show that super-affine methods outperform traditional affine approaches and models like LieGAN, particularly in non-affine tasks and scalability.
Super-affine transformations generalize classical affine transformations in Euclidean spaces by permitting higher-order polynomial or more complex basis expansions in the infinitesimal generator coefficients of transformation groups. This extension allows the discovery and exploitation of continuous symmetry structures in data that fall strictly outside the affine regime, interpolating methodically between affine symmetry and the full diffeomorphism group of ℝⁿ. The super-affine framework encompasses a versatile spectrum of polynomial and other function-based transformations, enabling symmetry-based analyses, model invariance, and geometric reasoning in machine learning beyond traditional affine groups (Shaw et al., 5 Jun 2024).
1. Mathematical Structure of Super-Affine Transformation Groups
The real affine group in n dimensions, Aff(n), has the standard action with and , associated infinitesimally with vector fields for and . The super-affine construction replaces the affine constraint of first-degree polynomial coefficients by an ansatz involving richer feature bases (e.g., monomials up to degree d or even non-polynomial bases), leading to vector fields
where spans the chosen feature space. The resulting transformation group, denoted here as , is a (potentially infinite-dimensional) Lie group strictly containing as a subgroup and lying within the diffeomorphism group of . By systematically enriching the basis , the framework interpolates from affine actions to higher-order polynomial or other functional symmetries.
2. Optimization and Null-Space Learning Objectives
Super-affine symmetry discovery relies on the principle that a vector field generates a symmetry of a function if and only if everywhere (or at all sampled points). By parameterizing with unknown weights spanning the chosen basis, and representing features as , this symmetry condition is transcribed for data samples into a linear system , where accumulates over all samples and coordinate directions.
The concrete optimization problem is
which seeks an orthonormal basis for the null-space of , with orthonormality imposed on the Stiefel manifold. Manifold-aware optimizers, as implemented in McTorch, enforce these constraints efficiently. Symmetry-invariant features are subsequently constructed by forming a similar null-space problem over the Jacobians of the basis functions under the identified generators, again with Stiefel or spherical manifold optimization.
3. Algorithmic Framework
The super-affine discovery algorithm proceeds via the following steps:
- Basis Selection: Choose scalar feature functions (e.g., all monomials up to degree ).
- Function Estimation: Fit (e.g., regression, density, or level-set models).
- Gradient Evaluation: Compute for all data points .
- Extended Matrix Construction: For each sample and each coordinate, set , yielding an extended feature matrix .
- Infinitesimal Generator Recovery: Solve for as in the optimization above to obtain symmetry generators.
- Invariant Computation (Optional): Construct a secondary matrix from the Jacobians under each ; solve a constrained null-space problem to obtain scalar invariants .
Both generator and invariant learning steps utilize manifold optimization to preserve the orthonormality (for ) or norm (for ) constraints.
4. Specialization to the Affine Case
Super-affine methods recover the classical affine framework as a special case. If the basis contains only constants and coordinate functions, i.e., , , then the ansatz for reduces to
which forms precisely the Lie algebra . In this reduction, the extended matrix encodes only first-degree features, and the solution space has dimension , corresponding exactly to existing "Lie-algebra learning" methods such as LieGAN but within a vector field null-space framework.
5. Empirical Evaluation and Benchmarks
Comparative experiments demonstrate super-affine symmetry detection on both affine and genuinely non-affine tasks, benchmarked primarily against LieGAN. The following table summarizes salient results:
| Task | LieGAN Similarity | Super-affine Similarity | LieGAN Time (s) | Super-affine Time (s) |
|---|---|---|---|---|
| Affine (N=200) | 0.50 ± 0.28 | 0.99 ± 0.03 | 1.94 ± 0.06 | 1.28 ± 0.08 |
| Affine (N=2,000) | 0.998 ± 0.002 | 0.97 ± 0.08 | 15.4 ± 1.0 | 1.69 ± 0.11 |
| Affine (N=20,000) | 0.9999 ± 5e-5 | 0.97 ± 0.07 | 177 ± 7 | 3.90 ± 0.09 |
| Super-affine (poly 2) | 0.19 ± 0.11 | 0.992 ± 0.011 | - | - |
For finite , the super-affine method more reliably locates the generator, achieving substantially higher similarity for small sample sizes and is significantly faster at large . On super-affine (e.g., quadratic) symmetry targets, LieGAN is incapable () while super-affine recovers near the ground truth ().
Additional applications include:
- Supervised Classification Invariants: On the Palmer Penguins dataset, the scalar invariant
computed via super-affine symmetry aligns with model probability contours for the "Adelie" class.
- Symmetry-based Regression: For functions with $7$-fold discrete rotational symmetry, symmetry-adapted features increase from $0.5965$ (test) in the original coordinates to $0.9965$ (test) after symmetry-based transformation.
- Infinitesimal Isometries in Pulled-back Metrics: Recovery of the 6-D Killing algebra of a nontrivial Riemannian metric on ℝ³ is achieved to cosine similarity .
6. Scope, Flexibility, and Limitations
The super-affine transformation framework extends symmetry discovery capabilities:
- The inclusion of polynomial or otherwise enriched feature bases allows continuous interpolation from classical affine symmetry to broader polynomial and isometric symmetries.
- Null-space learning with manifold constraints produces explicit generators and invariants.
- Reduction to affine forms recovers existing algorithms exactly as a special case.
- Capability to detect high-order ("beyond-affine") continuous symmetries empirically distinguishes it.
- Super-affine methods match or outperform LieGAN for both sample efficiency and computational cost in all tested domains, particularly excelling at capturing non-affine symmetries.
A plausible implication is that this framework defines a new algorithmic regime for learning and exploiting symmetry in data manifolds, with potential applicability across regression, density modeling, geometric learning, and invariant representation extraction (Shaw et al., 5 Jun 2024).
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free