JacNet: Jacobian Neural Function Approximation
- JacNet is a neural architecture that approximates functions by directly learning their Jacobian fields, enabling robust enforcement of constraints like invertibility and Lipschitz continuity.
- It employs structured Jacobian parameterizations, such as positive-definite and spectral-norm bounded formulations, to guarantee controlled derivative properties and theoretical guarantees.
- The network reconstructs functions through numerical integration of the Jacobian, facilitating accurate forward and inverse computations in supervised learning settings.
JacNet is a neural architecture designed for function approximation by directly parameterizing the Jacobian field of the target mapping, rather than the function itself. The key insight is that by learning a structured Jacobian, constraints such as invertibility and Lipschitz continuity can be imposed and controlled through neural network activations and output parameterizations, enabling the encoding of qualitative functional priors into neural models. Integration of the network-predicted Jacobian reconstructs the original function, and invertibility is ensured—or -Lipschitz priors enforced—by appropriate structural modifications to the network output. This approach provides fine-grained control over the differentiable properties of the learned mapping, facilitating both forward and inverse function computations with theoretical guarantees (Lorraine et al., 2024).
1. Core Principle: Jacobian Field Learning
JacNet targets the supervised learning problem where one seeks to model an unknown diffeomorphism . Rather than directly learning as a neural network, JacNet posit that the Jacobian is instead learned via a separate network. Using the fundamental theorem of line integrals, the function can be reconstructed from its Jacobian field as: where is a fixed base-point with known value.
The learning objective is empirical risk minimization over the data: where is a loss (e.g., or ), optimized via backpropagation through a numerical integration procedure.
This construction allows typical neural network machinery—activation functions, normalization constraints, positive-definite factorizations—to directly operate on the Jacobian, thereby imparting desired structural priors onto the function's derivatives.
2. Structuring the Jacobian Output
The network parameterizing can be constructed to enforce particular structural properties essential for qualitative function control:
- Unconstrained Jacobian: .
- Positive-Definite (PD) Jacobian for Invertibility: Let be an unconstrained output, then
for small . This guarantees strict positive-definiteness, ensuring all eigenvalues and , precluding negative eigen-directions.
- Block-Triangular or Triangular Factorization: (Proposed for future work.) The network outputs a lower-triangular matrix (with positive diagonals via softplus activations) and sets . This again ensures PD-ness and invertibility.
- Spectral-Norm-Bounded (Lipschitz) Jacobian: To enforce a -Lipschitz bound,
where is unconstrained; the elementwise tanh scales entries to and provides (tightened to with spectral normalization).
These constructions allow invertibility, -Lipschitz, and more general derivative-based priors to be imposed via architectural choice.
3. Theoretical Guarantees: Invertibility and Properness
JacNet's approach is supported by the inverse function and Hadamard's global inverse theorems. Specifically, if is and for all , then is locally invertible; if is also proper (i.e., pre-images of compact sets are compact) and the domains are connected, is a global -diffeomorphism.
In JacNet, parameterizing as strictly PD via the method above enforces . Properness can be ensured by domain clamping or by constructing to be coercive (i.e., as ), which is often achieved in practice.
4. Enforcing Lipschitz and Additional Priors
A function is -Lipschitz if for all . By the mean-value inequality, a sufficient condition is: JacNet realizes this by bounding the spectral norm of the output Jacobian through:
- Elementwise tanh scaling, plus spectral normalization on intermediate weights.
- Direct orthonormal factorization with singular values clamped to .
- Householder or Cayley-based parameterizations.
Because the Jacobian is directly output by the network, these constraints are applied as output layer activations, leaving the reconstruction procedure by line integration unchanged. Other priors, such as convexity or Cauchy–Riemann conditions, can be encoded similarly.
5. Training and Optimization Procedure
Training JacNet involves choosing a base-point (and if available), defining the reconstructed function via: The training objective is: where may be a penalty or additional regularization, but experiments are conducted with . Optimization leverages backpropagation through the numerical integrator for end-to-end learning.
6. Inverting the Learned Mapping
Given that is a global diffeomorphism under the enforced constraints, inversion is performed by the analogous integral in output space: where . For the matrix-factorized Jacobian form, inversion is efficiently computed via Cholesky or Woodbury identities (complexity , reduced to for low-rank ).
An Euler scheme is proposed for numerical inversion:
1 2 3 4 5 6 7 |
Δy = (y - y0) / T x = x0 for t in range(T): Y = x0 + t * Δy J = Jθ(Y) x = x + inv(J) @ Δy return x |
7. Empirical Illustration and Comparative Performance
JacNet is empirically validated on 1D problems, including invertible and Lipschitz function approximation:
- Invertible task (, ): Using the positive-definite parameterization, JacNet recovers precisely after 50 Adam iterations on training points. The unconstrained baseline fails, collapsing its derivative to zero and making inversion intractable.
- 1-Lipschitz task (, ): With so , JacNet matches exactly with bounded slope; the unconstrained baseline fails to maintain Lipschitz bounds and generalizes poorly.
Results are summarized as follows:
| Task | Unconstrained | JacNet-PD inv. | JacNet-Lipschitz |
|---|---|---|---|
| inv. | N/A | ||
| Lipschitz | N/A |
Explicit control of derivative properties, as provided by JacNet, is demonstrated to be essential for correct qualitative behavior (monotonicity, slope-bounds), not merely for minimizing training error.
8. Functional Priors and Modeling Implications
JacNet conceptually unifies a wide range of structural constraints as output constraints on the learned Jacobian field. By this construction, entire families of functional priors (invertibility, Lipschitz continuity, convexity, Cauchy–Riemann, among others) can be incorporated directly into end-to-end differentiable models. This allows modeling of highly flexible, universal-approximation-style derivatives, with precise control over the qualitative behavior of the resulting function mappings (Lorraine et al., 2024).