Papers
Topics
Authors
Recent
Search
2000 character limit reached

Globally Injective ReLU Networks

Published 15 Jun 2020 in cs.LG and stat.ML | (2006.08464v4)

Abstract: Injectivity plays an important role in generative models where it enables inference; in inverse problems and compressed sensing with generative priors it is a precursor to well posedness. We establish sharp characterizations of injectivity of fully-connected and convolutional ReLU layers and networks. First, through a layerwise analysis, we show that an expansivity factor of two is necessary and sufficient for injectivity by constructing appropriate weight matrices. We show that global injectivity with iid Gaussian matrices, a commonly used tractable model, requires larger expansivity between 3.4 and 10.5. We also characterize the stability of inverting an injective network via worst-case Lipschitz constants of the inverse. We then use arguments from differential topology to study injectivity of deep networks and prove that any Lipschitz map can be approximated by an injective ReLU network. Finally, using an argument based on random projections, we show that an end-to-end -- rather than layerwise -- doubling of the dimension suffices for injectivity. Our results establish a theoretical basis for the study of nonlinear inverse and inference problems using neural networks.

Citations (22)

Summary

  • The paper demonstrates that a ReLU layer is injective when its weight matrix meets the Directed Spanning Set condition across all input directions.
  • It establishes that an expansivity factor of 2 is both necessary and sufficient for injectivity, with empirical validation showing higher factors for Gaussian weights.
  • The study connects theoretical injectivity conditions to improved generative model inference and universal function approximation in multilayer architectures.

Globally Injective ReLU Networks

The paper "Globally Injective ReLU Networks" addresses the problem of determining injectivity for ReLU neural networks, especially focusing on fully-connected and convolutional networks. Injectivity is a crucial property in applications such as generative models and inverse problems. This work provides theoretical insights on conditions for neural network layers to be injective, establishes lower bounds on expansivity, and connects these findings to practical enhancements in generative model inference.

Layerwise Injectivity

Directed Spanning Set

The concept of a Directed Spanning Set (DSS) is introduced to understand when a ReLU-layered network is injective. This involves determining whether the weight matrix has enough vectors with positive inner products across all directions in the input space. Specifically, a ReLU layer is injective if for each input direction, it spans the output space. Figure 1

Figure 1

Figure 1: Illustration showing non-injective configuration due to absence of DSS.

Minimal Expansivity for Injective Layers

A critical result shows that an expansivity factor of 2 is both necessary and sufficient for the injectivity of ReLU layers. This presents a practical guideline for designing injective architectures, particularly for layers with Gaussian weights. Figure 2

Figure 2: Visualization of weight matrix mappings in an injective vs. non-injective scenario.

Multilayer Networks and Universal Function Approximation

Injectivity Across Layers

While each layer can be injective, the composition (entire network) may not be. The paper proves a universal approximation theorem for injective networks, showing that any continuous function can be approximated by an injective ReLU network with doubled output space dimensions. This results in a deeper understanding of how to preserve injectivity across layers.

Practical Implementation: Generative Inference

In practice, ensuring injectivity across a generative model facilitates better inversions within generative adversarial networks (GANs). The study shows that imposing injective constraints can lead to improved inference without sacrificing expressivity.

Empirical Validation

The work empirically validates the necessity of sufficient expansivity (greater than 3.4) for Gaussian weights to achieve injectivity. This demonstrates a sharp characterization of layerwise and end-to-end injectivity, challenging previous work that suggested lower factors for injectivity assurance. The numerical experiments confirm the theoretical predictions, while also highlighting the benefits in the context of generative model inference. Figure 3

Figure 3: Empirical probability of injectivity as a function of expansivity for various dimensions.

Conclusion

Overall, the paper provides a comprehensive theoretical treatment of injectivity in ReLU networks, proposing practical conditions for ensuring injectivity and illustrating the role of expansivity and layer design. This lays the groundwork for more robust applications in generative modeling and inverse problems where injectivity is critical for mapping back from the range to the domain. Figure 4

Figure 4: Mapping illustration demonstrating ReLU layer injectivity on latent spaces.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.