Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 43 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 18 tok/s Pro
GPT-5 High 16 tok/s Pro
GPT-4o 95 tok/s Pro
Kimi K2 198 tok/s Pro
GPT OSS 120B 464 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

How Much Training Data is Memorized in Overparameterized Autoencoders? An Inverse Problem Perspective on Memorization Evaluation (2310.02897v2)

Published 4 Oct 2023 in cs.LG

Abstract: Overparameterized autoencoder models often memorize their training data. For image data, memorization is often examined by using the trained autoencoder to recover missing regions in its training images (that were used only in their complete forms in the training). In this paper, we propose an inverse problem perspective for the study of memorization. Given a degraded training image, we define the recovery of the original training image as an inverse problem and formulate it as an optimization task. In our inverse problem, we use the trained autoencoder to implicitly define a regularizer for the particular training dataset that we aim to retrieve from. We develop the intricate optimization task into a practical method that iteratively applies the trained autoencoder and relatively simple computations that estimate and address the unknown degradation operator. We evaluate our method for blind inpainting where the goal is to recover training images from degradation of many missing pixels in an unknown pattern. We examine various deep autoencoder architectures, such as fully connected and U-Net (with various nonlinearities and at diverse train loss values), and show that our method significantly outperforms previous memorization-evaluation methods that recover training data from autoencoders. Importantly, our method greatly improves the recovery performance also in settings that were previously considered highly challenging, and even impractical, for such recovery and memorization evaluation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (25)
  1. Fast image recovery using variable splitting and constrained optimization. IEEE Trans. Image Process., 19(9):2345–2356, 2010.
  2. Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine Learning, 3(1):1–122, 2011.
  3. Turning a denoiser into a super-resolver using plug and play priors. In 2016 IEEE International Conference on Image Processing (ICIP), 2016.
  4. Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21), pp. 2633–2650, 2021.
  5. Membership inference attacks from first principles. In 2022 IEEE Symposium on Security and Privacy (SP), pp. 1897–1914, 2022.
  6. Extracting training data from diffusion models. In 32nd USENIX Security Symposium (USENIX Security 23), pp. 5253–5270, 2023.
  7. Plug-and-play ADMM for image restoration: Fixed-point convergence and applications. IEEE Trans. Comput. Imag., 3(1):84–98, 2017.
  8. Postprocessing of compressed images via sequential denoising. IEEE Transactions on Image Processing, 25(7):3044–3058, 2016.
  9. Subspace fitting meets regression: The effects of supervision and orthonormality constraints on double descent of generalization errors. In International Conference on Machine Learning (ICML), pp. 2366–2375, 2020.
  10. Convolutional proximal neural networks and plug-and-play algorithms. Linear Algebra and its Applications, 631:203–234, 2021.
  11. Membership inference attacks on machine learning: A survey. ACM Comput. Surv., 54(11s), sep 2022.
  12. Y. Jiang and C. Pehlevan. Associative memory in iterated overparameterized sigmoid autoencoders. In Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pp.  4828–4838. PMLR, 13–18 Jul 2020.
  13. A plug-and-play priors approach for solving nonlinear imaging inverse problems. IEEE Signal Processing Letters, 24(12):1872–1876, 2017.
  14. J.-J. Moreau. Proximité et dualité dans un espace hilbertien. Bulletin de la Société mathématique de France, 93:273–299, 1965.
  15. A. Nouri and S. A. Seyyedsalehi. Eigen value based loss function for training attractors in iterated autoencoders. Neural Networks, 2023.
  16. Downsampling leads to image memorization in convolutional autoencoders. 2018a.
  17. Memorization in overparameterized autoencoders. arXiv preprint arXiv:1810.10333, 2018b.
  18. Overparameterized neural networks implement associative memory. Proceedings of the National Academy of Sciences, 117(44):27162–27170, 2020.
  19. Poisson inverse problems by the plug-and-play scheme. Journal of Visual Communication and Image Representation, 41:96–108, 2016.
  20. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pp.  234–241. Springer, 2015.
  21. Nonlinear total variation based noise removal algorithms. Physica D: nonlinear phenomena, 60(1-4):259–268, 1992.
  22. Membership inference attacks against machine learning models. In 2017 IEEE symposium on security and privacy (SP), pp. 3–18. IEEE, 2017.
  23. Plug-and-play priors for bright field electron tomography and sparse interpolation. IEEE Transactions on Computational Imaging, 2(4):408–423, 2016.
  24. Plug-and-play priors for model based reconstruction. In IEEE GlobalSIP, 2013.
  25. Zero-shot image restoration using denoising diffusion null-space model. International Conference on Learning Representations (ICLR), 2023.
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces an inverse problem strategy to evaluate memorization by recovering degraded training images using overparameterized autoencoders.
  • The paper leverages ADMM and plug-and-play priors to optimize both reconstruction and degradation estimation, outperforming existing methods.
  • The paper demonstrates robust recovery and noise resilience across multiple architectures, including fully connected models and U-Net designs.

How Much Training Data is Memorized in Overparameterized Autoencoders? An Inverse Problem Perspective on Memorization Evaluation

Introduction

The paper investigates the phenomenon of memorization in overparameterized autoencoders, focusing on their ability to reconstruct degraded training images. The authors propose an inverse problem perspective, where the recovery of a degraded training image is formulated as an optimization task, using the trained autoencoder to implicitly define a regularizer for the dataset. This approach is tested on various architectures, demonstrating superior performance compared to previous methods.

Inverse Problem Formulation

The core contribution of the paper is the reframing of memorization as an inverse problem. Given a degraded training image, the objective is to recover the original image by minimizing a cost function that balances the reconstruction error and a regularization term. This regularization is implicitly defined by the trained autoencoder, reflecting the degree of memorization inherent in the model. Figure 1

Figure 1: Iterative recovery of a degraded training image using our proposed approach (top frame) and the method from previous works (bottom frame).

Implementation Details

Algorithmic Approach

The proposed method leverages the Alternating Direction Method of Multipliers (ADMM) combined with plug-and-play priors. The ADMM decomposition enables separate optimization of the image reconstruction and the estimation of the degradation operator. Notably, the framework extends the plug-and-play paradigm, traditionally used with denoisers, to incorporate arbitrary autoencoders.

Architectural Considerations

Experiments are conducted on fully connected and U-Net autoencoders with various nonlinearities. The architectures include:

  1. Fully Connected Autoencoders: 10 and 20-layer configurations.
  2. U-Net: Applied to CIFAR-10 and SVHN datasets, demonstrating adaptability to various image scales and complexities. Figure 2

    Figure 2: Architecture of 10 layers and 20 layers fully connected autoencoders for the Tiny ImageNet dataset (a subset of images, at 64×64×364 \times 64 \times 3 pixel size).

Experimental Results

Recovery Performance

The method shows a significant advantage in recovery performance, particularly in scenarios with unknown degradation masks. It achieves high recovery rates even under challenging conditions previously deemed impractical. Figure 3 illustrates that the proposed approach significantly outperforms both autoencoder iterations and generic inpainting techniques, especially in accurate recovery scenarios. Figure 3

Figure 3: Accurate recovery rates for recovery from degradation due to various missing pixel masks, tested on different architectures.

Noise Robustness

The evaluation also considers additive noise, with results indicating robust recovery capabilities under moderate noise levels. This robustness emphasizes the method's potential in real-world applications, where imaging conditions can often introduce noise. Figure 4

Figure 4: Recovery results of degraded samples with additive noise.

Conclusion

The paper provides a novel perspective on evaluating memorization in autoencoders by framing it as an inverse problem. The proposed methodological framework significantly enhances recovery rates of training data, thereby offering a more detailed empirical evaluation of memorization phenomena. This approach highlights the potential for further research into the intersection of inverse problems and deep learning, particularly concerning overparameterization and data memorization. Future work may explore extensions to other forms of neural networks and applications beyond image recovery.