Papers
Topics
Authors
Recent
2000 character limit reached

Model Inversion Attacks Meet Cryptographic Fuzzy Extractors (2510.25687v1)

Published 29 Oct 2025 in cs.CR and cs.LG

Abstract: Model inversion attacks pose an open challenge to privacy-sensitive applications that use ML models. For example, face authentication systems use modern ML models to compute embedding vectors from face images of the enrolled users and store them. If leaked, inversion attacks can accurately reconstruct user faces from the leaked vectors. There is no systematic characterization of properties needed in an ideal defense against model inversion, even for the canonical example application of a face authentication system susceptible to data breaches, despite a decade of best-effort solutions. In this paper, we formalize the desired properties of a provably strong defense against model inversion and connect it, for the first time, to the cryptographic concept of fuzzy extractors. We further show that existing fuzzy extractors are insecure for use in ML-based face authentication. We do so through a new model inversion attack called PIPE, which achieves a success rate of over 89% in most cases against prior schemes. We then propose L2FE-Hash, the first candidate fuzzy extractor which supports standard Euclidean distance comparators as needed in many ML-based applications, including face authentication. We formally characterize its computational security guarantees, even in the extreme threat model of full breach of stored secrets, and empirically show its usable accuracy in face authentication for practical face distributions. It offers attack-agnostic security without requiring any re-training of the ML model it protects. Empirically, it nullifies both prior state-of-the-art inversion attacks as well as our new PIPE attack.

Summary

  • The paper introduces the PIPE attack and demonstrates that defenses such as Facial-FE and MRP are highly vulnerable under full leakage, with attack success rates often exceeding 89%.
  • The study employs formal threat modeling and empirical evaluations using metrics like ASR, LPIPS, and FID to rigorously assess the strengths and weaknesses of existing post-processing schemes.
  • The proposed L2FE-Hash, a lattice-based fuzzy extractor, offers strong cryptographic security with practical authentication accuracy, setting a new standard for privacy-preserving ML systems.

Model Inversion Attacks and Cryptographic Fuzzy Extractors: A Formal and Empirical Analysis

Introduction and Motivation

This paper addresses the persistent vulnerability of ML systems to model inversion attacks, particularly in privacy-sensitive applications such as face authentication. Embedding vectors generated by deep models (e.g., Facenet, ArcFace) are commonly stored for authentication purposes. If these embeddings are leaked, adversaries can reconstruct the original input (e.g., face images) with high fidelity, posing severe privacy risks. The work formalizes the requirements for a robust, attack-agnostic defense and establishes a connection to the cryptographic primitive of fuzzy extractors, specifically in the context of Euclidean (â„“2\ell_2) distance comparators.

Threat Model and Problem Formalization

The paper considers a full-leakage threat model, analogous to password database breaches, where all persistently stored data—including protected embeddings and any auxiliary randomness—is compromised. The adversary is assumed to have black-box access to the target ML model and public datasets representative of the input distribution. The central challenge is to design a post-processing defense that:

  • Tolerates noise (i.e., allows authentication with slightly different inputs),
  • Remains secure even if all stored data is leaked,
  • Preserves sufficient entropy for utility (i.e., avoids trivial compression).

Existing Post-Processing Defenses and Their Limitations

Two prominent post-processing schemes are analyzed:

  1. Facial-Fuzzy Extractor (Facial-FE): A lattice-based fuzzy extractor designed for approximate â„“2\ell_2 correction. It uses deterministic decoding, which leaks significant information about the original embedding via the helper string. This makes it vulnerable to inversion attacks.
  2. Multispace Random Projection (MRP): Projects embeddings onto a lower-dimensional random subspace. While non-invertible in theory, it is susceptible to approximate recovery when both the projection matrix and projected vector are leaked.

Empirical evaluation demonstrates that both schemes are highly vulnerable under full-leakage. Attack success rates (ASR) exceed 89% in most cases, indicating that these defenses do not provide cryptographically strong protection.

The PIPE Attack: Adaptive Model Inversion

The paper introduces the PIPE (Post-leakage Inversion of Protected Embeddings) attack, which exploits the weaknesses of existing post-processing schemes. PIPE operates in two stages:

  1. Surrogate Embedding Recovery: For Facial-FE, the helper string is used directly as a surrogate for the original embedding. For MRP, the Moore-Penrose pseudo-inverse of the projection matrix is used to recover an approximate embedding.
  2. Image Reconstruction: A diffusion model, conditioned on the surrogate embedding, is trained to reconstruct images close to the original input.

Empirical results show that PIPE achieves ASR above 89% for Facial-FE and above 47% for MRP, far exceeding random guessing baselines. Figure 1

Figure 1: An example inverting the FacialFE-protected embedding vector to obtain a face image similar to the original using PIPE.

Formalizing the Ideal Defense: Fuzzy One-Way Hashes

The paper formalizes the requirements for an ideal post-processing primitive, termed a fuzzy one-way hash (FOWH), which must satisfy:

  • Correctness: Noise tolerance for authentication.
  • Security: Fuzzy one-wayness—hardness of recovering any input close to the original, even with full leakage.
  • Utility: Output entropy comparable to input entropy.

It is shown that a fuzzy extractor for â„“2\ell_2 distance satisfies these properties, provided the input distribution is well-behaved (e.g., high min-entropy, bounded support).

L2FE-Hash: A Secure â„“2\ell_2-Norm Fuzzy Extractor

The paper proposes L2FE-Hash, a lattice-based fuzzy extractor supporting â„“2\ell_2 distance comparators. The construction uses random q-ary lattices and Babai's Nearest Plane (BNP) algorithm for error correction. The secret is hashed before storage, ensuring security under full leakage.

Theoretical analysis demonstrates that L2FE-Hash is computationally secure under the assumption that the input distribution is uniform over a union of disjoint ϵ\epsilon-balls (representing user clusters). The number of secure output bits is a function of the input entropy and lattice parameters. Figure 2

Figure 2: A face authentication system with Enroll and Auth functions. MM represents models that output unprotected embeddings and MprotM_{prot} represents models with post-processing protection mechanisms that output protected embeddings.

Empirical Evaluation of L2FE-Hash

L2FE-Hash is evaluated against PIPE and prior inversion attacks (Bob, KED-MI, GMI) using state-of-the-art face recognition models and multiple datasets (Celeb-A, LFW, CASIA-Webface). Key findings:

  • Attack Resistance: ASR for PIPE and all prior attacks drops to below 7% for ArcFace and below 0.6% for Facenet, matching random guessing rates.
  • Authentication Accuracy: L2FE-Hash achieves true positive rates above 60% and false positive rates below 26%. Majority voting over multiple samples can further boost accuracy. Figure 3

    Figure 3: Reconstructed images by PIPE and Bob attacks against different protection schemes for Facenet embeddings. Check mark indicates successful authentication while cross indicates failure.

Comparative Analysis and Metrics

Visual similarity metrics (LPIPS, FID) confirm that PIPE reconstructions are perceptually close to originals for vulnerable schemes, but L2FE-Hash reconstructions are not, indicating effective privacy protection. Cross-model evaluation shows that PIPE reconstructions generalize across different feature extractors, underscoring the severity of the privacy risk.

Theoretical and Practical Implications

The work establishes that:

  • Partial defenses (Facial-FE, MRP) are insufficient under realistic threat models.
  • Fuzzy extractors, specifically L2FE-Hash, provide attack-agnostic, model-agnostic security for embedding-based authentication, without requiring model retraining.
  • Security guarantees are formalized and empirically validated for real-world face distributions.

The approach generalizes to other modalities (audio, text) and security applications (copyright, CSAM detection) where embedding vectors are used.

Future Directions

Open problems include:

  • Characterizing real-world input distributions to tighten theoretical security bounds.
  • Extending fuzzy extractor constructions to other distance metrics and modalities.
  • Optimizing authentication accuracy for practical deployment, possibly via ensemble methods or adaptive parameter tuning.

Conclusion

This paper provides a rigorous formalization and empirical validation of cryptographic fuzzy extractors as a defense against model inversion attacks in ML-based authentication systems. The proposed L2FE-Hash construction achieves strong security and practical utility, marking a significant advance in privacy-preserving ML system design. The results highlight the necessity of cryptographically principled post-processing defenses and set a foundation for future research in secure ML deployments.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Youtube Logo Streamline Icon: https://streamlinehq.com