Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 45 tok/s
Gemini 2.5 Pro 52 tok/s Pro
GPT-5 Medium 30 tok/s Pro
GPT-5 High 24 tok/s Pro
GPT-4o 96 tok/s Pro
Kimi K2 206 tok/s Pro
GPT OSS 120B 457 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

Protecting Privacy in Multimodal Large Language Models with MLLMU-Bench (2410.22108v2)

Published 29 Oct 2024 in cs.CL and cs.AI

Abstract: Generative models such as LLMs (LLM) and Multimodal LLMs (MLLMs) trained on massive web corpora can memorize and disclose individuals' confidential and private data, raising legal and ethical concerns. While many previous works have addressed this issue in LLM via machine unlearning, it remains largely unexplored for MLLMs. To tackle this challenge, we introduce Multimodal LLM Unlearning Benchmark (MLLMU-Bench), a novel benchmark aimed at advancing the understanding of multimodal machine unlearning. MLLMU-Bench consists of 500 fictitious profiles and 153 profiles for public celebrities, each profile feature over 14 customized question-answer pairs, evaluated from both multimodal (image+text) and unimodal (text) perspectives. The benchmark is divided into four sets to assess unlearning algorithms in terms of efficacy, generalizability, and model utility. Finally, we provide baseline results using existing generative model unlearning algorithms. Surprisingly, our experiments show that unimodal unlearning algorithms excel in generation and cloze tasks, while multimodal unlearning approaches perform better in classification tasks with multimodal inputs.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper’s main contribution is the introduction of MLLMU-Bench, a comprehensive benchmark for evaluating unlearning efficacy, generalizability, and model utility in multimodal LLMs.
  • It employs diverse datasets (Forget, Test, Retain, and Real Celebrity Sets) and compares methods like Gradient Ascent, KL Minimization, and NPO through classification, generation, and cloze tasks.
  • The study highlights the trade-off between achieving effective unlearning and preserving model utility, suggesting a need for further research in holistic privacy protection.

Protecting Privacy in Multimodal LLMs with MLLMU-Bench

Introduction

The research paper explores the intricate challenge of protecting privacy within multimodal LLMs (MLLMs), highlighting the risks of preserving sensitive information. Current efforts in unlearning for LLMs have primarily focused on textual data, leaving a gap in multimodal contexts. To address this, the paper proposes MLLMU-Bench, a comprehensive benchmark designed to evaluate the efficacy of unlearning methods across both textual and visual data, establishing a new standard for privacy protection in AI models. Figure 1

Figure 1: Demonstration of the multimodal unlearning task. MLLM is firstly fine-tuned on constructed profiles in the proposed benchmark. After fine-tuning, MLLM can answer multimodal questions related to profiles. We then conduct various unlearning methods on a portion of profiles (forget set). Finally, the performance on tasks related to the forget set and the remaining evaluation datasets are tested simultaneously.

MLLMU-Bench Framework

MLLMU-Bench introduces a diverse set of profiles—comprising fictitious individuals and public celebrities—to evaluate unlearning methods. Each profile features a series of customized question-answer pairs that are assessed through multimodal (image+text) and unimodal (text) tasks. Figure 2

Figure 2: Examples of question-answer pairs from all four distinct datasets used to assess model unlearning efficacy and model utility. The Forget, Test, Retain Set are fictitious individuals, while the Real Celebrity Set includes real public figures.

The benchmark is structured into four sets: Forget, Test, Retain, and Real Celebrity. It measures unlearning efficacy, generalizability, and model utility. The Forget Set focuses on the model's ability to erase specific knowledge, whereas the Test Set evaluates the extension of this unlearning to paraphrased or transformed data. The Retain Set ensures the preservation of unrelated knowledge, and the Real Celebrity Set assesses the impact on neighboring concepts.

Evaluation Metrics and Baseline Methods

Unlearning efficacy is tested through classification and generation tasks, where models must minimize the probability of retrieving correct answers from the Forget Set, ideally treating these as unknowns. Equally, generalizability is assessed by the model's reactions to the Test Set's transformed representations. Model utility evaluates accuracy in retaining important knowledge across different datasets. Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3: Classification, generation, and cloze performance of the GA algorithm applied to multimodal and unimodal setups with 5% forget data, using LLaVA as the base model.

Baseline methods considered include Gradient Ascent (GA), KL Minimization, Negative Preference Optimization (NPO), and system prompt-based strategies. The effectiveness of these methods is meticulously compared, illustrating that while GA shows strong unlearning results, it often compromises utility, making approaches like NPO more balanced in preserving model generalization while achieving unlearning goals.

Trade-offs and Challenges

A critical aspect of successful unlearning involves balancing the effectiveness with the preservation of model utility. The paper identifies that while GA can significantly enhance unlearning, the method also risks collapsing utility across datasets, including the Real Celebrity and Retain Sets. Conversely, systems like NPO demonstrate restrained effectiveness in unlearning but better preserve model utility, showcasing the trade-off between utility and data forgetting. Figure 4

Figure 4

Figure 4

Figure 4

Figure 4: The overall trade-off between unlearning effectiveness and model utility across all baselines using different forget data, with LLaVA as the base model.

Implications and Future Directions

The development of MLLMU-Bench marks a pivotal point for the evaluation of privacy-preserving algorithms in MLLMs. It highlights the necessity for further research in multimodal contexts—recognizing the added complexity due to interconnected modalities. Future directions may include tackling broader privacy scenarios—possibly integrating certified unlearning and more advanced attacks to test the robustness of MLLMs against data reconstruction or relearning.

Conclusion

"Protecting Privacy in Multimodal LLMs with MLLMU-Bench" provides a benchmark that offers comprehensive metrics for evaluating unlearning performance while reinforcing the importance of maintaining model utility. It brings critical insights into the challenges of privacy in AI, setting a precursor for future work to refine and enhance privacy-preserving techniques in multimodal frameworks.

X Twitter Logo Streamline Icon: https://streamlinehq.com