Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Removing Adversarial Noise in Class Activation Feature Space (2104.09197v1)

Published 19 Apr 2021 in cs.LG

Abstract: Deep neural networks (DNNs) are vulnerable to adversarial noise. Preprocessing based defenses could largely remove adversarial noise by processing inputs. However, they are typically affected by the error amplification effect, especially in the front of continuously evolving attacks. To solve this problem, in this paper, we propose to remove adversarial noise by implementing a self-supervised adversarial training mechanism in a class activation feature space. To be specific, we first maximize the disruptions to class activation features of natural examples to craft adversarial examples. Then, we train a denoising model to minimize the distances between the adversarial examples and the natural examples in the class activation feature space. Empirical evaluations demonstrate that our method could significantly enhance adversarial robustness in comparison to previous state-of-the-art approaches, especially against unseen adversarial attacks and adaptive attacks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Dawei Zhou (53 papers)
  2. Nannan Wang (106 papers)
  3. Chunlei Peng (20 papers)
  4. Xinbo Gao (194 papers)
  5. Xiaoyu Wang (200 papers)
  6. Jun Yu (234 papers)
  7. Tongliang Liu (251 papers)
Citations (24)

Summary

We haven't generated a summary for this paper yet.