Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MixACM: Mixup-Based Robustness Transfer via Distillation of Activated Channel Maps (2111.05073v1)

Published 9 Nov 2021 in cs.LG, cs.AI, and cs.CV

Abstract: Deep neural networks are susceptible to adversarially crafted, small and imperceptible changes in the natural inputs. The most effective defense mechanism against these examples is adversarial training which constructs adversarial examples during training by iterative maximization of loss. The model is then trained to minimize the loss on these constructed examples. This min-max optimization requires more data, larger capacity models, and additional computing resources. It also degrades the standard generalization performance of a model. Can we achieve robustness more efficiently? In this work, we explore this question from the perspective of knowledge transfer. First, we theoretically show the transferability of robustness from an adversarially trained teacher model to a student model with the help of mixup augmentation. Second, we propose a novel robustness transfer method called Mixup-Based Activated Channel Maps (MixACM) Transfer. MixACM transfers robustness from a robust teacher to a student by matching activated channel maps generated without expensive adversarial perturbations. Finally, extensive experiments on multiple datasets and different learning scenarios show our method can transfer robustness while also improving generalization on natural images.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Muhammad Awais (59 papers)
  2. Fengwei Zhou (21 papers)
  3. Chuanlong Xie (23 papers)
  4. Jiawei Li (116 papers)
  5. Sung-Ho Bae (29 papers)
  6. Zhenguo Li (195 papers)
Citations (16)

Summary

We haven't generated a summary for this paper yet.