Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Instance Enhancement Batch Normalization: an Adaptive Regulator of Batch Noise (1908.04008v2)

Published 12 Aug 2019 in cs.LG, cs.CV, and stat.ML

Abstract: Batch Normalization (BN)(Ioffe and Szegedy 2015) normalizes the features of an input image via statistics of a batch of images and hence BN will bring the noise to the gradient of the training loss. Previous works indicate that the noise is important for the optimization and generalization of deep neural networks, but too much noise will harm the performance of networks. In our paper, we offer a new point of view that self-attention mechanism can help to regulate the noise by enhancing instance-specific information to obtain a better regularization effect. Therefore, we propose an attention-based BN called Instance Enhancement Batch Normalization (IEBN) that recalibrates the information of each channel by a simple linear transformation. IEBN has a good capacity of regulating noise and stabilizing network training to improve generalization even in the presence of two kinds of noise attacks during training. Finally, IEBN outperforms BN with only a light parameter increment in image classification tasks for different network structures and benchmark datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Senwei Liang (27 papers)
  2. Zhongzhan Huang (25 papers)
  3. Mingfu Liang (17 papers)
  4. Haizhao Yang (99 papers)
Citations (53)

Summary

We haven't generated a summary for this paper yet.