Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adversarial Dropout for Recurrent Neural Networks (1904.09816v1)

Published 22 Apr 2019 in cs.LG and stat.ML

Abstract: Successful application processing sequential data, such as text and speech, requires an improved generalization performance of recurrent neural networks (RNNs). Dropout techniques for RNNs were introduced to respond to these demands, but we conjecture that the dropout on RNNs could have been improved by adopting the adversarial concept. This paper investigates ways to improve the dropout for RNNs by utilizing intentionally generated dropout masks. Specifically, the guided dropout used in this research is called as adversarial dropout, which adversarially disconnects neurons that are dominantly used to predict correct targets over time. Our analysis showed that our regularizer, which consists of a gap between the original and the reconfigured RNNs, was the upper bound of the gap between the training and the inference phases of the random dropout. We demonstrated that minimizing our regularizer improved the effectiveness of the dropout for RNNs on sequential MNIST tasks, semi-supervised text classification tasks, and LLMing tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sungrae Park (17 papers)
  2. Kyungwoo Song (38 papers)
  3. Mingi Ji (8 papers)
  4. Wonsung Lee (5 papers)
  5. Il-Chul Moon (39 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.