Dreaming is All You Need (2409.01633v3)
Abstract: In classification tasks, achieving a harmonious balance between exploration and precision is of paramount importance. To this end, this research introduces two novel deep learning models, SleepNet and DreamNet, to strike this balance. SleepNet seamlessly integrates supervised learning with unsupervised sleep" stages using pre-trained encoder models. Dedicated neurons within SleepNet are embedded in these unsupervised features, forming intermittentsleep" blocks that facilitate exploratory learning. Building upon the foundation of SleepNet, DreamNet employs full encoder-decoder frameworks to reconstruct the hidden states, mimicking the human "dreaming" process. This reconstruction process enables further exploration and refinement of the learned representations. Moreover, the principle ideas of our SleepNet and DreamNet are generic and can be applied to both computer vision and natural language processing downstream tasks. Through extensive empirical evaluations on diverse image and text datasets, SleepNet and DreanNet have demonstrated superior performance compared to state-of-the-art models, showcasing the strengths of unsupervised exploration and supervised precision afforded by our innovative approaches.
- The sleep-immune crosstalk in health and disease. Physiological reviews, 2019.
- Mark S Blumberg. Beyond dreams: do sleep-related movements contribute to brain development? Frontiers in Neurology, 1:140, 2010.
- Louis Breger. Function of dreams. Journal of Abnormal Psychology, 72(5p2):1, 1967.
- Language models are few-shot learners. ArXiv, abs/2005.14165, 2020.
- An empirical survey of data augmentation for limited data learning in nlp. Transactions of the Association for Computational Linguistics, 11:191–211, 2023.
- Randaugment: Practical automated data augmentation with a reduced search space. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pages 702–703, 2020.
- Coatnet: Marrying convolution and attention for all data sizes. Advances in neural information processing systems, 34:3965–3977, 2021.
- Bert: Pre-training of deep bidirectional transformers for language understanding. ArXiv, abs/1810.04805, 2019.
- G William Domhoff. A new neurocognitive theory of dreams. Dreaming, 11:13–33, 2001.
- An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- An image is worth 16x16 words: Transformers for image recognition at scale. ArXiv, abs/2010.11929, 2020.
- The neuroprotective aspects of sleep. MEDtube science, 3(1):35, 2015.
- The role of sleep in emotional brain function. Annual review of clinical psychology, 10:679–708, 2014.
- Generative adversarial nets. In NIPS, 2014.
- Masked autoencoders are scalable vision learners. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 15979–15988, 2021.
- Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770–778, 2015.
- Masking augmentation for supervised learning. arXiv preprint arXiv:2306.11339, 2023.
- Mobilenets: Efficient convolutional neural networks for mobile vision applications. ArXiv, abs/1704.04861, 2017.
- Trans-blstm: Transformer with bidirectional lstm for language understanding. arXiv preprint arXiv:2003.07000, 2020.
- Adversarial examples are not bugs, they are features. ArXiv, abs/1905.02175, 2019.
- Highly accurate protein structure prediction with alphafold. Nature, 596:583 – 589, 2021.
- Yoon Kim. Convolutional neural networks for sentence classification. In Conference on Empirical Methods in Natural Language Processing, 2014.
- Adam: A method for stochastic optimization. CoRR, abs/1412.6980, 2014.
- Auto-encoding variational bayes. CoRR, abs/1312.6114, 2013.
- Learning multiple layers of features from tiny images. 2009.
- Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60:84 – 90, 2012.
- Ya Le and Xuan S. Yang. Tiny imagenet visual recognition challenge. 2015.
- Roberta: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692, 2019.
- Learning word vectors for sentiment analysis. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pages 142–150, Portland, Oregon, USA, June 2011. Association for Computational Linguistics.
- Efficient estimation of word representations in vector space. In International Conference on Learning Representations, 2013.
- Textattack: A framework for adversarial attacks, data augmentation, and adversarial training in nlp. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 119–126, 2020.
- National Geographic. Giant panda eating, 2024. Accessed: 2024-06-04.
- Deep contextualized word representations. ArXiv, abs/1802.05365, 2018.
- About sleep’s role in memory. Physiological reviews, 93 2:681–766, 2013.
- ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 115(3):211–252, 2015.
- Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. ArXiv, abs/1910.01108, 2019.
- Very deep convolutional networks for large-scale image recognition. CoRR, abs/1409.1556, 2014.
- Going deeper with convolutions. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1–9, 2014.
- Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning, pages 6105–6114. PMLR, 2019.
- Augmenting convolutional networks with attention-based aggregation. arXiv preprint arXiv:2112.13692, 2021.
- Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Cvt: Introducing convolutions to vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pages 22–31, 2021.
- Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems, 32, 2019.
- Understanding how pretraining regularizes deep learning algorithms. IEEE Transactions on Neural Networks and Learning Systems, 2021.
- Scaling vision transformers. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 12104–12113, 2022.
- mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412, 2017.
- Character-level convolutional networks for text classification. In NIPS, 2015.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.