DIRA: Dynamic Domain Incremental Regularised Adaptation (2205.00147v5)
Abstract: Autonomous systems (AS) often use Deep Neural Network (DNN) classifiers to allow them to operate in complex, high-dimensional, non-linear, and dynamically changing environments. Due to the complexity of these environments, DNN classifiers may output misclassifications during operation when they face domains not identified during development. Removing a system from operation for retraining becomes impractical as the number of such AS increases. To increase AS reliability and overcome this limitation, DNN classifiers need to have the ability to adapt during operation when faced with different operational domains using a few samples (e.g. 2 to 100 samples). However, retraining DNNs on a few samples is known to cause catastrophic forgetting and poor generalisation. In this paper, we introduce Dynamic Incremental Regularised Adaptation (DIRA), an approach for dynamic operational domain adaption of DNNs using regularisation techniques. We show that DIRA improves on the problem of forgetting and achieves strong gains in performance when retraining using a few samples from the target domain. Our approach shows improvements on different image classification benchmarks aimed at evaluating robustness to distribution shifts (e.g.CIFAR-10C/100C, ImageNet-C), and produces state-of-the-art performance in comparison with other methods from the literature.
- Nidhi Kalra and Susan M. Paddock “Driving to Safety: How Many Miles of Driving Would It Take to Demonstrate Autonomous Vehicle Reliability?” Santa Monica, CA: RAND Corporation, 2016 DOI: 10.7249/RR1478
- “An agency-directed approach to test generation for simulation-based autonomous vehicle verification” In 2020 IEEE International Conference On Artificial Intelligence Testing (AITest), 2020, pp. 31–38 IEEE
- Kerstin I Eder, Wen-ling Huang and Jan Peleska “Complete Agent-driven Model-based System Testing for Autonomous Systems” In arXiv preprint arXiv:2110.12586, 2021
- “Assessing Trustworthiness of Autonomous Systems”, 2023 arXiv:2305.03411 [cs.AI]
- “Positive Trust Balance for Self-driving Car Deployment” Springer International Publishing, 2020, pp. 351–357 DOI: 10.1007/978-3-030-55583-2˙26
- Darryl Hond, Hamid Asgari and Daniel Jeffery “Verifying Artificial Neural Network Classifier Performance Using Dataset Dissimilarity Measures” In Proceedings - 19th IEEE International Conference on Machine Learning and Applications, ICMLA 2020, 2020, pp. 115–121 DOI: 10.1109/ICMLA51294.2020.00027
- J.David Schaffer and Walker H. Land “Predicting with Confidence: Classifiers that Know What They Don’t Know” In Procedia Computer Science 114 Elsevier B.V., 2017, pp. 200–207 DOI: 10.1016/j.procs.2017.09.061
- “Distance-based Confidence Score for Neural Network Classifiers”, 2017 arXiv: http://arxiv.org/abs/1709.09844
- “Distance-Based Learning from Errors for Confidence Calibration”, 2019, pp. 1–12 arXiv: http://arxiv.org/abs/1912.01730
- “The oracle problem in software testing: A survey” In IEEE Transactions on Software Engineering 41.5 IEEE, 2015, pp. 507–525 DOI: 10.1109/TSE.2014.2372785
- “Machine Learning Testing: Survey, Landscapes and Horizons” In IEEE Transactions on Software Engineering, 2020, pp. 1–1 DOI: 10.1109/tse.2019.2962027
- “An empirical investigation of catastrophic forgetting in gradient-based neural networks” In 2nd International Conference on Learning Representations, ICLR 2014 - Conference Track Proceedings, 2014 arXiv:1312.6211
- “Safety Validation of Autonomous Vehicles using Assertion-based Oracles” In arXiv preprint arXiv:2111.04611, 2021
- “A Trustworthiness Score to Evaluate DNN Predictions” In 2023 IEEE International Conference On Artificial Intelligence Testing (AITest), 2023, pp. 9–16 DOI: 10.1109/AITest58265.2023.00011
- Gido M. Ven, Tinne Tuytelaars and Andreas S Tolias “Three types of incremental learning” In Nature Machine Intelligence 4.12 Springer US, 2022, pp. 1185–1197 DOI: 10.1038/s42256-022-00568-3
- “Energy-based models for continual learning” In Conference on Lifelong Learning Agents, 2022, pp. 1–22 PMLR
- Timothée Lesort, Massimo Caccia and Irina Rish “Understanding continual learning settings with data distribution drift analysis” In arXiv preprint arXiv:2104.01678, 2021
- “Task agnostic continual learning using online variational bayes” In arXiv preprint arXiv:1803.10123, 2018
- “Incremental learning algorithms and applications” In European symposium on artificial neural networks (ESANN), 2016
- Li Deng “The mnist database of handwritten digit images for machine learning research” In IEEE Signal Processing Magazine 29.6 IEEE, 2012, pp. 141–142
- Han Xiao, Kashif Rasul and Roland Vollgraf “Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms”, 2017, pp. 1–6 arXiv: http://arxiv.org/abs/1708.07747
- “Model Zoo: A Growing” Brain” That Learns Continually” In arXiv preprint arXiv:2106.03027, 2021
- Nicolas Y Masse, Gregory D Grant and David J Freedman “Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization” In Proceedings of the National Academy of Sciences 115.44 National Acad Sciences, 2018, pp. E10467–E10475
- “ELLA: An efficient lifelong learning algorithm” In International conference on machine learning, 2013, pp. 507–515 PMLR
- “An Efficient Domain-Incremental Learning Approach to Drive in All Weather Conditions” In IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops 2022-June, 2022, pp. 3000–3010 DOI: 10.1109/CVPRW56347.2022.00339
- “CLASSIC: Continual and contrastive learning of aspect sentiment classification tasks” In arXiv preprint arXiv:2112.02714, 2021
- “Few-Shot Class-Incremental Learning” In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2020, pp. 12180–12189 DOI: 10.1109/CVPR42600.2020.01220
- “icarl: Incremental classifier and representation learning” In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2017, pp. 2001–2010
- “The norm must go on: dynamic unsupervised domain adaptation by normalization” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 14765–14775
- “Test-time training with self-supervision for generalization under distribution shifts” In International conference on machine learning, 2020, pp. 9229–9248 PMLR
- “Unsupervised domain adaptation through self-supervision” In arXiv preprint arXiv:1909.11825, 2019
- “Improving robustness against common corruptions by covariate shift adaptation” In Advances in neural information processing systems 33, 2020, pp. 11539–11551
- “Evaluating prediction-time batch normalization for robustness under covariate shift” In arXiv preprint arXiv:2006.10963, 2020
- “Autodial: Automatic domain alignment layers” In Proceedings of the IEEE international conference on computer vision, 2017, pp. 5067–5075
- Friedemann Zenke, Ben Poole and Surya Ganguli “Continual learning through synaptic intelligence” In 34th International Conference on Machine Learning, ICML 2017 8, 2017, pp. 6072–6082 arXiv:1703.04200
- “Learning without Forgetting” In IEEE Transactions on Pattern Analysis and Machine Intelligence 40.12 IEEE, 2018, pp. 2935–2947 DOI: 10.1109/TPAMI.2017.2773081
- “Overcoming catastrophic forgetting in neural networks” In Proceedings of the national academy of sciences 114.13 National Acad Sciences, 2017, pp. 3521–3526
- “Measuring catastrophic forgetting in neural networks” In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, 2018, pp. 3390–3398 arXiv:1708.02072
- Ian J. Goodfellow, Yoshua Bengio and Aaron Courville “Deep Learning” http://www.deeplearningbook.org Cambridge, MA, USA: MIT Press, 2016
- Ferenc Huszár “Note on the quadratic penalties in elastic weight consolidation” In Proceedings of the National Academy of Sciences of the United States of America 115.11, 2018, pp. E2496–E2497 DOI: 10.1073/pnas.1717042115
- David J.C. MacKay “Information Theory, Inference & Learning Algorithms” USA: Cambridge University Press, 2002
- Steven M. Kay “Fundamentals of Statistical Signal Processing: Estimation Theory”, 1993, pp. 180
- “A Tutorial on Fisher information” In Journal of Mathematical Psychology 80, 2017, pp. 40–55 DOI: 10.1016/j.jmp.2017.05.006
- James Martens “New insights and perspectives on the natural gradient method” In Journal of Machine Learning Research 21, 2020, pp. 1–76 arXiv:1412.1193
- “Pytorch: An imperative style, high-performance deep learning library” In Advances in neural information processing systems 32, 2019
- “Benchmarking neural network robustness to common corruptions and perturbations” In ICLR, 2019
- “Learning multiple layers of features from tiny images” Toronto, ON, Canada, 2009
- “Imagenet: A large-scale hierarchical image database” In 2009 IEEE conference on computer vision and pattern recognition, 2009, pp. 248–255 Ieee
- “Deep residual learning for image recognition” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778
- “Evaluation Metrics for DNNs Compression” In arXiv preprint arXiv:2305.10616, 2023
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.