Papers
Topics
Authors
Recent
2000 character limit reached

DIRA: Dynamic Domain Incremental Regularised Adaptation (2205.00147v5)

Published 30 Apr 2022 in cs.LG

Abstract: Autonomous systems (AS) often use Deep Neural Network (DNN) classifiers to allow them to operate in complex, high-dimensional, non-linear, and dynamically changing environments. Due to the complexity of these environments, DNN classifiers may output misclassifications during operation when they face domains not identified during development. Removing a system from operation for retraining becomes impractical as the number of such AS increases. To increase AS reliability and overcome this limitation, DNN classifiers need to have the ability to adapt during operation when faced with different operational domains using a few samples (e.g. 2 to 100 samples). However, retraining DNNs on a few samples is known to cause catastrophic forgetting and poor generalisation. In this paper, we introduce Dynamic Incremental Regularised Adaptation (DIRA), an approach for dynamic operational domain adaption of DNNs using regularisation techniques. We show that DIRA improves on the problem of forgetting and achieves strong gains in performance when retraining using a few samples from the target domain. Our approach shows improvements on different image classification benchmarks aimed at evaluating robustness to distribution shifts (e.g.CIFAR-10C/100C, ImageNet-C), and produces state-of-the-art performance in comparison with other methods from the literature.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Nidhi Kalra and Susan M. Paddock “Driving to Safety: How Many Miles of Driving Would It Take to Demonstrate Autonomous Vehicle Reliability?” Santa Monica, CA: RAND Corporation, 2016 DOI: 10.7249/RR1478
  2. “An agency-directed approach to test generation for simulation-based autonomous vehicle verification” In 2020 IEEE International Conference On Artificial Intelligence Testing (AITest), 2020, pp. 31–38 IEEE
  3. Kerstin I Eder, Wen-ling Huang and Jan Peleska “Complete Agent-driven Model-based System Testing for Autonomous Systems” In arXiv preprint arXiv:2110.12586, 2021
  4. “Assessing Trustworthiness of Autonomous Systems”, 2023 arXiv:2305.03411 [cs.AI]
  5. “Positive Trust Balance for Self-driving Car Deployment” Springer International Publishing, 2020, pp. 351–357 DOI: 10.1007/978-3-030-55583-2˙26
  6. Darryl Hond, Hamid Asgari and Daniel Jeffery “Verifying Artificial Neural Network Classifier Performance Using Dataset Dissimilarity Measures” In Proceedings - 19th IEEE International Conference on Machine Learning and Applications, ICMLA 2020, 2020, pp. 115–121 DOI: 10.1109/ICMLA51294.2020.00027
  7. J.David Schaffer and Walker H. Land “Predicting with Confidence: Classifiers that Know What They Don’t Know” In Procedia Computer Science 114 Elsevier B.V., 2017, pp. 200–207 DOI: 10.1016/j.procs.2017.09.061
  8. “Distance-based Confidence Score for Neural Network Classifiers”, 2017 arXiv: http://arxiv.org/abs/1709.09844
  9. “Distance-Based Learning from Errors for Confidence Calibration”, 2019, pp. 1–12 arXiv: http://arxiv.org/abs/1912.01730
  10. “The oracle problem in software testing: A survey” In IEEE Transactions on Software Engineering 41.5 IEEE, 2015, pp. 507–525 DOI: 10.1109/TSE.2014.2372785
  11. “Machine Learning Testing: Survey, Landscapes and Horizons” In IEEE Transactions on Software Engineering, 2020, pp. 1–1 DOI: 10.1109/tse.2019.2962027
  12. “An empirical investigation of catastrophic forgetting in gradient-based neural networks” In 2nd International Conference on Learning Representations, ICLR 2014 - Conference Track Proceedings, 2014 arXiv:1312.6211
  13. “Safety Validation of Autonomous Vehicles using Assertion-based Oracles” In arXiv preprint arXiv:2111.04611, 2021
  14. “A Trustworthiness Score to Evaluate DNN Predictions” In 2023 IEEE International Conference On Artificial Intelligence Testing (AITest), 2023, pp. 9–16 DOI: 10.1109/AITest58265.2023.00011
  15. Gido M. Ven, Tinne Tuytelaars and Andreas S Tolias “Three types of incremental learning” In Nature Machine Intelligence 4.12 Springer US, 2022, pp. 1185–1197 DOI: 10.1038/s42256-022-00568-3
  16. “Energy-based models for continual learning” In Conference on Lifelong Learning Agents, 2022, pp. 1–22 PMLR
  17. Timothée Lesort, Massimo Caccia and Irina Rish “Understanding continual learning settings with data distribution drift analysis” In arXiv preprint arXiv:2104.01678, 2021
  18. “Task agnostic continual learning using online variational bayes” In arXiv preprint arXiv:1803.10123, 2018
  19. “Incremental learning algorithms and applications” In European symposium on artificial neural networks (ESANN), 2016
  20. Li Deng “The mnist database of handwritten digit images for machine learning research” In IEEE Signal Processing Magazine 29.6 IEEE, 2012, pp. 141–142
  21. Han Xiao, Kashif Rasul and Roland Vollgraf “Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms”, 2017, pp. 1–6 arXiv: http://arxiv.org/abs/1708.07747
  22. “Model Zoo: A Growing” Brain” That Learns Continually” In arXiv preprint arXiv:2106.03027, 2021
  23. Nicolas Y Masse, Gregory D Grant and David J Freedman “Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization” In Proceedings of the National Academy of Sciences 115.44 National Acad Sciences, 2018, pp. E10467–E10475
  24. “ELLA: An efficient lifelong learning algorithm” In International conference on machine learning, 2013, pp. 507–515 PMLR
  25. “An Efficient Domain-Incremental Learning Approach to Drive in All Weather Conditions” In IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops 2022-June, 2022, pp. 3000–3010 DOI: 10.1109/CVPRW56347.2022.00339
  26. “CLASSIC: Continual and contrastive learning of aspect sentiment classification tasks” In arXiv preprint arXiv:2112.02714, 2021
  27. “Few-Shot Class-Incremental Learning” In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2020, pp. 12180–12189 DOI: 10.1109/CVPR42600.2020.01220
  28. “icarl: Incremental classifier and representation learning” In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2017, pp. 2001–2010
  29. “The norm must go on: dynamic unsupervised domain adaptation by normalization” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 14765–14775
  30. “Test-time training with self-supervision for generalization under distribution shifts” In International conference on machine learning, 2020, pp. 9229–9248 PMLR
  31. “Unsupervised domain adaptation through self-supervision” In arXiv preprint arXiv:1909.11825, 2019
  32. “Improving robustness against common corruptions by covariate shift adaptation” In Advances in neural information processing systems 33, 2020, pp. 11539–11551
  33. “Evaluating prediction-time batch normalization for robustness under covariate shift” In arXiv preprint arXiv:2006.10963, 2020
  34. “Autodial: Automatic domain alignment layers” In Proceedings of the IEEE international conference on computer vision, 2017, pp. 5067–5075
  35. Friedemann Zenke, Ben Poole and Surya Ganguli “Continual learning through synaptic intelligence” In 34th International Conference on Machine Learning, ICML 2017 8, 2017, pp. 6072–6082 arXiv:1703.04200
  36. “Learning without Forgetting” In IEEE Transactions on Pattern Analysis and Machine Intelligence 40.12 IEEE, 2018, pp. 2935–2947 DOI: 10.1109/TPAMI.2017.2773081
  37. “Overcoming catastrophic forgetting in neural networks” In Proceedings of the national academy of sciences 114.13 National Acad Sciences, 2017, pp. 3521–3526
  38. “Measuring catastrophic forgetting in neural networks” In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, 2018, pp. 3390–3398 arXiv:1708.02072
  39. Ian J. Goodfellow, Yoshua Bengio and Aaron Courville “Deep Learning” http://www.deeplearningbook.org Cambridge, MA, USA: MIT Press, 2016
  40. Ferenc Huszár “Note on the quadratic penalties in elastic weight consolidation” In Proceedings of the National Academy of Sciences of the United States of America 115.11, 2018, pp. E2496–E2497 DOI: 10.1073/pnas.1717042115
  41. David J.C. MacKay “Information Theory, Inference & Learning Algorithms” USA: Cambridge University Press, 2002
  42. Steven M. Kay “Fundamentals of Statistical Signal Processing: Estimation Theory”, 1993, pp. 180
  43. “A Tutorial on Fisher information” In Journal of Mathematical Psychology 80, 2017, pp. 40–55 DOI: 10.1016/j.jmp.2017.05.006
  44. James Martens “New insights and perspectives on the natural gradient method” In Journal of Machine Learning Research 21, 2020, pp. 1–76 arXiv:1412.1193
  45. “Pytorch: An imperative style, high-performance deep learning library” In Advances in neural information processing systems 32, 2019
  46. “Benchmarking neural network robustness to common corruptions and perturbations” In ICLR, 2019
  47. “Learning multiple layers of features from tiny images” Toronto, ON, Canada, 2009
  48. “Imagenet: A large-scale hierarchical image database” In 2009 IEEE conference on computer vision and pattern recognition, 2009, pp. 248–255 Ieee
  49. “Deep residual learning for image recognition” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778
  50. “Evaluation Metrics for DNNs Compression” In arXiv preprint arXiv:2305.10616, 2023
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.