Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey on Knowledge Editing of Neural Networks (2310.19704v3)

Published 30 Oct 2023 in cs.LG and cs.AI

Abstract: Deep neural networks are becoming increasingly pervasive in academia and industry, matching and surpassing human performance on a wide variety of fields and related tasks. However, just as humans, even the largest artificial neural networks make mistakes, and once-correct predictions can become invalid as the world progresses in time. Augmenting datasets with samples that account for mistakes or up-to-date information has become a common workaround in practical applications. However, the well-known phenomenon of catastrophic forgetting poses a challenge in achieving precise changes in the implicitly memorized knowledge of neural network parameters, often requiring a full model re-training to achieve desired behaviors. That is expensive, unreliable, and incompatible with the current trend of large self-supervised pre-training, making it necessary to find more efficient and effective methods for adapting neural network models to changing data. To address this need, knowledge editing is emerging as a novel area of research that aims to enable reliable, data-efficient, and fast changes to a pre-trained target model, without affecting model behaviors on previously learned tasks. In this survey, we provide a brief review of this recent artificial intelligence field of research. We first introduce the problem of editing neural networks, formalize it in a common framework and differentiate it from more notorious branches of research such as continuous learning. Next, we provide a review of the most relevant knowledge editing approaches and datasets proposed so far, grouping works under four different families: regularization techniques, meta-learning, direct model editing, and architectural strategies. Finally, we outline some intersections with other fields of research and potential directions for future works.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (110)
  1. Branch-specific dendritic ca2+ spikes cause persistent synaptic plasticity. Nature, 520(7546):180–185, 2015.
  2. Roger Ratcliff. Connectionist models of recognition memory: constraints imposed by learning and forgetting functions. Psychological review, 97(2):285, 1990.
  3. Computational principles of synaptic memory consolidation. Nature neuroscience, 19(12):1697–1706, 2016.
  4. Stably maintained dendritic spines are associated with lifelong memories. Nature, 462(7275):920–924, 2009.
  5. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13):3521–3526, 2017.
  6. Transformer-patcher: One mistake worth one neuron. arXiv preprint arXiv:2301.09785, 2023.
  7. The bottom-up evolution of representations in the transformer: A study with machine translation and language modeling objectives. In 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, pages 4387–4397. Association for Computational Linguistics (ACL), 2019.
  8. Analysis methods in neural language processing: A survey. Transactions of the Association for Computational Linguistics, 7:49–72, 2019.
  9. Editable neural networks. arXiv preprint arXiv:2004.00345, 2020.
  10. Detecting formal thought disorder by deep contextualized word representations. Psychiatry Research, 304:114135, 2021.
  11. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  12. Dinov2: Learning robust visual features without supervision. arXiv preprint arXiv:2304.07193, 2023.
  13. A survey of transfer learning. Journal of Big data, 3(1):1–40, 2016.
  14. A survey of large language models. arXiv preprint arXiv:2303.18223, 2023.
  15. Pali: A jointly-scaled multilingual language-image model. arXiv preprint arXiv:2209.06794, 2022.
  16. Mind the gap: Assessing temporal generalization in neural language models. Advances in Neural Information Processing Systems, 34:29348–29363, 2021.
  17. Editing factual knowledge in language models. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6491–6506, 2021.
  18. Modifying memories in transformer models. arXiv preprint arXiv:2012.00363, 2020.
  19. Fast model editing at scale. In International Conference on Learning Representations, 2022a. URL https://openreview.net/pdf?id=0DcZxeWfOPt.
  20. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  21. Alexatm 20b: Few-shot learning using a large-scale multilingual seq2seq model. arXiv preprint arXiv:2208.01448, 2022.
  22. Zero-shot relation extraction via reading comprehension. In Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL 2017), pages 333–342, 2017.
  23. Fever: a large-scale dataset for fact extraction and verification. In Proceedings of the 16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Sheffield, 2018.
  24. Machine learning in medical applications: A review of state-of-the-art methods. Computers in Biology and Medicine, 145:105458, 2022.
  25. Artificial intelligence, machine learning and deep learning in advanced robotics, a review. Cognitive Robotics, 2023.
  26. Machine learning applications for precision agriculture: A comprehensive review. IEEE Access, 9:4843–4873, 2020.
  27. Editing large language models: Problems, methods, and opportunities. arXiv preprint arXiv:2305.13172, 2023.
  28. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  29. Mass-editing memory in a transformer. arXiv preprint arXiv:2210.07229, 2022a.
  30. Memory-based model editing at scale. In International Conference on Machine Learning, pages 15817–15831. PMLR, 2022b.
  31. Aging with grace: Lifelong model editing with discrete key-value adaptors. arXiv preprint arXiv:2211.11031, 2022.
  32. Tom Michael Mitchell et al. Machine learning, volume 1. McGraw-hill New York, 2007.
  33. Locating and editing factual associations in gpt. Advances in Neural Information Processing Systems, 35:17359–17372, 2022b.
  34. Conditional neural processes. In International conference on machine learning, pages 1704–1713. PMLR, 2018.
  35. Richard Szeliski. Computer vision: algorithms and applications. Springer Nature, 2022.
  36. Deep learning for computer vision: A brief review. Computational intelligence and neuroscience, 2018, 2018.
  37. Deep learning vs. traditional computer vision. In Advances in Computer Vision: Proceedings of the 2019 Computer Vision Conference (CVC), Volume 1 1, pages 128–144. Springer, 2020.
  38. Michael Firman. Rgbd datasets: Past, present and future. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pages 19–31, 2016.
  39. Provable repair of deep neural networks. In Proceedings of the 42nd ACM SIGPLAN International Conference on Programming Language Design and Implementation, pages 588–603, 2021.
  40. Mnist-c: A robustness benchmark for computer vision. arXiv preprint arXiv:1906.02337, 2019.
  41. Mnist handwritten digit database. ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, 2, 2010.
  42. Alex Krizhevsky. Learning multiple layers of features from tiny images. 2009. URL https://api.semanticscholar.org/CorpusID:18268744.
  43. Cifar-10 (canadian institute for advanced research). 2009. URL http://www. cs. toronto. edu/kriz/cifar. html, 5, 2009.
  44. Overcoming catastrophic forgetting with unlabeled data in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 312–321, 2019.
  45. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 248–255, 2009a. doi:10.1109/CVPR.2009.5206848.
  46. ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 115(3):211–252, 2015. doi:10.1007/s11263-015-0816-y.
  47. Natural adversarial examples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15262–15271, 2021.
  48. Deep learning face attributes in the wild. In Proceedings of International Conference on Computer Vision (ICCV), December 2015.
  49. Jacob Eisenstein. Introduction to Natural Language Processing. MIT Press, 2019.
  50. A survey of the usages of deep learning for natural language processing. IEEE transactions on neural networks and learning systems, 32(2):604–624, 2020.
  51. Alexa teacher model: Pretraining and distilling multi-billion-parameter encoders for natural language understanding systems. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD ’22, page 2893–2902, New York, NY, USA, 2022. Association for Computing Machinery. ISBN 9781450393850. doi:10.1145/3534678.3539173. URL https://doi.org/10.1145/3534678.3539173.
  52. Pre-trained language models and their applications. Engineering, 2022.
  53. A survey on automated fact-checking. Transactions of the Association for Computational Linguistics, 10:178–206, 2022.
  54. Shaking the foundations: delusions in sequence models for interaction and control. arXiv preprint arXiv:2110.10819, 2021.
  55. Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12):1–38, 2023.
  56. Get your vitamin C! robust fact verification with contrastive evidence. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 624–643, Online, June 2021. Association for Computational Linguistics. doi:10.18653/v1/2021.naacl-main.52. URL https://aclanthology.org/2021.naacl-main.52.
  57. How much knowledge can you pack into the parameters of a language model? arXiv preprint arXiv:2002.08910, 2020.
  58. T-REx: A large scale alignment of natural language with knowledge base triples. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018), Miyazaki, Japan, May 2018. European Language Resources Association (ELRA). URL https://www.aclweb.org/anthology/L18-1544.
  59. Natural questions: A benchmark for question answering research. Transactions of the Association for Computational Linguistics, 7:452–466, 2019. doi:10.1162/tacl_a_00276. URL https://aclanthology.org/Q19-1026.
  60. Pointer sentinel mixture models. arXiv preprint arXiv:1609.07843, 2016.
  61. FairLex: A multilingual benchmark for evaluating fairness in legal text processing. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4389–4406, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi:10.18653/v1/2022.acl-long.301. URL https://aclanthology.org/2022.acl-long.301.
  62. John C Knight. Safety critical systems: challenges and directions. In Proceedings of the 24th international conference on software engineering, pages 547–550, 2002.
  63. Deep learning for object detection and scene perception in self-driving cars: Survey, challenges, and open issues. Array, 10:100057, 2021.
  64. Ahmet İlker Tekkeşin et al. Artificial intelligence in healthcare: past, present and future. Anatol J Cardiol, 22(Suppl 2):8–9, 2019.
  65. Banavar Sridhar. Applications of machine learning techniques to aviation operations: Promises and challenges. In 2020 International Conference on Artificial Intelligence and Data Analytics for Air Transportation (AIDA-AT), pages 1–12. IEEE, 2020.
  66. Guaranteeing safety for neural network-based aircraft collision avoidance systems. In 2019 IEEE/AIAA 38th Digital Avionics Systems Conference (DASC), pages 1–10. IEEE, 2019.
  67. Sound and complete neural network repair with minimality and locality guarantees. arXiv preprint arXiv:2110.07682, 2021.
  68. Repairing deep neural networks based on behavior imitation. arXiv preprint arXiv:2305.03365, 2023.
  69. Mykel J Kochenderfer and JP Chryssanthacopoulos. Robust airborne collision avoidance through dynamic programming. Massachusetts Institute of Technology, Lincoln Laboratory, Project Report ATC-371, 130, 2011.
  70. Reluplex: An efficient smt solver for verifying deep neural networks. In Computer Aided Verification: 29th International Conference, CAV 2017, Heidelberg, Germany, July 24-28, 2017, Proceedings, Part I 30, pages 97–117. Springer, 2017.
  71. Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine, 34(4):18–42, 2017.
  72. Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv preprint arXiv:2104.13478, 2021.
  73. A comprehensive survey on graph neural networks. IEEE transactions on neural networks and learning systems, 32(1):4–24, 2020.
  74. Editable graph neural network for node classifications. arXiv preprint arXiv:2305.15529, 2023.
  75. Automating the construction of internet portals with machine learning. Information Retrieval, 3:127–163, 2000.
  76. Inductive representation learning on large graphs. Advances in neural information processing systems, 30, 2017.
  77. Measuring catastrophic forgetting in neural networks. In Proceedings of the AAAI conference on artificial intelligence, volume 32, 2018.
  78. Lifelong learning with dynamically expandable networks. arXiv preprint arXiv:1708.01547, 2017.
  79. Overcoming catastrophic forgetting with hard attention to the task. In International Conference on Machine Learning, pages 4548–4557. PMLR, 2018.
  80. A wholistic view of continual learning with deep neural networks: Forgotten lessons and the bridge to active and open world learning. Neural Networks, 2023.
  81. Model-agnostic meta-learning for fast adaptation of deep networks. In International conference on machine learning, pages 1126–1135. PMLR, 2017.
  82. Meta-learning for semi-supervised few-shot classification. In International Conference on Learning Representations, 2018.
  83. Meta-learning for low-resource neural machine translation. In 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018, pages 3622–3631. Association for Computational Linguistics, 2020.
  84. Evolved policy gradients. Advances in Neural Information Processing Systems, 31, 2018.
  85. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009b.
  86. Report on the 11th iwslt evaluation campaign. In Proceedings of the 11th International Workshop on Spoken Language Translation: Evaluation Campaign, pages 2–17, 2014.
  87. Hypernetworks. arXiv preprint arXiv:1609.09106, 2016.
  88. Long short-term memory. Neural Comput, 9(8):1735–1780, 1997.
  89. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880, 2020.
  90. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pages 249–256. JMLR Workshop and Conference Proceedings, 2010.
  91. FiLM: Visual Reasoning with a General Conditioning Layer. In AAAI Conference on Artificial Intelligence, New Orleans, United States, February 2018. URL https://inria.hal.science/hal-01648685.
  92. Improving language understanding by generative pre-training. 2018.
  93. A mathematical framework for transformer circuits. Transformer Circuits Thread, 1, 2021.
  94. Analyzing transformers in embedding space. In Annual Meeting of the Association for Computational Linguistics, 2023.
  95. Transformer memory as a differentiable search index. Advances in Neural Information Processing Systems, 35:21831–21843, 2022.
  96. End-to-end memory networks. Advances in neural information processing systems, 28, 2015.
  97. Knowledge neurons in pretrained transformers. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8493–8502, 2022.
  98. Judea Pearl. Direct and indirect effects. In Proceedings of the Seventeenth Conference on Uncertainty in Artificial Intelligence, UAI’01, page 411–420, San Francisco, CA, USA, 2001. Morgan Kaufmann Publishers Inc. ISBN 1558608001.
  99. Investigating gender bias in language models using causal mediation analysis. Advances in neural information processing systems, 33:12388–12401, 2020.
  100. Generating informative and diverse conversational responses via adversarial information maximization. Advances in Neural Information Processing Systems, 31, 2018.
  101. Editing commonsense knowledge in gpt. arXiv preprint arXiv:2305.14956, 2023.
  102. Does localization inform editing? surprising differences in causality-based localization vs. knowledge editing in language models. arXiv preprint arXiv:2301.04213, 2023.
  103. Neural turing machines. arXiv preprint arXiv:1410.5401, 2014.
  104. Meta-learning with memory-augmented neural networks. In International conference on machine learning, pages 1842–1850. PMLR, 2016.
  105. Calibrating factual knowledge in pretrained language models. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 5937–5947, 2022.
  106. Measuring and improving consistency in pretrained language models. Transactions of the Association for Computational Linguistics, 9:1012–1031, 2021.
  107. LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=nZeVKeeFYf9.
  108. A principled approach to failure analysis and model repairment: Demonstration in medical imaging. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part III 24, pages 509–518. Springer, 2021.
  109. Backpack language models. arXiv preprint arXiv:2305.16765, 2023.
  110. A comprehensive survey on pretrained foundation models: A history from bert to chatgpt. arXiv preprint arXiv:2302.09419, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Vittorio Mazzia (17 papers)
  2. Alessandro Pedrani (1 paper)
  3. Andrea Caciolai (1 paper)
  4. Kay Rottmann (3 papers)
  5. Davide Bernardi (6 papers)
Citations (19)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets