Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 90 tok/s
Gemini 2.5 Pro 57 tok/s Pro
GPT-5 Medium 27 tok/s
GPT-5 High 22 tok/s Pro
GPT-4o 101 tok/s
GPT OSS 120B 467 tok/s Pro
Kimi K2 163 tok/s Pro
2000 character limit reached

Learning Interpretable Models Through Multi-Objective Neural Architecture Search (2112.08645v4)

Published 16 Dec 2021 in cs.LG, cs.AI, and cs.NE

Abstract: Monumental advances in deep learning have led to unprecedented achievements across various domains. While the performance of deep neural networks is indubitable, the architectural design and interpretability of such models are nontrivial. Research has been introduced to automate the design of neural network architectures through neural architecture search (NAS). Recent progress has made these methods more pragmatic by exploiting distributed computation and novel optimization algorithms. However, there is little work in optimizing architectures for interpretability. To this end, we propose a multi-objective distributed NAS framework that optimizes for both task performance and "introspectability," a surrogate metric for aspects of interpretability. We leverage the non-dominated sorting genetic algorithm (NSGA-II) and explainable AI (XAI) techniques to reward architectures that can be better comprehended by domain experts. The framework is evaluated on several image classification datasets. We demonstrate that jointly optimizing for task error and introspectability leads to more disentangled and debuggable architectures that perform within tolerable error.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Understanding neural architecture search techniques. CoRR, abs/1904.00438, 2019.
  2. An evolutionary algorithm that constructs recurrent neural networks. IEEE Trans. Neural Networks, 5(1):54–65, 1994.
  3. Deephyper: Asynchronous hyperparameter search for deep neural networks. In 25th IEEE International Conference on High Performance Computing, HiPC 2018, Bengaluru, India, December 17-20, 2018, pages 42–51. IEEE, 2018.
  4. Understanding the role of individual units in a deep neural network. Proc. Natl. Acad. Sci. USA, 117(48):30071–30078, 2020.
  5. Making a science of model search: Hyperparameter optimization in hundreds of dimensions for vision architectures. In Proceedings of the 30th International Conference on Machine Learning, ICML 2013, Atlanta, GA, USA, 16-21 June 2013, volume 28 of JMLR Workshop and Conference Proceedings, pages 115–123. JMLR.org, 2013.
  6. pymoo: Multi-objective optimization in python. IEEE Access, 8:89497–89509, 2020.
  7. Efficient architecture search by network transformation. In Sheila A. McIlraith and Kilian Q. Weinberger, editors, Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, pages 2787–2794. AAAI Press, 2018.
  8. A downsampled variant of imagenet as an alternative to the CIFAR datasets. CoRR, abs/1707.08819, 2017.
  9. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Transactions on Evolutionary Computation, 6(2):182–197, 2002.
  10. Self-adaptive simulated binary crossover for real-parameter optimization. In Proceedings of the 9th Annual Conference on Genetic and Evolutionary Computation, GECCO ’07, page 1187–1194, New York, NY, USA, 2007. Association for Computing Machinery.
  11. Reference point based multi-objective optimization using evolutionary algorithms. In Proceedings of the 8th Annual Conference on Genetic and Evolutionary Computation, GECCO ’06, page 635–642, New York, NY, USA, 2006. Association for Computing Machinery.
  12. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, pages 248–255. IEEE Computer Society, 2009.
  13. Improved regularization of convolutional neural networks with cutout. CoRR, abs/1708.04552, 2017.
  14. Speeding up automatic hyperparameter optimization of deep neural networks by extrapolation of learning curves. In Qiang Yang and Michael J. Wooldridge, editors, Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, IJCAI, pages 3460–3468. AAAI Press, 2015.
  15. PPP-net: Platform-aware progressive search for Pareto-optimal neural architectures. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Workshop Track Proceedings. OpenReview.net, 2018.
  16. Nas-bench-201: Extending the scope of reproducible neural architecture search. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020.
  17. Efficient multi-objective neural architecture search via lamarckian evolution. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
  18. Neural architecture search: A survey. The Journal of Machine Learning Research, 20(1):1997–2017, 2019.
  19. Visualizing higher-layer features of a deep network. University of Montreal, 1341(3):1, 2009.
  20. An improved dimension-sweep algorithm for the hypervolume indicator. In IEEE International Conference on Evolutionary Computation, CEC 2006, part of WCCI 2006, Vancouver, BC, Canada, 16-21 July 2006, pages 1157–1163. IEEE, 2006.
  21. Neuron shapley: Discovering the responsible neurons. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin, editors, Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020.
  22. David Gunning. Darpa’s explainable artificial intelligence (xai) program. In Proceedings of the 24th International Conference on Intelligent User Interfaces, IUI ’19, page ii, New York, NY, USA, 2019. Association for Computing Machinery.
  23. Identity mappings in deep residual networks. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, Computer Vision - ECCV 2016 - 14th European Conference, volume 9908 of Lecture Notes in Computer Science, pages 630–645. Springer, 2016.
  24. beta-vae: Learning basic visual concepts with a constrained variational framework. In 5th International Conference on Learning Representations, ICLR 2017. OpenReview.net, 2017.
  25. John H. Holland. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence. The MIT Press, 04 1992.
  26. MONAS: multi-objective neural architecture search using reinforcement learning. CoRR, abs/1806.10332, 2018.
  27. Densely connected convolutional networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pages 2261–2269. IEEE Computer Society, 2017.
  28. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (TCAV). In Jennifer G. Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 2673–2682. PMLR, 2018.
  29. Auto-encoding variational bayes. In Yoshua Bengio and Yann LeCun, editors, 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014.
  30. Understanding black-box predictions via influence functions. In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, volume 70 of Proceedings of Machine Learning Research, pages 1885–1894. PMLR, 2017.
  31. DARTS: differentiable architecture search. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
  32. SGDR: stochastic gradient descent with warm restarts. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net, 2017.
  33. NSGA-Net: neural architecture search using multi-objective genetic algorithm. In Anne Auger and Thomas Stützle, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), pages 419–427. ACM, 2019.
  34. Survey of multi-objective optimization methods for engineering. Structural and Multidisciplinary Optimization, 26(6):369–395, Apr 2004.
  35. Towards automatically-tuned neural networks. In Frank Hutter, Lars Kotthoff, and Joaquin Vanschoren, editors, Proceedings of the 2016 Workshop on Automatic Machine Learning (AutoML), volume 64 of JMLR Workshop and Conference Proceedings, pages 58–65. JMLR.org, 2016.
  36. George A Miller. WordNet: a lexical database for English. Communications of the ACM, 38(11):39–41, 1995.
  37. Designing neural networks using genetic algorithms. In J. David Schaffer, editor, Proceedings of the 3rd International Conference on Genetic Algorithms, pages 379–384. Morgan Kaufmann, 1989.
  38. Ray: A distributed framework for emerging AI applications. In Andrea C. Arpaci-Dusseau and Geoff Voelker, editors, 13th USENIX Symposium on Operating Systems Design and Implementation, pages 561–577. USENIX Association, 2018.
  39. Regularized evolution for image classifier architecture search. In The Thirty-Third AAAI Conference on Artificial Intelligence, pages 4780–4789. AAAI Press, 2019.
  40. Large-scale evolution of image classifiers. In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 2902–2911. PMLR, 2017.
  41. Interpretable neural architecture search via bayesian optimisation with weisfeiler-lehman kernels. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021.
  42. Why are we using black box models in ai when we don’t need to? a lesson from an explainable ai competition. Harvard Data Science Review, 1(2), 11 2019.
  43. Evolving neural networks through augmenting topologies. Evolutionary Computation, 10(2):99–127, 2002.
  44. Modularized morphing of deep convolutional neural networks: A graph approach. IEEE Trans. Computers, 70(2):305–315, 2021.
  45. Understanding neural networks through deep visualization. CoRR, abs/1506.06579, 2015.
  46. Interpretable convolutional neural networks. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, pages 8827–8836. Computer Vision Foundation / IEEE Computer Society, 2018.
  47. Disentangled neural architecture search. CoRR, abs/2009.13266, 2020.
  48. Multiobjective optimization using evolutionary algorithms - A comparative case study. In A. E. Eiben, Thomas Bäck, Marc Schoenauer, and Hans-Paul Schwefel, editors, Parallel Problem Solving from Nature - PPSN V, 5th International Conference, volume 1498 of Lecture Notes in Computer Science, pages 292–304. Springer, 1998.
  49. Neural architecture search with reinforcement learning. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net, 2017.
  50. Learning transferable architectures for scalable image recognition. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, pages 8697–8710. Computer Vision Foundation / IEEE Computer Society, 2018.
Citations (9)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube