Interactive Continual Learning Architecture for Long-Term Personalization of Home Service Robots (2403.03462v1)
Abstract: For robots to perform assistive tasks in unstructured home environments, they must learn and reason on the semantic knowledge of the environments. Despite a resurgence in the development of semantic reasoning architectures, these methods assume that all the training data is available a priori. However, each user's environment is unique and can continue to change over time, which makes these methods unsuitable for personalized home service robots. Although research in continual learning develops methods that can learn and adapt over time, most of these methods are tested in the narrow context of object classification on static image datasets. In this paper, we combine ideas from continual learning, semantic reasoning, and interactive machine learning literature and develop a novel interactive continual learning architecture for continual learning of semantic knowledge in a home environment through human-robot interaction. The architecture builds on core cognitive principles of learning and memory for efficient and real-time learning of new knowledge from humans. We integrate our architecture with a physical mobile manipulator robot and perform extensive system evaluations in a laboratory environment over two months. Our results demonstrate the effectiveness of our architecture to allow a physical robot to continually adapt to the changes in the environment from limited data provided by the users (experimenters), and use the learned knowledge to perform object fetching tasks.
- O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, “Imagenet large scale visual recognition challenge,” Int. J. Comput. Vision, vol. 115, no. 3, pp. 211–252, Dec. 2015.
- R. Girshick, “Fast r-cnn,” in The IEEE International Conference on Computer Vision (ICCV), December 2015.
- N. Fulda, N. Tibbetts, Z. Brown, and D. Wingate, “Harvesting common-sense navigational knowledge for robotics from uncurated text corpora,” in Conference on Robot Learning. PMLR, 2017, pp. 525–534.
- J. Wang, V. A. Shim, R. Yan, H. Tang, and F. Sun, “Automatic object searching and behavior learning for mobile robots in unstructured environment by deep belief networks,” IEEE Transactions on Cognitive and Developmental Systems, vol. 11, no. 3, pp. 395–404, 2018.
- A. Ayub and A. R. Wagner, “Centroid based concept learning for rgb-d indoor scene classification,” in British Machine Vision Conference (BMVC), 2020.
- W. Liu, A. Daruna, M. Patel, K. Ramachandruni, and S. Chernova, “A survey of semantic reasoning frameworks for robotic systems,” Robotics and Autonomous Systems, vol. 159, p. 104294, 2023.
- J. Saunders, D. S. Syrdal, K. L. Koay, N. Burke, and K. Dautenhahn, “Teach Me–Show Me”—End-user personalization of a smart home and companion robot,” IEEE Transactions on Human-Machine Systems, vol. 46, no. 1, pp. 27–40, 2016.
- A. L. Thomaz and M. Cakmak, “Learning about objects with human teachers,” in 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI), 2009, pp. 15–22.
- S.-A. Rebuffi, A. Kolesnikov, G. Sperl, and C. H. Lampert, “iCaRL: Incremental classifier and representation learning,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
- A. Ayub and A. R. Wagner, “Cognitively-inspired model for incremental learning using a few examples,” in The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2020.
- T. L. Hayes, G. P. Krishnan, M. Bazhenov, H. T. Siegelmann, T. J. Sejnowski, and C. Kanan, “Replay in deep learning: Current approaches and missing biological elements,” Neural computation, vol. 33, no. 11, pp. 2908–2950, 2021.
- A. Ayub and A. R. Wagner, “Tell me what this is: Few-shot incremental object learning by a robot,” arXiv:2008.00819, 2020.
- A. Ayub, J. Mehta, Z. De Francesco, P. Holthaus, K. Dautenhahn, and C. L. Nehaniv, “How do human users teach a continual learning robot in repeated interactions?” in IEEE International Conference on Robot and Human Interactive Communication (RO-MAN) (Accepted), 2023.
- A. Daruna, M. Gupta, M. Sridharan, and S. Chernova, “Continual learning of knowledge graph embeddings,” IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 1128–1135, 2021.
- Y. Cui, Y. Wang, Z. Sun, W. Liu, Y. Jiang, K. Han, and W. Hu, “Lifelong embedding learning and transfer for growing knowledge graphs,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 4, 2023, pp. 4217–4224.
- Z. Wang, G. Tian, and X. Shao, “Home service robot task planning using semantic knowledge and probabilistic inference,” Knowledge-Based Systems, vol. 204, p. 106174, 2020.
- A. Daruna, L. Nair, W. Liu, and S. Chernova, “Towards robust one-shot task execution using knowledge graph embeddings,” in 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2021, pp. 11 118–11 124.
- W. Liu, D. Bansal, A. Daruna, and S. Chernova, “Learning instance-level n-ary semantic knowledge at scale for robots operating in everyday environments,” Autonomous Robots, pp. 1–19, 2023.
- S. Liu, G. Tian, Y. Zhang, M. Zhang, and S. Liu, “Service planning oriented efficient object search: A knowledge-based framework for home service robot,” Expert Systems with Applications, vol. 187, p. 115853, 2022.
- M. Wise, M. Ferguson, D. King, E. Diehr, and D. Dymesich, “Fetch and freight: Standard platforms for service robot applications,” in IJCAI, Workshop on Autonomous Mobile Service Robots, 2016.
- A. Ayub and A. R. Wagner, “EEC: Learning to encode and regenerate images for continual learning,” in International Conference on Learning Representations (ICLR), 2021. [Online]. Available: https://openreview.net/forum?id=lWaz5a9lcFU
- X. Shi, D. Li, P. Zhao, Q. Tian, Y. Tian, Q. Long, C. Zhu, J. Song, F. Qiao, L. Song, Y. Guo, Z. Wang, Y. Zhang, B. Qin, W. Yang, F. Wang, R. H. M. Chan, and Q. She, “Are we ready for service robots? the openloris-scene datasets for lifelong slam,” arXiv:1911.05603, 2019.
- V. Lomonaco and D. Maltoni, “Core50: a new dataset and benchmark for continuous object recognition,” in Proceedings of the 1st Annual Conference on Robot Learning, vol. 78, 2017, pp. 17–26.
- R. M. French, “Dynamically constraining connectionist networks to produce distributed, orthogonal representations to reduce catastrophic interference,” Proceedings of the Sixteenth Annual Conference of the Cognitive Science Society, pp. 335–340, 2019.
- Y. Wu, Y. Chen, L. Wang, Y. Ye, Z. Liu, Y. Guo, and Y. Fu, “Large scale incremental learning,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
- J. Kirkpatrick, R. Pascanu, N. C. Rabinowitz, J. Veness, G. Desjardins, A. A. Rusu, K. Milan, J. Quan, T. Ramalho, A. Grabska-Barwinska, D. Hassabis, C. Clopath, D. Kumaran, and R. Hadsell, “Overcoming catastrophic forgetting in neural networks,” Proceedings of the National Academy of Sciences of the United States of America, vol. 114, no. 13, pp. 3521–3526, 2017.
- Z. Li and D. Hoiem, “Learning without forgetting,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 12, pp. 2935–2947, Dec 2018.
- O. Ostapenko, M. Puscas, T. Klein, P. Jahnichen, and M. Nabi, “Learning to remember: A synaptic plasticity driven framework for continual learning,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019, pp. 11 321–11 329.
- X. Tao, X. Chang, X. Hong, X. Wei, and Y. Gong, “Topology-preserving class-incremental learning,” in Computer Vision – ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIX. Springer-Verlag, 2020, p. 254–270.
- M. Dehghan, Z. Zhang, M. Siam, J. Jin, L. Petrich, and M. Jagersand, “Online object and task learning via human robot interaction,” in 2019 International Conference on Robotics and Automation (ICRA), May 2019, pp. 2132–2138.
- A. Ayub and C. Fendley, “Few-shot continual active learning by a robot,” Advances in Neural Information Processing Systems, vol. 35, pp. 30 612–30 624, 2022.
- M. Scheutz, E. Krause, B. Oosterveld, T. Frasca, and R. Platt, “Spoken instruction-based one-shot object and action learning in a cognitive robotic architecture,” in Proceedings of the 16th Conference on Autonomous Agents and MultiAgent Systems, ser. AAMAS ’17. Richland, SC: International Foundation for Autonomous Agents and Multiagent Systems, 2017, p. 1378–1386.
- J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016.
- W.-Y. Chen, Y.-C. Liu, Z. Kira, Y.-C. F. Wang, and J.-B. Huang, “A closer look at few-shot classification,” in International Conference on Learning Representations, 2019.
- J. E. Laird, A. Newell, and P. S. Rosenbloom, “Soar: An architecture for general intelligence,” Artificial intelligence, vol. 33, no. 1, pp. 1–64, 1987.
- B. C. Love, D. L. Medin, and T. M. Gureckis, “Sustain: A network model of category learning.” Psychological Review, vol. 111, no. 2, p. 309–332, 2004.
- M. L. Mack, B. C. Love, and A. R. Preston, “Building concepts one episode at a time: The hippocampus and concept formation,” Neuroscience Letters, vol. 680, pp. 31–38, 2018.
- M. G. Collins, F. Sense, M. Krusmark, and T. S. Jastrzembski, “Improving predictive accuracy of models of learning and retention through bayesian hierarchical modeling: An exploration with the predictive performance equation.” in CogSci, 2020.
- M. M. Walsh, K. A. Gluck, G. Gunzelmann, T. Jastrzembski, and M. Krusmark, “Evaluating the theoretic adequacy and applied potential of computational models of the spacing effect,” Cognitive science, vol. 42, pp. 644–691, 2018.
- J. L. McClelland, B. L. McNaughton, and R. C. O’Reilly, “Why there are complementary learning systems in the hippocampus and neocortex: Insights from the successes and failures of connectionist models of learning and memory.” Psychological Review, vol. 102, no. 3, p. 419–457, 1995.
- T. Kitamura, S. K. Ogawa, D. S. Roy, T. Okuyama, M. D. Morrissey, L. M. Smith, R. L. Redondo, and S. Tonegawa, “Engrams and circuits crucial for systems consolidation of a memory,” Science, vol. 356, no. 6333, pp. 73–78, 2017.
- J. Shah, A. Ayub, C. L. Nehaniv, and K. Dautenhahn, “Where is my phone? towards developing an episodic memory model for companion robots to track users’ salient objects,” in Companion of the 2023 ACM/IEEE International Conference on Human-Robot Interaction, ser. HRI ’23. New York, NY, USA: Association for Computing Machinery, 2023, p. 621–624.
- P. Gärdenfors, “Conceptual spaces as a framework for knowledge representation,” Mind and Matter, vol. 2, no. 2, pp. 9–27, 2004.
- M. Csorba, “Simultaneous localisation and map building,” Ph.D. dissertation, University of Oxford, 1997.
- Y. Jiang, N. Walker, J. Hart, and P. Stone, “Open-world reasoning for service robots,” in Proceedings of the International Conference on Automated Planning and Scheduling, vol. 29, 2019, pp. 725–733.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016.
- C. Zhang, N. Song, G. Lin, Y. Zheng, P. Pan, and Y. Xu, “Few-shot incremental learning with continually evolved classifiers,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2021, pp. 12 455–12 464.
- Ali Ayub (22 papers)
- Chrystopher Nehaniv (2 papers)
- Kerstin Dautenhahn (25 papers)