Robust Fusion for Bayesian Semantic Mapping (2303.07836v2)
Abstract: The integration of semantic information in a map allows robots to understand better their environment and make high-level decisions. In the last few years, neural networks have shown enormous progress in their perception capabilities. However, when fusing multiple observations from a neural network in a semantic map, its inherent overconfidence with unknown data gives too much weight to the outliers and decreases the robustness. To mitigate this issue we propose a novel robust fusion method to combine multiple Bayesian semantic predictions. Our method uses the uncertainty estimation provided by a Bayesian neural network to calibrate the way in which the measurements are fused. This is done by regularizing the observations to mitigate the problem of overconfident outlier predictions and using the epistemic uncertainty to weigh their influence in the fusion, resulting in a different formulation of the probability distributions. We validate our robust fusion strategy by performing experiments on photo-realistic simulated environments and real scenes. In both cases, we use a network trained on different data to expose the model to varying data distributions. The results show that considering the model's uncertainty and regularizing the probability distribution of the observations distribution results in a better semantic segmentation performance and more robustness to outliers, compared with other methods. Video - https://youtu.be/5xVGm7z9c-0
- S. Duan, G. Tian, Z. Wang, S. Liu, and C. Feng, “A semantic robotic grasping framework based on multi-task learning in stacking scenes,” Engineering Applications of Artificial Intelligence, vol. 121, 2023.
- A. Geiger, P. Lenz, and R. Urtasun, “Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite,” in IEEE Conf. on Computer Vision and Pattern Recognition, 2012.
- Z. Wang, G. Tian, and X. Shao, “Home service robot task planning using semantic knowledge and probabilistic inference,” Knowledge-Based Systems, vol. 204, 2020.
- N. Sünderhauf, O. Brock, W. Scheirer, R. Hadsell, D. Fox, J. Leitner, B. Upcroft, P. Abbeel, W. Burgard, M. Milford, et al., “The limits and potentials of deep learning for robotics,” The International journal of robotics research, vol. 37, no. 4-5, 2018.
- C. Guo, G. Pleiss, Y. Sun, and K. Q. Weinberger, “On calibration of modern neural networks,” in Int. Conf. on Machine Learning, 2017.
- J. McCormac, A. Handa, A. Davison, and S. Leutenegger, “SemanticFusion: Dense 3D semantic mapping with convolutional neural networks,” in IEEE Int. Conf. on Robotics and Automation, 2017.
- N. Sunderhauf, T. T. Pham, Y. Latif, M. Milford, and I. Reid, “Meaningful maps with object-oriented semantic mapping,” in IEEE Int. Conf. on Intelligent Robots and Systems, 2017.
- A. Rosinol, M. Abate, Y. Chang, and L. Carlone, “Kimera: An Open-Source Library for Real-Time Metric-Semantic Localization and Mapping,” in IEEE Int. Conf. on Robotics and Automation, may 2020.
- A. Asgharivaskasi and N. Atanasov, “Active Bayesian Multi-class Mapping from Range and Semantic Segmentation Observations,” in IEEE Int. Conf. on Robotics and Automation, 2021.
- D. S. Chaplot, M. Dalal, S. Gupta, J. Malik, and R. R. Salakhutdinov, “Seal: Self-supervised embodied active learning using exploration and 3d consistency,” Adv. in neural information proc. systems, vol. 34, 2021.
- D. Nilsson, A. Pirinen, E. Gärtner, and C. Sminchisescu, “Embodied visual active learning for semantic segmentation,” in AAAI Conference on Artificial Intelligence, no. 3, 2021.
- J. Rückin, L. Jin, F. Magistri, C. Stachniss, and M. Popović, “Informative path planning for active learning in aerial semantic mapping,” in IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, 2022.
- R. Zurbrügg, H. Blum, C. Cadena, R. Siegwart, and L. Schmid, “Embodied active domain adaptation for semantic segmentation via informative path planning,” IEEE Robotics and Automation Letters, vol. 7, 2022.
- Y. Xiang and D. Fox, “DA-RNN: Semantic Mapping with Data Associated Recurrent Neural Networks,” in Robotics: Science and Systems, 2017.
- Z. Yang and C. Liu, “TUPPer-Map: Temporal and Unified Panoptic Perception for 3D Metric-Semantic Mapping,” in IEEE Int. Conf. on Intelligent Robots and Systems, 2021.
- M. Grinvald, F. Furrer, T. Novkovic, J. J. Chung, C. Cadena, R. Siegwart, and J. Nieto, “Volumetric instance-aware semantic mapping and 3d object discovery,” IEEE Robotics and Automation Letters, vol. 4, 2019.
- L. Schmid, J. Delmerico, J. L. Schonberger, J. Nieto, M. Pollefeys, R. Siegwart, and C. Cadena, “Panoptic Multi-TSDFs: a Flexible Representation for Online Multi-resolution Volumetric Mapping and Long-term Dynamic Scene Consistency,” in IEEE ICRA, 2022.
- J. McCormac, R. Clark, M. Bloesch, A. Davison, and S. Leutenegger, “Fusion++: Volumetric Object-Level SLAM,” in Int. Conf. on 3D Vision, 2018.
- G. Narita, T. Seno, T. Ishikawa, and Y. Kaji, “PanopticFusion: Online Volumetric Semantic Mapping at the Level of Stuff and Things,” in IEEE International Conference on Intelligent Robots and Systems. Institute of Electrical and Electronics Engineers Inc., nov 2019.
- Y. Gal and Z. Ghahramani, “Dropout as a bayesian approximation: Representing model uncertainty in deep learning,” in Int. conf. on machine learning, 2016.
- F. K. Gustafsson, M. Danelljan, and T. B. Schon, “Evaluating scalable bayesian deep learning methods for robust computer vision,” in IEEE/CVF CVPR Workshops, 2020.
- B. Lakshminarayanan, A. Pritzel, and C. Blundell, “Simple and scalable predictive uncertainty estimation using deep ensembles,” Advances in neural information processing systems, vol. 30, 2017.
- K. Lee, K. Lee, H. Lee, and J. Shin, “A simple unified framework for detecting out-of-distribution samples and adversarial attacks,” Advances in neural information processing systems, vol. 31, 2018.
- J. Mukhoti, A. Kirsch, J. van Amersfoort, P. H. Torr, and Y. Gal, “Deep deterministic uncertainty: A new simple baseline,” in IEEE/CVF Conf. on Computer Vision and Pattern Recognition, June 2023.
- J. Postels, H. Blum, C. Cadena, R. Siegwart, L. V. Gool, and F. Tombari, “Quantifying aleatoric and epistemic uncertainty using density estimation in latent space,” CoRR, vol. abs/2012.03082, 2020.
- A. Kendall and Y. Gal, “What uncertainties do we need in bayesian deep learning for computer vision?” Advances in neural information processing systems, vol. 30, 2017.
- L. Mur-Labadia, R. Martinez-Cantin, and J. Guerrero, “Bayesian deep learning for affordance segmentation in images,” in 2023 Int. Conf. on Robotics and Automation, 2023.
- D. Miller, N. Sünderhauf, M. Milford, and F. Dayoub, “Uncertainty for identifying open-set errors in visual object detection,” IEEE Robotics and Automation Letters, vol. 7, no. 1, 2021.
- K. Shridhar, F. Laumann, and M. Liwicki, “Uncertainty estimations by softplus normalization in bayesian convolutional neural networks with variational inference,” arXiv preprint arXiv:1806.05978, 2018.
- G. Lenczner, A. Chan-Hon-Tong, B. Le Saux, N. Luminari, and G. Le Besnerais, “Dial: Deep interactive and active learning for semantic segmentation in remote sensing,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 15, 2022.
- G. Georgakis, B. Bucher, K. Schmeckpeper, S. Singh, and K. Daniilidis, “Learning to map for active semantic goal navigation,” in Int. Conf. on Learning Representations, 2022.
- Y. Feldman and V. Indelman, “Bayesian viewpoint-dependent robust classification under model and localization uncertainty,” in 2018 IEEE International Conference on Robotics and Automation (ICRA), 2018.
- V. Tchuiev and V. Indelman, “Epistemic uncertainty aware semantic localization and mapping for inference and belief space planning,” Artificial Intelligence, vol. 319, 2023.
- A. Dai, A. X. Chang, M. Savva, M. Halber, T. Funkhouser, and M. Nießner, “Scannet: Richly-annotated 3d reconstructions of indoor scenes,” in Proc. Comp. Vision and Pattern Recog. (CVPR), IEEE, 2017.
- L. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking atrous convolution for semantic image segmentation,” CoRR, vol. abs/1706.05587, 2017.
- M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman, “The PASCAL Visual Object Classes Challenge 2012 (VOC2012).”
- P. K. Nathan Silberman, Derek Hoiem and R. Fergus, “Indoor segmentation and support inference from rgbd images,” in ECCV, 2012.
- David Morilla-Cabello (4 papers)
- Lorenzo Mur-Labadia (9 papers)
- Ruben Martinez-Cantin (36 papers)
- Eduardo Montijano (34 papers)