Towards Autonomous Driving: a Multi-Modal 360$^{\circ}$ Perception Proposal (2008.09672v1)
Abstract: In this paper, a multi-modal 360${\circ}$ framework for 3D object detection and tracking for autonomous vehicles is presented. The process is divided into four main stages. First, images are fed into a CNN network to obtain instance segmentation of the surrounding road participants. Second, LiDAR-to-image association is performed for the estimated mask proposals. Then, the isolated points of every object are processed by a PointNet ensemble to compute their corresponding 3D bounding boxes and poses. Lastly, a tracking stage based on Unscented Kalman Filter is used to track the agents along time. The solution, based on a novel sensor fusion configuration, provides accurate and reliable road environment detection. A wide variety of tests of the system, deployed in an autonomous vehicle, have successfully assessed the suitability of the proposed perception stack in a real autonomous driving application.
- Jorge Beltrán (8 papers)
- Carlos Guindel (6 papers)
- Irene Cortés (3 papers)
- Alejandro Barrera (4 papers)
- Armando Astudillo (1 paper)
- Jesús Urdiales (1 paper)
- Farid Bekka (1 paper)
- Vicente Milanés (2 papers)
- Fernando García (15 papers)
- Mario Álvarez (1 paper)