Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
153 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fully Onboard Low-Power Localization with Semantic Sensor Fusion on a Nano-UAV using Floor Plans (2310.12536v2)

Published 19 Oct 2023 in cs.RO

Abstract: Nano-sized unmanned aerial vehicles (UAVs) are well-fit for indoor applications and for close proximity to humans. To enable autonomy, the nano-UAV must be able to self-localize in its operating environment. This is a particularly-challenging task due to the limited sensing and compute resources on board. This work presents an online and onboard approach for localization in floor plans annotated with semantic information. Unlike sensor-based maps, floor plans are readily-available, and do not increase the cost and time of deployment. To overcome the difficulty of localizing in sparse maps, the proposed approach fuses geometric information from miniaturized time-of-flight sensors and semantic cues. The semantic information is extracted from images by deploying a state-of-the-art object detection model on a high-performance multi-core microcontroller onboard the drone, consuming only 2.5mJ per frame and executing in 38ms. In our evaluation, we globally localize in a real-world office environment, achieving 90% success rate. We also release an open-source implementation of our work.

Citations (2)

Summary

  • The paper presents a fusion method of annotated floor plans and onboard ToF sensor data, achieving a 90% localization success rate in indoor environments.
  • The paper employs a low-power multi-core microcontroller that processes semantic cues at 20 fps with just 2.5 mJ per frame.
  • The paper demonstrates that integrating sparse geometric data with semantic insights overcomes challenges in symmetric and cluttered indoor settings.

Fully Onboard Low-Power Localization with Semantic Sensor Fusion on a Nano-UAV using Floor Plans: An Expert Overview

The paper under discussion presents a novel approach to the challenging task of autonomously localizing nano-sized Unmanned Aerial Vehicles (UAVs) in indoor environments. Given the constraints of limited sensing and computing resources on these small platforms, the authors leverage a fusion of geometric and semantic data to enhance localization capabilities, achieving promising results in complex indoor settings.

The proposed method employs floor plans annotated with semantic information as the primary map reference. This strategy circumvents the time and resource demands associated with map creation typically required for localization processes. The technique integrates geometric data from Time-of-Flight (ToF) sensors with semantic cues by deploying a state-of-the-art object detection model onboard the UAV. This integration capitalizes on the ready availability of annotated floor plans while enhancing data richness with semantic information.

Technical Approach and System Architecture

The authors deploy a high-performance, low-power multi-core microcontroller on the UAV, capable of executing object detection at 20 frames per second with minimal energy consumption (2.5mJ per frame in 38ms). This substantial efficiency facilitates real-time processing of semantic information extracted from images, which is crucial for the mission scenarios considered.

The nano-UAV is equipped with miniaturized ToF sensors providing sparse yet crucial geometric data for the localization task. The geometric data's limited range and resolution are effectively compensated for by semantic inference, thus enhancing the robustness and accuracy of localization across large and structured environments.

Results and Evaluation

The approach was evaluated in a real-world office scenario, demonstrating a 90% success rate in global localization. The fusion of semantic and geometric information significantly improves accuracy compared to setups relying solely on one type of data. The evaluation included detecting and localizing the UAV in structurally symmetric and sparse environments, showcasing the method's effectiveness in overcoming typical challenges.

The method achieved these results with stringent computational and power constraints. Specifically, the authors emphasize the system’s capability to operate fully onboard without offloading computations, a critical feature given the power budget constraints typical of nano-UAVs.

Implications and Future Research Directions

The paper outlines several implications for both theoretical research in SLAM (Simultaneous Localization and Mapping) and practical applications in robotics. The work contributes to autonomous navigation by demonstrating that semantic information can be harness to improve localization in environments where traditional geometric approaches face limitations.

Future research could explore extending this method to incorporate dynamic environments, where semantic maps or models may be updated in real time to adapt to changes. Additionally, enhancing the recognition capabilities for a broader array of semantic classes could further bolster the localization robustness and expand the UAV's operational contexts, including more diverse and cluttered indoor scenes.

In summary, this research exemplifies an effective utilization of semantic sensor fusion for nano-UAV localization, setting a promising precedent for future work aimed at achieving autonomy in constrained robotic systems.

Github Logo Streamline Icon: https://streamlinehq.com

GitHub

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com