Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
149 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Automatic Robot Path Planning for Visual Inspection from Object Shape (2312.02603v1)

Published 5 Dec 2023 in cs.RO

Abstract: Visual inspection is a crucial yet time-consuming task across various industries. Numerous established methods employ machine learning in inspection tasks, necessitating specific training data that includes predefined inspection poses and training images essential for the training of models. The acquisition of such data and their integration into an inspection framework is challenging due to the variety in objects and scenes involved and due to additional bottlenecks caused by the manual collection of training data by humans, thereby hindering the automation of visual inspection across diverse domains. This work proposes a solution for automatic path planning using a single depth camera mounted on a robot manipulator. Point clouds obtained from the depth images are processed and filtered to extract object profiles and transformed to inspection target paths for the robot end-effector. The approach relies on the geometry of the object and generates an inspection path that follows the shape normal to the surface. Depending on the object size and shape, inspection paths can be defined as single or multi-path plans. Results are demonstrated in both simulated and real-world environments, yielding promising inspection paths for objects with varying sizes and shapes. Code and video are open-source available at: https://github.com/CuriousLad1000/Auto-Path-Planner

Definition Search Book Streamline Icon: https://streamlinehq.com
References (25)
  1. J. E. See, C. G. Drury, A. E. Speed, A. Williams, and N. Khalandi, “The role of visual inspection in the 21st century,” in Proc. of the Human Factors and Ergonomics Society Annual Meeting, vol. 61, 2017, pp. 262 – 266.
  2. S. Agnisarman, S. Lopes, K. Chalil Madathil, K. Piratla, and A. Gramopadhye, “A survey of automation-enabled human-in-the-loop systems for infrastructure visual inspection,” Automation in Construction, vol. 97, pp. 52–76, 2019.
  3. N. Kato, M. Inoue, M. Nishiyama, and Y. Iwai, “Comparing the recognition accuracy of humans and deep learning on a simple visual inspection task,” in 5th Asian Conference on Pattern Recognition, 2020, pp. 184–197.
  4. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 770–778.
  5. W. Liu et al., “SSD: Single Shot MultiBox Detector,” in European Conference on Computer Vision (ECCV), 2016, pp. 21–37.
  6. O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in 18th International Conference on Medical Image Computing and Computer-Assisted Intervention, 2015, pp. 234–241.
  7. S. F. Dodge and L. Karam, “A study and comparison of human and deep learning recognition performance under visual distortions,” in 26th International Conference on Computer Communication and Networks (ICCCN), 2017, pp. 1–7.
  8. S. R. Kheradpisheh, M. Ghodrati, M. Ganjtabesh, and T. Masquelier, “Deep networks can resemble human feed-forward vision in invariant object recognition,” Scientific Reports, vol. 6, p. 32672, 2015.
  9. Q. Yu, Y. Yang, F. Liu, Y.-Z. Song, T. Xiang, and T. M. Hospedales, “Sketch-a-net: A deep neural network that beats humans,” International Journal of Computer Vision, vol. 122, pp. 411–425, 2017.
  10. C.-C. Ho, J.-C. Li, T.-H. Kuo, and C. Peng, “Multicamera fusion-based leather defects marking system,” Advances in Mechanical Engineering, vol. 5, 2013.
  11. Z. Ren, F. Fang, N. Yan, and Y. Wu, “State of the art in defect detection based on machine vision,” International Journal of Precision Engineering and Manufacturing-Green Technology, vol. 9, pp. 661–691, 2021.
  12. M. A. H. Ali and A. K. Lun, “A cascading fuzzy logic with image processing algorithm–based defect detection for automatic visual inspection of industrial cylindrical object’s surface,” Int. Journal of Advanced Manufacturing Technology, vol. 102, pp. 81–94, 2019.
  13. M. Hanses, R. Behrens, and N. Elkmann, “Hand-guiding robots along predefined geometric paths under hard joint constraints,” in 21st International Conference on Emerging Technologies and Factory Automation (ETFA), 2016, pp. 1–5.
  14. Z. Loncarevic, A. Gams, S. Rebersek, B. Nemec, J. Skrabar, J. Skvarc, and A. Ude, “Specifying and optimizing robotic motion for visual quality inspection,” Robotics and Computer Integrated Manufacturing, vol. 72, p. 102200, 2021.
  15. M. Roberts, S. Shah, D. Dey, A. T. Truong, S. N. Sinha, A. Kapoor, P. Hanrahan, and N. Joshi, “Submodular trajectory optimization for aerial 3d scanning,” in IEEE International Conference on Computer Vision (ICCV), 2017, pp. 5334–5343.
  16. R. Monica and J. Aleotti, “Surfel-based next best view planning,” IEEE Robotics and Automation Letters, vol. 3, pp. 3324–3331, 2018.
  17. M. Naazare, F. G. Rosas, and D. Schulz, “Online next-best-view planner for 3d-exploration and inspection with a mobile manipulator robot,” IEEE Robotics and Automation Letters, vol. 7, pp. 3779–3786, 2022.
  18. X. Fan, L. Zhang, B. J. Brown, and S. Rusinkiewicz, “Automated view and path planning for scalable multi-object 3d scanning,” ACM Transactions on Graphics TOG, vol. 35, pp. 1 – 13, 2016.
  19. R. Arav, S. Filin, and N. Pfeifer, “Content-aware point cloud simplification of natural scenes,” IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–12, 2022.
  20. X. Yu, Z. Cheng, Y. Zhang, and L. Ou, “Point cloud modeling and slicing algorithm for trajectory planning of spray painting robot,” Robotica, vol. 39, pp. 2246 – 2267, 2021.
  21. G. Wang, J. Cheng, R. Li, and K. Chen, “A new point cloud slicing based path planning algorithm for robotic spray painting,” in IEEE International Conference on Robotics and Biomimetics (ROBIO), 2015, pp. 1717–1722.
  22. M. Ester, H.-P. Kriegel, J. Sander, and X. Xu, “A density-based algorithm for discovering clusters in large spatial databases with noise,” in Knowledge Discovery and Data Mining, 1996.
  23. Q.-Y. Zhou, J. Park, and V. Koltun, “Open3D: A modern library for 3D data processing,” ArXiv, vol. abs/1801.09847, 2018.
  24. M. Pauly, “Point primitives for interactive modeling and processing of 3d-geometry,” Ph.D. dissertation, ETH Zurich, 2003.
  25. I. A. Şucan, M. Moll, and L. E. Kavraki, “The Open Motion Planning Library,” IEEE Robotics & Automation Magazine, vol. 19, no. 4, pp. 72–82, 2012.

Summary

  • The paper presents a novel methodology that automatically generates inspection paths by processing 3D point cloud data of object surfaces.
  • The system integrates a depth camera with a robot manipulator to extract and filter geometric profiles from objects of varying shapes and sizes.
  • Tests in both simulated and real-world environments demonstrate efficient path planning, reducing reliance on manual trajectory design.

Introduction to Robot Path Planning

Robot path planning plays an integral role in the field of automation and assists in enhancing the efficiency and safety of tasks such as visual inspection. Traditional methods of visual inspection, predominantly performed by humans, are often time-consuming and potentially hazardous, requiring the presence of a person in unsafe environments. Advances in AI and robotics have paved the way for more autonomous solutions in various industries, notably influencing the procedures associated with quality control and maintenance tasks.

Automating Visual Inspection

The paper presents a novel methodology for automating robot path planning for visual inspection purposes. The system integrates a single depth camera mounted onto a robot manipulator to capture 3D data in the form of point clouds, which represent the surface details of objects. These point clouds are then processed and filtered to extract the object's profile, regardless of its shape or size.

This innovative approach leverages the geometry of the objects to generate inspection paths that follow their shape. The methodology allows both single and multiple inspection path plans to adapt to objects with different dimensions and geometric complexities. The paper's approach takes steps to address a crucial issue in traditional visual inspection methods—dependence on training data, which often requires labor-intensive gathering and is hindered by variability in objects' sizes and shapes.

Methodology and Results

The proposed system operates by capturing multiple point clouds, refining the point cloud data through filtering and clustering, and deriving normals to determine the robot camera's optimal positioning. The collected point cloud data is processed to identify the cluster most representative of the object profile. From this profile, target poses are generated, guiding the robot manipulator along the calculated inspection path.

The methodology was tested with both simulated and real-world objects, demonstrating that it could successfully generate appropriate path plans for objects of various shapes and sizes. In experiments involving complex objects like car models and tunnels in virtual environments as well as physical items like benches and suction rolls, the system proved to be effective in path planning without the need for manual data collection or predefined trajectories.

Conclusion and Discussion

In drawing conclusions, the research indicates the system's potential to automate visual inspection and 3D modeling without depending on CAD models, existing data sets, or human-prescribed paths. Although subject to certain limitations, such as the occasional appearance of "ghost" targets and residual points in the filtered point cloud, the system provides a malleable solution adaptable to different inspection requirements.

Furthermore, the paper discusses the implications of the number of point cloud samples and filtering thresholds on the quality and density of the target paths, offering insight into optimizing the system for various operational conditions. The authors also highlight their confidence in the open-source availability of their code and invite further advancement by the research community.

In conclusion, this approach to robot path planning embraces the fusion of 3D point cloud data and robotics to streamline the visual inspection process, which aims to shift repetitive and dangerous tasks from humans to robots, leading to safer and more efficient industrial operations.