Papers
Topics
Authors
Recent
Search
2000 character limit reached

Assisting MoCap-Based Teleoperation of Robot Arm using Augmented Reality Visualisations

Published 9 Jan 2025 in cs.RO and cs.HC | (2501.05153v1)

Abstract: Teleoperating a robot arm involves the human operator positioning the robot's end-effector or programming each joint. Whereas humans can control their own arms easily by integrating visual and proprioceptive feedback, it is challenging to control an external robot arm in the same way, due to its inconsistent orientation and appearance. We explore teleoperating a robot arm through motion-capture (MoCap) of the human operator's arm with the assistance of augmented reality (AR) visualisations. We investigate how AR helps teleoperation by visualising a virtual reference of the human arm alongside the robot arm to help users understand the movement mapping. We found that the AR overlay of a humanoid arm on the robot in the same orientation helped users learn the control. We discuss findings and future work on MoCap-based robot teleoperation.

Summary

  • The paper introduces an AR system overlaying a virtual arm on MoCap-based teleoperation to bridge human-robot control mapping gaps.
  • It evaluates various AR configurations, with the HumanVertical setup yielding preferred alignment and enhanced system understanding.
  • Significant reductions in physical demand and effort were observed during posture-control tasks when using AR feedback.

Assisting MoCap-Based Teleoperation of Robot Arm using Augmented Reality Visualisations

Introduction

The paper "Assisting MoCap-Based Teleoperation of Robot Arm using Augmented Reality Visualisations" (2501.05153) explores the integration of augmented reality (AR) into the teleoperation of robot arms, specifically addressing the challenges posed by the inconsistencies between human and robotic arm orientations and appearances. By utilizing AR to render virtual arms as visual references, the authors aim to mediate these inconsistencies, enhancing user performance and experience during teleoperation tasks.

MoCap-Based AR Teleoperation System

The study employs a sophisticated teleoperation system that integrates a 7-DOF robotic arm, the Franka Research 3 (FR3), with a motion capture system and AR. The system architecture involves a Linux machine running ROS 2 and the Franka Control Interface, a Windows machine running Unity and Motive, a Franka Research 3 robot, and a HoloLens 2 device for AR rendering (Figure 1). Figure 1

Figure 1: System architecture: a Linux machine running ROS 2 and the Franka Control Interface, a Windows machine running Unity and Motive, a Franka Research 3 robot, and a HoloLens 2 that renders through a remoting player.

The AR system renders a virtual arm corresponding to the physical robot arm's movements, providing visual feedback on the mapping between human and robot arm joint rotations. This setup is designed to facilitate intuitive teleoperation, allowing users to control the robot arm using their naturalistic arm movements, thereby enhancing the ease of use and understanding of the control mappings.

Study 1: Configuring the AR Visualisation

In the first study, the authors investigate various configurations of AR visualisation to determine the optimal setup for assisting with MoCap-based teleoperation. The study compares four visualisation conditions: HumanHorizontal (HH), HumanVertical (HV), RobotHorizontal (RH), and a Baseline with no AR visualisation (Figure 2). Figure 2

Figure 2: AR Visualisation conditions and apparatus: (a) Participant situated away from the robot; (b) HumanHorizontal arm in AR next to the physical robot; (c) HumanVertical; (d) RobotHorizontal.

The study employs a target-reaching task to evaluate user performance. While no statistically significant differences were found in movement time or NASA-TLX scores among the visualisation conditions, qualitative feedback indicated a preference for the HumanVertical configuration. Participants reported that this setup provided a consistent visual reference with the physical robot, aiding in their understanding of control mappings.

Study 2: Posture Control

The second study further explores the efficacy of the AR visualisation in a posture-matching task, comparing two conditions: AR Arm and No Arm. Participants were required to match predefined postures using the robot arm, with visual feedback provided through spheres rendered on the robot's key joints (Figures 4 and 5). Figure 3

Figure 3: S2 postures: a) Elbow up, wrist up; b) Elbow up, wrist down; c) Elbow down, wrist up; d) Elbow down, wrist down.

Figure 4

Figure 4: Study 2: participant, condition No Arm and AR Arm. Blue and red sphere rendered on elbow and wrist respectively. Lighter blue and red denote targets to match posture.

The results indicated significant reductions in perceived physical demand, effort, and frustration when using the AR Arm configuration compared to the No Arm baseline. Qualitative feedback suggested that the AR visualisation served as an effective learning tool for understanding teleoperation control mappings, particularly during the initial stages of learning.

Conclusion

The research demonstrates the potential of AR visualisations to enhance MoCap-based teleoperation of robot arms by bridging the perceptual gaps between human operators and robotic systems. While the study identifies the HumanVertical configuration as optimal for visual alignment, it also highlights the effectiveness of AR as a transitional learning aid rather than a perpetual guidance tool. This work lays a foundation for future research into the incorporation of AR in robotic teleoperation, particularly in contexts requiring intuitive and seamless human-robot interaction.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.