Skip to main content

EDITORIAL article

Front. Robot. AI, 27 April 2022
Sec. Robotic Control Systems
This article is part of the Research Topic Shared Control for Tele-Operation Systems View all 5 articles

Editorial: Shared Control for Tele-Operation Systems

  • 1Department of Engineering and Design, University of Sussex, Brighton, United Kingdom
  • 2Communication Science Laboratories, Nippon Telegraph and Telephone, Tokyo, Japan
  • 3Institute for Infocomm Research, Agency for Science, Technology and Research, Singapore, Singapore

Editorial on the Research Topic
Shared Control for Tele-Operation Systems

Tele-operation systems have been extensively studied in the literature and used in various robotic applications, such as in surgery, search and rescue, space exploration, nuclear decommissioning, etc. However, most traditional tele-operation systems adopt a leader-follower paradigm, where the tele-operated robot follows the human operator’s guidance. This unilateral interaction does not leverage the robot’s autonomy and imposes cognitive and physical loads on the human operator. Conversely, if the robot’s strengths such as local sensing and accurate, fast execution capabilities are fully utilized, the human’s control effort can be reduced while simultaneously increasing task efficiency and performance. Therefore, shared control between a human and a robot is essential to develop an advanced teleoperation system. However, designing shared control for teleoperation systems raises many challenges. From the robot’s perspective, decision marking must be in place to address questions such as when the robot should follow the human, when it should not, and what to do if there are conflicts between the two. The designer must also be aware of how the human adapts to the robot’s behavior, and how the latter’s design should incorporate such human motor adaptation.

In this Research Topic, we have invited researchers and practitioners in relevant fields to discuss these challenging issues. Four articles have been collected, which provide interesting insights on this topic and present promising results.

Zolotas et al. introduce a virtual reality system to assist human users in manipulation tasks. Their system provides human users visualization of joint limit and environmental constraints of the robot follower, by displaying manipulability polytopes at the teleoperated robot’s end-effector. They first use a pilot study to find graphical cues and virtual reality setup that are suitable for the task of screwing in a set of bolts. Then, through comparative experiments, they conclude that their system increases safety in terms of preventing erratic motion, despite reducing the task completion speed, compared to teleoperation without shared control.

Costi et al. investigate how shared control can mitigate the negative effects of time delay that is inevitable in teleoperation. They propose and compare four different control modalities of increasing autonomy: non-predictive human control (HC), predictive human control (PHC), (shared) predictive human-robot control (PHRC), and predictive robot control (PRC). They consider an object reaching and recognition task, and develop an internal model to predict the sensor’s output that is used to increase the robot’s autonomy. Their experimental results show that the two control architectures with increasing autonomy, PHRC and PRC, outperform the other two in terms of faster task completion and increased performance. By further comparing PHRC and PRC, they show that PHRC can avoid undesired hard contact with the environment that is observed under PRC, thus suggesting the advantage of shared control and concluding that PHRC represents a good trade-off between reaching accuracy, task completion speed and contact safety.

Hagmann et al. confirm the usefulness of virtual fixtures in teleoperation in minimally invasive robotic surgery, while they propose a shared control parametrization engine to retrieve procedural context information from a digital twin. Their idea is to abstract surgical training from real surgical operations, which offers a relatively well-defined environment and is thus easier to model. They first use a digital twin to estimate geometric and semantic task states, and then based on the estimation they parameterize assistance functions that are used to provide assistance in surgical training. Results from a pilot study demonstrate that novel users profit from haptic augmentation during training of certain tasks, such that they perform a task faster, more accurately and with less cognitive load.

Hu et al. take an overview of shared control for teleoperation, which they refer to as human–machine telecollaboration paradigm in their opinion article. They separate the telecollaboration paradigm into three sub-frameworks: human–machine bidirectional augmentation framework, where human and machine contribute to a same task with flexible roles; augmented machine intelligence under human guidance, where the machine infers the human’s intent to enhance its intelligence; augmented human performance leveraging machine intelligence, where the machine provides assistance to the human. The authors particularly review the deployment of large-scale autonomous robots during the Covid-19 pandemic and show the feasibility and great potential of telecollaboration.

Author Contributions

YL drafted the manuscript. All authors read and edited the manuscript, and agreed with its content.

Funding

This work was supported in part by the Royal Society Grant IES/R3/193136 and United Kingdom EPSRC Grant EP/T006951/1.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Keywords: shared control, tele-operation, physical human-robot interaction, human-robot collaboration, robot control

Citation: Li Y, Takagi A and Tee KP (2022) Editorial: Shared Control for Tele-Operation Systems. Front. Robot. AI 9:915187. doi: 10.3389/frobt.2022.915187

Received: 07 April 2022; Accepted: 12 April 2022;
Published: 27 April 2022.

Edited and reviewed by:

Kostas J. Kyriakopoulos, National Technical University of Athens, Greece

Copyright © 2022 Li, Takagi and Tee. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yanan Li, yl557@sussex.ac.uk

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.