Teleoperation of the Industrial Robot: Augmented reality
application
Chung Xue Er (Shamaine)
Department of Computer &
Software Engineering
The Technological University of
the Shannon: Midlands Midwest
Athlone, Ireland
xechung@research.ait.ie
John Henry
Robotics and Drives
Mullingar Business Park,
Mullingar, Ireland
jhenry@rdservices.ie
Yuansong Qiao
Department of Computer &
Software Engineering
The Technological University of
the Shannon: Midlands Midwest
Athlone, Ireland
ysqiao@research.ait.ie
Ken McNevin
Robotics and Drives
Mullingar Business Park,
Mullingar, Ireland
ken@rdservices.ie
Vladimir Kuts†
Department of Computer &
Software Engineering
The Technological University of
the Shannon: Midlands Midwest
Athlone, Ireland
vkuts@ait.ie
Niall Murray
Department of Computer &
Software Engineering
The Technological University of
the Shannon: Midlands Midwest
Athlone, Ireland
nmurray@research.ait.ie
Augmented reality application. In Proceedings of ACM MMSys conference
(MMSys2022). ACM, Athlone, Ireland, 5 pages.
ABSTRACT
Industry 4.0 is aimed at the full manufacturing domain
automatization and digitalization. Humans and robots working
together are being discussed widely both in the academic and
industrial sectors. As being discussed, there is a need for a more
novel type of interaction method between humans and robots.
This demonstrational paper is presenting the technical
advancement and prototype of remote control and reprogramming of the Industrial robot. This development is safe and
efficient and allows the operator to focus rather on the final task
than the programming process.
1
Introduction
The fourth industrial revolution often referred to as Industry 4.0, is
a general term for a new industrial model [1]. Industry 4.0 aims at
building cyber-physical production systems (CPPS) (which
comprise intelligent, real-time-capable, networked sensors and
actuators) that unite both digital and physical worlds to make
manufacturing increasingly smart by utilising the internet of things
(IoT)[1][2]. Industry 4.0 [3] includes a range of digital technologies
that includes but is not limited to Augmented Reality (AR), Virtual
Reality (VR), Digital Twins (DT), predictive maintenance, cloud
computing, Internet of Things (IoT), Artificial Intelligence (AI),
and big data. This stack is essential in achieving smart factories of
the future (FoF).
However, little research has been conducted to develop a usercentred human-robot interaction (HRI) in the context of industry
4.0 environment requirements. HRI is the study of how humans
communicate with robotic systems. It informs us on how to best
design, evaluate, understand and implement robotic systems that
are competent enough for carrying out collaborative tasks with the
human stakeholder [4][5]. One of the major challenges encountered
CCS CONCEPTS
• Human-centered computing • Human-computer interaction
(HCI) • Interactive systems and tools • User interface management
systems
KEYWORDS
Augmented Reality, Human-Robot Interaction, Industrial Robot
ACM Reference format:
Chung Xue Er Shamaine, Yuansong Qiao, Vladimir Kuts, John Henry, Ken
McNevin and Niall Murray. 2022. Teleoperation of the Industrial Robot:
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are not
made or distributed for profit or commercial advantage and that copies bear
this notice and the full citation on the first page. Copyrights for components
of this work owned by others than ACM must be honored. Abstracting with
credit is permitted. To copy otherwise, or republish, to post on servers or to
redistribute to lists, requires prior specific permission and/or a fee. Request
permissions from permissions@acm.org.
MMSys '22, June 14–17, 2022, Athlone, Ireland
© 2022 Association for Computing Machinery.
ACM ISBN 978-1-4503-9283-9/22/06…$15.00
https://doi.org/10.1145/3524273.3532901
299
ACM MMSys, June, 2022, Athlone, Ireland
Chung Xue Er (Shamaine) et al.
by HRI is that both humans and robots use extremely distinct
methods of communication and data representation [6]. Current
work in HRI seeks to discover a less training-intensive, more direct,
safe, effective, natural and stress-free interaction [7].
Teleoperation or often defined as telerobotic [8] when controlling
robotics systems, is commonly used in human-robot interaction
(HRI) research. It is an approach that allows human operators to
control machines or robotic systems from a distance. This approach
has been intensively developed in the last decade. Teleoperated
robots are mostly used in telesurgery [9], military search and rescue
tasks [10], mining [11] as well as space exploration [12]. According
to [13], teleoperated robots helps to reduce the risk and costs
associated with human exposure to life-threatening conditions. The
capabilities spectrum for telerobotic systems has been expanded
from traditional 2D computer interfaces to immersive media-based
[14] VR and AR controlled teleoperated robots and has attracted
considerable attention in recent years [15][16]. AR provides more
spatial awareness for the users if compared with the more
immersive VR experience which may cause them to lose
connection with the real-world surroundings.
The novel approach of using Augmented Reality (AR) creates a
new design space and an opportunity for facilitating more
naturalistic human-robotic interaction for applications such as
teleoperation in real-time [17]. The key inspiration is that AR has
the potential of displaying visual objects directly (from the field of
view of the robotic system) into the field of view of the user without
being disconnected from the physical world. A clear
communication orientation and information visualisation are
needed to mediate robot teleoperation [18][19]. On account of the
ability of AR to augment the operator view and visualise the robotic
corresponding task space, this allows the operator to retrieve the
robot’s sensor state information (temperature level, position, joint
angle etc) via graphic overlays result in AR. This provides great
potential as an interface for robotic teleoperation [20].
To the best of this demo paper authors knowledge, no studies have
been performed to thoroughly analyse AR interfaces’ effectiveness
and accuracy in more complex path following tasks.
2
Figure 1: Developed system architecture
On the right is our physical ABB robot arm connected to the ROS
server over an ethernet cable. The integration of ROS and Unity 3D
is enabled via ROS-Sharp, through a WebSocket connection on
port 9090. All development tools used in this project are
summarized in Table 1. On the HoloLens 2 client-side, the Mixed
Reality Toolkit (MRTK) library [21] was imported to control the
user hand gestures on a HoloLens 2. The Final IK asset attached to
the holographic robot arm consisted of a powerful Inverse
Kinematics solution (FABRIK algorithm is chosen) for advanced
robot animation.
Vuforia Image Target is used to correctly overlay the holographic
robot arm on top of the real robot arm. On the server-side, two
virtual switches were created and connected to the 31 Hyper V
machine at the same time. A file server will be the first to launch
opening port 9090, waiting for the HoloLens Client to connect over
the mobile Wi-Fi hotspot. Secondly, both MoveIT and Rviz are
launch together. MoveIT is mainly for Motion Planning. Basically,
it creates a sequence of movements that all joints have to perform
in order for the end effector to reach the desired position. While
Rviz is for visualization purposes. On the ABB client-side, a WAN
port is chosen to connect the ABB controller over the ROS server.
Table 1: Tools used for the solution development
System Architecture and Technology Stack
The proposed HRI system architecture is presented in Fig.1. On the
left is a Microsoft HoloLens 2 user interface with Unity
2019.4.21f1 (64-bit) game engine. The Microsoft HoloLens 2 AR
HMD is employed because of its built-in localization and mapping
(SLAM) abilities in addition to its self-supporting nature. It also
currently represents state of the art in AR HMD technology. In the
Centre, is a ROS system running on a virtual Ubuntu16.04 LTS
ROS Kinetic virtual machine.
The integration of ROS and Unity 3D is enabled by implementing
the open-source ROS-Sharp library [22]. A WebSocket connection
is established on port 9090 between Unity and the ROS arm
simulator. It allows ROS-Sharp to serialize a topic. A pose stamped
publisher script in Unity will create a topic. The topic is activated
300
Teleoperation of the Industrial Robot: Augmented reality
application
ACM MMSys, June, 2022, Athlone, Ireland
each time when Unity tries to send end-effector pose information
as JSON [23] to ROS. Within the script, the Unity end effector
Euler rotation x, y and z will automatically be converted into
quaternion x, y, z and w before sending. In this case, Unity will act
as a publisher, while the ROS simulator is the subscriber. Unity
converts C# data structures into JSON messages, and the pose data
is passed across the bridge. The rosbridge library which contains
the core rosbridge package then retrieves the pose data through the
same topic by converting the JSON message back into geometry
messages.
2.1
ROS is a versatile, multilingual middleware. It provides a
collection of tools and libraries for robotic developers to simplify
their effort of designing complicated and robust robot actions [26].
ROS is open-source and is language agnostic. It has become the
common automation language for robotics and research focus in
recent years. It is widely used in various robotic applications.
Gazebo, as shown in Fig. 3(a). serves as a robotic simulator while
ROS is the interface to the robot. The key benefit of using Gazebo
is its ready integration with ROS. Similar to ROS, Gazebo is also
open source. It is one of the most popular robotic simulators [68].
RViz, as shown in Fig. 3(b), is not a simulator but a 3D
visualisation tool for ROS applications. It offers real-time sensor
information from robot sensors and a view of the robot model. It
can be used to display 2D and 3D point clouds as well as camera or
lasers data [27].
Client-side stack
The Client-Side stack mainly responsible for handling the user’s
request and interaction. This section covers the client-side
technology such as the Unity game engine, the Mixed Reality
Toolkit (MRTK) and finally, the Vuforia SDK used to build the
application on a HoloLens 2.
The Unity3D game engine is deployed to create the AR application.
A suite of development tools is provided to create different
experiences, and in this case, the Mixed Reality Toolkit (MRTK)
[24] was used. MRTK is a Microsoft- motivated open-source
development kit that offers a set of components and features used
to accelerate application
development for the Microsoft HoloLens 2 [25]. The MRTK ineditor hand input simulation allows the developer to test a scene in
the Unity game view. The MRTK supports rapid AR application
development. MRTK automatically sets up a default scene for a
new project when the “Add to Scene and Configure” is selected.
The HoloLens 2 implemented the style buttons provided by the
MRTK example package to create a set of tag-along buttons as
shown on figure 2. A Radial View script is attached to the parent
button. This solver script allows the buttons to follow the user's
head movement.
Figure 3. (a)ABB robot arm in Gazebo simulator (b) ABB robot
arm in RViz with transform frames.
3
Experiment
This chapter is a description of the measures taken to evaluate the
effectiveness of AR application, those will be applied in the future
steps within the user test experiments.
The main design of the experiment as illustrated in Fig. 4 consists
of a computer running Ubuntu16, HMD Microsoft HoloLens 2, and
an ABB robot arm version IRB 1200-5 /0.9. We will place an image
target marker on a table 15cm in front of the physical robot arm.
The data collection flowchart of the RoSTAR system is shown in
Fig. 4-2. The evaluation of the proposed RoSTAR system was
achieved by running two experiments. It is the same person that
performs the testing sequences throughout the first and second
experiments.
A series of objective measures for the ABB robot arm was
conducted to test the robot path accuracy, end-effector’s position
accuracy and path completion time.
1. Path Accuracy: Path accuracy was calculated as the
number of correct trajectories executed by the robot
divided by the total number of trials (10 trials). We
treated the robot arm movement as timeout when the
robot arm does not move as predicted/ spinning 180
degrees along the pre-specified trajectories
Figure2: (a) Virtual Tag-Along Buttons Created with MRTK
Examples package (b) Draggable 3D waypoint with MRTK
articulated hand tracking input.
2.1 Server-side stack
The Server-Side stack is mainly responsible for receiving the user’s
request and feedback from the Client-Side. This section covers the
server-side technology such as the Robot Operating System (ROS),
MoveIT, ROSBridge and finally, the Robot Model used to
teleoperate the physical ABB robot arm.
301
ACM MMSys, June, 2022, Athlone, Ireland
2.
3.
Chung Xue Er (Shamaine) et al.
5
(Singularities). The path accuracy results will be plotted
in the form of graphs.
Absolute
Position
Accuracy:
The
position
measurements across four different complexities are
stored only after the end-effector arrived at each pose
(irrespective of the path taken by the robot to reach the
goal position). The robot arm will execute the same
trajectories 10 times after receiving the pose data.
Average Path Completion Time: Each trial began when
the robot moves from start to goal position. The time
measure was calculated as the average time of the
interaction across all 10 trials.
Conclusion
In this technical demonstration overview, we present a novel opensource HRI system based on the Robot Operating System (ROS)
and Augmented Reality (AR). It is demonstrated that the HoloLens
2 is capable of visualizing the robot movement and safety zone
within a cutting-edge AR device. We propose a prototype
application that successfully displayed key parameters of userdefined robot trajectory within the Microsoft HoloLens 2. As a
result, it is possible to control and re-program robot in real-time
from remote distances. The safety aspect is also considered, as the
safety gate near the robot is always prioritized.
REFERENCES
[1] A. C. Pereira and F. Romero, ‘A review of the meanings and the implications of
the Industry 4.0 concept’, Procedia Manuf., vol. 13, pp. 1206–1214, Jan. 2017,
doi: 10.1016/j.promfg.2017.09.032.
[2] D. Kerpen, M. Löhrer, M. Saggiomo, M. Kemper, J. Lemm, and Y.-S. Gloy,
‘Effects of cyber-physical production systems on human factors in a weaving
mill: Implementation of digital working environments based on augmented
reality’, in 2016 IEEE International Conference on Industrial Technology (ICIT),
Mar. 2016, pp. 2094–2098. doi: 10.1109/ICIT.2016.7475092.
[3] J. Egger and T. Masood, ‘Augmented reality in support of intelligent manufacturing
– A systematic literature review’, Comput. Ind. Eng., vol. 140, p. 106195, Feb.
2020, doi: 10.1016/j.cie.2019.106195.
[4] D. Feil-Seifer and M. Mataric, ‘Human-Robot Interaction’, in Encyclopedia of
Complexity and Systems Science, 2009, pp. 4643–4659. doi: 10.1007/978-0387-30440-3_274.[5] M. Goodrich and A. Schultz, ‘Human-Robot Interaction:
A Survey’, Found. Trends Hum.-Comput. Interact., vol. 1, pp. 203–275, Jan.
2007, doi: 10.1561/1100000005.
[6] T. Williams, D. Szafir, T. Chakraborti, and E. Phillips, ‘Virtual, Augmented, and
Mixed Reality for Human-Robot Interaction (VAM-HRI)’, in 2019 14th
ACM/IEEE International Conference on Human-Robot Interaction (HRI), Mar.
2019, pp. 671–672. doi: 10.1109/HRI.2019.8673207.
[7] L. H. Manring et al., ‘Augmented Reality for Interactive Robot Control’, 2020. doi:
10.1007/978-3-030-12243-0_2.[8] G. Niemeyer, C. Preusche, and G. Hirzinger,
‘Telerobotics’, in Automatica, vol. 25, 2008, pp. 741–757. doi: 10.1007/978-3540-30301-5_32.
[9] T. Haidegger, J. Sándor, and Z. Benyó, ‘Surgery in space: the future of robotic
telesurgery’, Surg. Endosc., vol. 25, no. 3, pp. 681–690, Mar. 2011, doi:
10.1007/s00464-010-1243-3.
[10] T. Kot and P. Novák, ‘Application of virtual reality in teleoperation of the military
mobile robotic system TAROS’, 2018, doi: 10.1177/1729881417751545.
[11] D. W. Hainsworth, ‘Teleoperation User Interfaces for Mining Robotics’, Auton.
Robots, vol. 11, pp. 19–28, Jul. 2001, doi: 10.1023/A:1011299910904.
[12] G. Liu, X. Geng, L. Liu, and Y. Wang, ‘Haptic based teleoperation with masterslave motion mapping and haptic rendering for space exploration’, Chin. J.
Aeronaut., vol. 32, no. 3, pp. 723–736, Mar. 2019, doi:
10.1016/j.cja.2018.07.009.
[13] N. Small, K. Lee, and G. Mann, ‘An assigned responsibility system for robotic
teleoperation control’, Int. J. Intell. Robot. Appl., vol. 2, no. 1, pp. 81–97, Mar.
2018, doi: 10.1007/s41315-018-0043-0.58
[14] A. Perkis et al., QUALINET White Paper on Definitions of Immersive Media
Experience (IMEx). 2020.
[15] D. Krupke, F. Steinicke, P. Lubos, Y. Jonetzko, M. Görner, and J. Zhang,
Comparison of Multimodal Heading and Pointing Gestures for Co-Located
Mixed
Reality
Human-Robot
Interaction.
2018.
doi:
10.1109/IROS.2018.8594043.
[16] L. Peppoloni, F. Brizzi, C. A. Avizzano, and E. Ruffaldi, ‘Immersive ROSintegrated framework for robot teleoperation’, in 2015 IEEE Symposium on 3D
User
Interfaces
(3DUI),
Mar.
2015,
pp.
177–178.
doi:
10.1109/3DUI.2015.7131758.
[17] E. Dima et al., ‘View Position Impact on QoE in an Immersive Telepresence
System for Remote Operation’, in 2019 Eleventh International Conference on
Quality of Multimedia Experience (QoMEX), Jun. 2019, pp. 1–3. doi:
10.1109/QoMEX.2019.8743147.
[18] S. Kortekamp, S. Werning, I. Ingmar, and O. Thomas, THE FUTURE OF
DIGITAL WORK - USE CASES FOR AUGMENTED REALITY GLASSES.
2019.
[19] A. Monferrer and D. Bonyuet, Cooperative robot teleoperation through virtual
reality interfaces. 2002, p. 248. doi: 10.1109/IV.2002.1028783.
Figure 4: Experiment setup
The performance of the HoloLens 2 in the task was evaluated by
monitoring the frame rate at four different stages mainly: Active
Tracking Stage, Fixed Position Stage, Trajectory Execution Stage,
End Trajectory Stage. The objective measures for the HoloLens 2
are frame rate and path completion time.
1. Frame Rate: The frame rate is calculated by averaging the
total frame rates received every second at each stage. The
HoloLens 2 frame rate are collected every second at the
start up and classified into four main stages: The tracking
stage, planning stage, execution stage and reset the stage.
2. Average Path Completion Time: Each trial began when the
virtual robot arm moves from start to goal position. The
time measure was calculated as the average time of the
interaction across all 10 trials. The average path
completion time across 4 levels of trajectory complexity
for both HoloLens 2 virtual robot arm and real ABB robot
arm based on the number of waypoints are compared via a
3D-clustered column chart. A shorter path completion time
indicates fast inverse kinematics solve times.
4
Demonstration
For the demonstration in the ACM MMSys2022, the visitors will
be able to program and control in real-time the ABB industrial robot
remotely and follow in the PC monitor how the physical robot is
being moved in the university (TUS: Athlone campus). The aim of
the demonstration is to show a safe and efficient method for
industrial hardware teleoperation from remote distances.
302
Teleoperation of the Industrial Robot: Augmented reality
application
ACM MMSys, June, 2022, Athlone, Ireland
[20] S. Green, M. Billinghurst, X. Chen, and J. Chase, ‘Human Robot Collaboration:
An Augmented Reality Approach A Literature Review And Analysis’, Jan. 2007,
doi: 10.1115/DETC2007-34227
[21] polar-kev, ‘MRTK-Unity Developer Documentation - Mixed Reality Toolkit’.
https://docs.microsoft.com/en-us/windows/mixed-reality/mrtk-unity/ (accessed
Sep. 22, 2021).
[22] ‘Home · siemens/ros-sharp Wiki’, GitHub. https://github.com/siemens/ros-sharp
(accessed Sep. 22, 2021)
[23] ‘JSON’. https://www.json.org/json-en.html (accessed Sep. 12, 2020)
[24] polar-kev, ‘MRTK-Unity Developer Documentation - Mixed Reality Toolkit’.
https://docs.microsoft.com/en-us/windows/mixed-reality/mrtk-unity/ (accessed
Sep. 22, 2021).
[25] ‘HoloLens 2—Pricing and Options | Microsoft HoloLens’.
https://www.microsoft.com/en-us/hololens/buy (accessed Sep. 02, 2019).
[26] ‘ROS.org | About ROS’. https://www.ros.org/about-ros/ (accessed May. 21,
2020).
[27] ‘rviz - AWS RoboMaker’.
https://docs.aws.amazon.com/robomaker/latest/dg/simulation-tools-rviz.html
(accessed Aug. 22, 2019).
303