1 s2.0 S0007850624000131 Main
1 s2.0 S0007850624000131 Main
1 s2.0 S0007850624000131 Main
A R T I C L E I N F O A B S T R A C T
Article history: Flexible and rapidly adaptable automated production processes, e.g. with collaborative lightweight robots, are key
Available online xxx aspects for the competitiveness of companies in globalised markets. However, software adaptation of the robot
still requires specific programming skills. We have developed a human-centred programming by demonstration
Keywords: approach based on augmented reality to enable operators to intuitively adapt the robot programme. The devel-
Augmented reality
oped hand-interaction model overcomes the challenge of object tracking during the assembly demonstration
Human robot collaboration
phase. This allows quick programme modifications by the operator using the head-mounted augmented reality
Assembly
device. The human-in-the-loop concept ensures a highly reliable and robust programming process.
© 2024 The Authors. Published by Elsevier Ltd on behalf of CIRP. This is an open access article under the CC BY
license (http://creativecommons.org/licenses/by/4.0/)
https://doi.org/10.1016/j.cirp.2024.04.001
0007-8506/© 2024 The Authors. Published by Elsevier Ltd on behalf of CIRP. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/)
Please cite this article as: S. Blankemeyer et al., A hand-interaction model for augmented reality enhanced human-robot collaboration, CIRP
Annals - Manufacturing Technology (2024), https://doi.org/10.1016/j.cirp.2024.04.001
JID: CIRP
ARTICLE IN PRESS [m191;May 23, 2024;17:50]
Our framework for AR enhanced HRC consists of three different requires object specific training. To use this method, a training data-
layers (see Fig. 1). The shop floor consists of lightweight robots set must be created and trained prior to pose estimation. On the other
designed to interact with humans. These are already commercially hand, the MegaPose [12] method is able to recognise novel, not pre-
available in large numbers from various robot manufacturers. Periph- known, objects. This means that only CAD files of the objects need to
erals, such as edge devices or databases, are connected in a different be provided during pose estimation. The trade-off is that the initial
layer. The operator with the necessary interfaces to the technical sys- processing time of approximately 27 s is comparatively long. In order
tems is located in the HRC-HUB. Through the use of universal APIs to overcome this disadvantage, we use a fixed sensor, which makes it
and middleware-based approaches in between the layers, our system possible to determine the poses even before the PbD.
ensures flexible extensibility and adaptability. Table 1 shows a comparison between the two methods and their
respective image processing time, as well as the average recall on the
Industrial 3D Object Detection Dataset (ARITODD). The average recall
score indicates the accuracy of the object’s pose estimation within spe-
cific parameters. A higher score is desirable. This score was determined
for the ITODD dataset, which has a strong focus on objects and require-
ments from industrial use cases. Therefore, we chose this evaluation
metric as it best meets the requirements of our applications.
Table 1
Comparison of used pose estimation methods.
Fig. 2. Relation between state, operation virtual indicator and object collider (green
2.1. 6D pose estimation as a key component for PBD with dashed line) within the hand-interaction model.
Please cite this article as: S. Blankemeyer et al., A hand-interaction model for augmented reality enhanced human-robot collaboration, CIRP
Annals - Manufacturing Technology (2024), https://doi.org/10.1016/j.cirp.2024.04.001
JID: CIRP
ARTICLE IN PRESS [m191;May 23, 2024;17:50]
the dimensions of the CAD model, as deviations in pose estimation or initial position, capturing during motion and the final position. Each
fingertip detection can lead to deviations between the assumed and method has specific advantages and disadvantages. Therefore, the most
actual distance between the hand and the object. As soon as two or suitable one depends on the respective requirements such as tasks and
more fingertips are in constant contact, the operation is set to grasp and environmental conditions. We have selected the methods to best match
enters the “in-contact” transition state. The points of contact are the requirements of SMEs with a rapidly changing product range.
recorded and can be used later to determine a possible gripping position
for the end effector. Gripping connects the CAD object to the root of the
operator’s hand. During object manipulation, the operator’s fingertips
may be covered, making the root of the hand a more stable option as it
can be continuously tracked. The position of the movement is continu-
ously recorded during the “in-hand” state. To recognise when the dem-
onstration is finished, a small box collider with a distance of 1.5 cm to
the object is applied. Once the fingertips are no longer in contact with
the collider, the state returns to ‘in-contact’, concluding the PbD.
During the demonstration, the interaction between hand and
object colliders is used to derive MTM base operations (reach, grasp,
move, position and release). Virtual indicators display the respective
active operation to the operator. If there are any deviations between
the real and derived operations, the operator can recognise them and
take appropriate action. To complete the PbD, a robot program can be Fig. 4. Stages of object tracking and respective different methods of pose estimation
created based on the demonstrated MTM sequence. A similar (used methods are highlighted).
approach was presented in our previous work [5].
Fig. 3. Procedure and necessary steps for our programming by demonstration approach.
Please cite this article as: S. Blankemeyer et al., A hand-interaction model for augmented reality enhanced human-robot collaboration, CIRP
Annals - Manufacturing Technology (2024), https://doi.org/10.1016/j.cirp.2024.04.001
JID: CIRP
ARTICLE IN PRESS [m191;May 23, 2024;17:50]
The operator’s workstation is shown in Fig. 5 and includes a light- The pressure to remain competitive is forcing companies to auto-
weight robot (in this case a LBR iiwa from KUKA), the components of mate their production processes. Lightweight robots already offer a suit-
the assembly and the operator with the head-mounted augmented able opportunity to take advantage of the strengths of automation while
reality device (HoloLens2 from Microsoft). The components are maintaining a high level of flexibility. We present an approach to sim-
placed on the markers that define the ROI for pose estimation. Once plify the programming of such workstations in this paper. The various
the pose of the real components has been determined, they are steps and components of the overall approach are briefly presented and
superimposed as virtual holograms in the HMD. This approach offers explained. The focus is on the human-machine interface with its main
the advantage, particularly for SMEs with a large number of variants, components of pose estimation and the novel hand-interaction model.
that production processes can be adapted quickly by exchanging The AR interface allows the operator to continuously monitor and adjust
markers and estimating the pose without the time-consuming crea- the process without in-depth programming knowledge. The resulting
tion of training data. process representation is available as an MTM sequence and can then
be transferred to a robot program or used for further process analysis.
The presented HRC Hub is still in activate development and will
be particularly useful for SMEs. The hand-interaction model with the
human-in-the-loop approach allows the specific process knowledge
of the operator to be used to obtain a reliable and correct demonstra-
tion output. Combined with the flexible hardware, this provides an
optimal solution for rapid process adaptation.
Further work can be carried out on selecting and combining
appropriate object tracking approaches based on application tasks.
This can enhance usability and streamline processes. Additionally, it
can be transferred to service robotics and skilled trades, where intui-
tive programming is also a challenge.
The functionality of the system is illustrated in Fig. 6. The figure Sebastian Blankemeyer: Conceptualization, Investigation, Meth-
shows the scenery that the operator can perceive during the execution odology, Writing original draft. David Wendorff: Investigation,
of the assembly process (real and virtual content). The operator has Software, Visualization, Writing original draft. Annika Raatz: Con-
picked up a component and is moving it to its target position (defined ceptualization, Funding acquisition, Project administration, Supervi-
by the operator). The red virtual indicator signals that the current sion, Writing review & editing.
operation is “Move”. Additional content about the components, such
as the name, can be displayed to the operator if required. As can be References
seen, there is a deviation between the real and virtual workpiece,
which is linked to the operator’s hand. This is due to an error in the [1] Keme ny Z, Vancza J, Wang L, Wang XV (2021) Human Robot Collaboration in
Manufacturing: A Multi-Agent View, Advanced Human Robot Collaboration in
previous pose estimation. However, the human-in-the-loop approach
Manufacturing, Springer, 3–43.
allows the correction by restarting the PbD. If the demonstration was [2] Wang L, Gao R, Vancza J, Kru € ger J, Wang XV, Makris S, Chryssolouris G (2019)
successful, the derived MTM-based representation can be translated Symbiotic Human-Robot Collaborative Assembly. CIRP Annals - Manufacturing
into a robot program and sent to the robot. Thus, our approach offers a Technology 68(2):701–726.
[3] Raatz A, Blankemeyer S, Recker T, Pischke D, Nyhuis P (2020) Task Scheduling
quick and intuitive method for utilising the operator’s process knowl- Method for HRC Workplaces Based on Capabilities and Execution Time Assump-
edge and combining it with the robot’s strengths (e.g. endurance) and tions for Robots. CIRP Annals - Manufacturing Technology 69(1):13–16.
the human’s strengths (e.g. ability to react to unexpected events). [4] Liu S, Wang XV, Wang L (2022) Digital Twin-Enabled Advance Execution for Human-
Robot Collaborative Assembly. CIRP Annals - Manufacturing Technology 71(1):25–28.
Using a HMD in HRC workstations has the advantage of ambidexterity, [5] Blankemeyer S, Wiemann R, Posniak L, Pregizer C, Raatz A (2018) Intuitive Robot
meeting accessibility requirements and enabling the use of both hands Programming Using Augmented Reality. Procedia CIRP 76:155–160.
for operational tasks. [6] Nee AYC, Ong SK, Chryssolouris G, Mourtzis D (2012) Augmented Reality Applica-
tions in Design and Manufacturing. CIRP Annals - Manufacturing Technology 61
(2):657–679.
[7] Overmeyer L, Ju € tte L, Poschke A (2023) A Real-Time Augmented Reality System to
See Through Forklift Components. CIRP Annals - Manufacturing Technology 72
(1):409–412.
[8] Makris S, Karagiannis P, Koukas S, Matthaiakis A-S (2016) Augmented Reality Sys-
tem for Operator Support in Human Robot Collaborative Assembly. CIRP Annals -
Manufacturing Technology 65(1):61–64.
[9] Wang XV, Wang L, Lei M, Zhao Y (2020) Closed-Loop Augmented Reality Towards
Accurate Human-Robot Collaboration. CIRP Annals - Manufacturing Technology 69
(1):425–428.
[10] Makris S, Pintzos G, Rentzos L, Chryssolouris G (2013) Assembly Support Using AR
Technology Based On Automatic Sequence Generation. CIRP Annals - Manufactur-
ing Technology 62(1):9–12.
[11] Su Y, Saleh M, Fetzer T, Rambach J, Navab N, Busam B, Stricker D, Tombari F
(2022) ZebraPose: Coarse to Fine Surface Encoding for 6DoF Object Pose Estima-
tion. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Rec-
ognition (CVPR), 6738–6748.
[12] Labbe Y, Manuelli L, Mousavian A, Tyree S, Birchfield S, Tremblay J, Carpentier J,
Aubry M, Fox D, Sivic J (2023) MegaPose: 6D Pose Estimation of Novel Objects Via
Fig. 6. Photo recorded with HoloLens 2 showing holographic virtual objects during the Render & Compare. In: Proceedings of the 6th Conference on Robot Learning (CoRL),
programming by demonstration. PMLR, 205:715–725.
Please cite this article as: S. Blankemeyer et al., A hand-interaction model for augmented reality enhanced human-robot collaboration, CIRP
Annals - Manufacturing Technology (2024), https://doi.org/10.1016/j.cirp.2024.04.001