Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

A real-time human-robot interaction framework with robust background invariant hand gesture detection

Published: 01 December 2019 Publication History

Highlights

Kinect V2 is integrated with OpenPose to get a 3D estimate of human skeleton.
Background invariant hand gesture detector is trained by background substitution.
Inception V3 is adapted for 10 static hand gestures from American Sign Language.
Our safe pHRI library OpenPHRI is integrated to ensure effective and safe robot control.
Teaching by gestures is presented while safety is guaranteed by separation distance monitoring.

Abstract

In the light of factories of the future, to ensure productive and safe interaction between robot and human coworkers, it is imperative that the robot extracts the essential information of the coworker. We address this by designing a reliable framework for real-time safe human-robot collaboration, using static hand gestures and 3D skeleton extraction. OpenPose library is integrated with Microsoft Kinect V2, to obtain a 3D estimation of the human skeleton. With the help of 10 volunteers, we recorded an image dataset of alpha-numeric static hand gestures, taken from the American Sign Language. We named our dataset OpenSign and released it to the community for benchmarking. Inception V3 convolutional neural network is adapted and trained to detect the hand gestures. To augment the data for training the hand gesture detector, we use OpenPose to localize the hands in the dataset images and segment the backgrounds of hand images, by exploiting the Kinect V2 depth map. Then, the backgrounds are substituted with random patterns and indoor architecture templates. Fine-tuning of Inception V3 is performed in three phases, to achieve validation accuracy of 99.1% and test accuracy of 98.9%. An asynchronous integration of image acquisition and hand gesture detection is performed to ensure real-time detection of hand gestures. Finally, the proposed framework is integrated in our physical human-robot interaction library OpenPHRI. This integration complements OpenPHRI by providing successful implementation of the ISO/TS 15066 safety standards for “safety rated monitored stop” and “speed and separation monitoring” collaborative modes. We validate the performance of the proposed framework through a complete teaching by demonstration experiment with a robotic manipulator.

References

[1]
D. Gorecky, M. Schmitt, M. Loskyll, D. Zühlke, Human-machine-interaction in the industry 4.0 era, Proceedings of the IEEE International Conference on Industrial Informatics, 2014, pp. 289–294.
[2]
B. Gleeson, K. MacLean, A. Haddadi, E. Croft, J. Alcazar, Gestures for industry: intuitive human-robot communication from human observation, Proceedings of the 8th ACM/IEEE International Conference on Human-robot Interaction, IEEE Press, Piscataway, NJ, USA, 2013, pp. 349–356.
[3]
A. Cherubini, R. Passama, A. Crosnier, A. Lasnier, P. Fraisse, Collaborative manufacturing with physical human–robot interaction, Robot. Comput. Integr. Manuf. 40 (2016) 1–13.
[4]
C. Schlenoff, Z. Kootbally, A. Pietromartire, M. Franaszek, S. Foufou, Intention recognition in manufacturing applications, Robot. Comput. Integr. Manuf. 33 (2015) 29–41.
[5]
I. Maurtua, A. Ibarguren, J. Kildal, L. Susperregi, B. Sierra, Human-robot collaboration in industrial applications: safety, interaction and trust, Int. J. Adv. Robot. Syst. 14 (2017) 1–10.
[6]
S.S. Rautaray, A. Agrawal, Vision based hand gesture recognition for human computer interaction: a survey, Artif. Intell. Rev. 43 (2015) 1–54.
[7]
P. Neto, J.N. Pires, A.P. Moreira, High-level programming and control for industrial robotics: using a hand-held accelerometer-based input device for gesture and posture recognition, Ind. Robot. 37 (2010) pp137–147.
[8]
V. Villani, F. Pini, F. Leali, C. Secchi, Survey on human–robot collaboration in industrial settings: safety, intuitive interfaces and applications, Mechatronics 55 (2018) 248–266.
[9]
P. Neto, D. Pereira, J.N. Pires, A.P. Moreira, Gesture recognition for human-robot collaboration: a review, Proceedings of the 7th Swedish Production Symposium, 2016, pp. 1–12.
[10]
J. Ruiz-del-Solar, P. Loncomilla, N. Soto, A survey on deep learning methods for robot vision, 2018, arxiv:1803.10862.
[11]
R. Yang, S. Sarkar, B. Loeding, Handling movement epenthesis and hand segmentation ambiguities in continuous sign language recognition using nested dynamic programming, IEEE Trans. Pattern Anal. Mach. Intell. 32 (2010) 462–477.
[12]
B. Navarro, A. Fonte, P. Fraisse, G. Poisson, A. Cherubini, In pursuit of safety: an open-Source library for physical human-robot interaction, IEEE Robot. Autom. Mag. 25 (2018) 39–50.
[13]
A. De Luca, F. Flacco, Integrated control for pHRI: collision avoidance, detection, reaction and collaboration, Proceedings of the 4th IEEE RAS & EMBS International Conference on Biomedical Robotics and Biomechatronics, IEEE, 2012, pp. 288–295.
[14]
ISO 10218-1:2011 - Robots and robotic devices - Safety requirements for industrial robots - Part 1: Robots. ISO; 2011(2011).
[15]
ISO 10218-2:2011 - Robots and robotic devices - Safety requirements for industrial robots - Part 2: Robot systems and integration. ISO; 2011 (2011).
[16]
ISO/TS 15066:2016 - Robots and robotic devices - Safety requirements for collaborative industrial robot systems and the work environment: Collaborative robots. ISO/TS; 2016 (2011).
[17]
A. Levratti, G. Riggio, A. De Vuono, C. Fantuzzi, C. Secchi, Safe navigation and experimental evaluation of a novel tire workshop assistant robot, Proceedings of the IEEE International Conference on Robotics and Automation, 2017, pp. 994–999.
[18]
L. Sabattini, A. Levratti, F. Venturi, E. Amplo, C. Fantuzzi, C. Secchi, Experimental comparison of 3D vision sensors for mobile robot localization for industrial application: Stereo-camera and RGB-D sensor, Proceedings of the 12th IEEE International Conference on Control Automation Robotics Vision, 2012, pp. 823–828.
[19]
J.A. Marvel, R. Norcross, Implementing speed and separation monitoring in collaborative robot workcells, Robot. Comput. Integr. Manuf. 44 (2017) pp.144–155.
[20]
V. Rampa, F. Vicentini, S. Savazzi, N. Pedrocchi, M. Ioppolo, M. Giussani, Safe human-robot cooperation through sensor-less radio localization, Proceedings of the 12th IEEE International Conference on Industrial Informatics, 2014, pp. 683–689.
[21]
F. Vicentini, M. Ruggeri, L. Dariz, A. Pecora, L. Maiolo, D. Polese, L. Pazzini, L.M. Tosatti, Wireless sensor networks and safe protocols for user tracking in human-robot cooperative workspaces, Proceedings of the 23rd IEEE International Symposium on Industrial Electronics, 2014, pp. 1274–1279.
[22]
F. Vicentini, M. Giussani, L.M. Tosatti, Trajectory-dependent safe distances in human-robot interaction, Proceedings of the Emerging Technology and Factory Automation, IEEE, 2014, pp. 1–4.
[23]
P.A. Lasota, G.F. Rossano, J.A. Shah, Toward safe close-proximity human-robot interaction with standard industrial robots, in: 2014 IEEE International Conference on Automation Science and Engineering (CASE), Taipei, 2014, pp. 339–344.
[24]
C. Morato, K.N. Kaipa, B. Zhao, S.K. Gupta, Toward safe human robot collaboration by using multiple kinects based real-time human tracking, J. Comput. Inf. Sci. Eng. 14 (2014).
[25]
S. Makris, P. Tsarouchi, D. Surdilovic, J. Krüger, Intuitive dual arm robot programming for assembly operations, CIRP Ann. Manuf. Technol. 63 (2014) 13–16.
[26]
Z. Zhang, Microsoft kinect sensor and its effect, IEEE Multimed. 19 (2012) 4–10.
[27]
Y. Yang, H. Yan, M. Dehghan, M.H. Ang, Real-time human-robot interaction in complex environment using kinect v2 image recognition, Proceedings of the IEEE International Conference on Cybernetics and Intelligent Systems and IEEE Conference on Robotics, Automation and Mechatronics, 2015, pp. 112–117.
[28]
M.G. Carmichael, D. Liu, A. Tran, R. Khonasty, S. Aldini, Robot co-worker for abrasive blasting: lessons learnt in worker posture estimation, Proceedings of the IEEE International Conference on Robotics and Automation, 2018.
[29]
J. Sell, P. O’Connor, The xbox one system on a chip and kinect sensor, IEEE Micro 34 (2014) 44–53.
[30]
L. Keselman, J.I. Woodfill, A. Grunnet-Jepsen, A. Bhowmik, Intel(R) RealSense(TM) Stereoscopic Depth Cameras, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017, pp. 1267–1276.
[31]
G. Du, P. Zhang, Markerless human–robot interface for dual robot manipulators using kinect sensor, Robot. Comput. Integr. Manuf. 30 (2014) pp.150–159.
[32]
G. Du, M. Chen, C. Liu, B. Zhang and P. Zhang, Online Robot Teaching With Natural Human–Robot Interaction, in IEEE Transactions on Industrial Electronics, vol. 65, no. 12, pp. 9571-9581, Dec. 2018.
[33]
E. Matsas, G.-C. Vosniakos, Design of a virtual reality training system for human–robot collaboration in manufacturing tasks, Int. J. Interact. Des. Manuf. 11 (2017) 139–153.
[34]
E. Matsas, G.-C. Vosniakos, D. Batras, Prototyping proactive and adaptive techniques for human-robot collaboration in manufacturing using virtual reality, Robot. Comput. Integr. Manuf. 50 (2018) pp.168–180.
[35]
M. Simão, P. Neto, O. Gibaru, Natural control of an industrial robot using hand gesture recognition with neural networks, Proceedings of the 42nd Annual Conference of the IEEE Industrial Electronics Society, 2016, pp. 5322–5327.
[36]
G. Canal, S. Escalera, C. Angulo, A real-time human-robot interaction system based on gestures for assistive scenarios, Comput. Vis. Image Underst. 149 (2016) pp.65–77.
[37]
J.L. Raheja, M. Chandra, A. Chaudhary, 3D Gesture based real-time object selection and recognition, Pattern Recognit. Lett. 115 (2017) 14–19.
[38]
Y. LeCun, F.J. Huang, L. Bottou, Learning methods for generic object recognition with invariance to pose and lighting, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2, 2004, pp. 97–104.
[39]
K. Ehlers, K. Brama, A human-robot interaction interface for mobile and stationary robots based on real-time 3D human body and hand-finger pose estimation, Proceedings of the IEEE International Conference on Emerging Technologies and Factory Automation, 2016, pp. 1–6.
[40]
T. Starner, J. Weaver, A. Pentland, Real-time american sign language recognition using desk and wearable computer based video, IEEE Trans Pattern Anal Mach Intell 20 (1998) 1371–1375.
[41]
P. Neto, D. Pereira, J.N. Pires, A.P. Moreira, Real-time and continuous hand gesture spotting: An approach based on artificial neural networks, Proceedings of the IEEE International Conference on Robotics and Automation, 2013, pp. 178–183.
[42]
Q. Munib, M. Habeeb, B. Takruri, H.A. Al-Malik, American sign language (ASL) recognition based on hough transform and neural networks, Expert Syst. Appl. 32 (2007) 24–37.
[43]
C. Charayaphan, A. Marble, Image processing system for interpreting motion in american sign language, Med. Eng. Phys. 14 (1992) pp.419–425.
[44]
D.H. Ballard, Generalizing the hough transform to detect arbitrary shapes, Pattern Recognit. 13 (1981) 111–122.
[45]
A. Joshi, H. Sierra, E. Arzuaga, American sign language translation using edge detection and cross correlation, Proceedings of the IEEE Colombian Conference on Communications and Computing (COLCOM), 2017, pp. 1–6.
[46]
P. Xu, A real-time hand gesture recognition and human-computer interaction system, 2017, arxiv:1704.07296.
[47]
Y. LeCun, L. Bottou, Y. Bengio, P. Haffner, Gradient-based learning applied to document recognition, Proc. IEEE 86 (1998) 2278–2324.
[48]
K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016.
[49]
H. Cheng, L. Yang, Z. Liu, Survey on 3D hand gesture recognition., IEEE Trans. Circuits Syst. Video Techn. 26 (2016) pp.1659–1673.
[50]
M. Van den Bergh, L. Van Gool, Combining RGB and ToF cameras for real-time 3D hand gesture interaction, Proceedings of the IEEE Workshop on Applications of Computer Vision (WACV), 2011, pp. 66–72.
[51]
I. Oikonomidis, N. Kyriazis, A.A. Argyros, Markerless and efficient 26-DOF hand pose recovery, Proceedings of the Asian Conference on Computer Vision, Springer, 2010, pp. 744–757.
[52]
S.-E. Wei, V. Ramakrishna, T. Kanade, Y. Sheikh, Convolutional Pose Machines, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016.
[53]
Z. Cao, T. Simon, S.-E. Wei, Y. Sheikh, Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
[54]
P. Panteleris, I. Oikonomidis, A.A. Argyros, Using a Single RGB Frame for Real Time 3D Hand Pose Estimation in the Wild, Proceedings of the IEEE Winter Conference on Applications of Computer Vision, IEEE, 2018, pp. 436–445.
[55]
O. Mazhar, S. Ramdani, B. Navarro, R. Passama, A. Cherubini, Towards Real-time Physical Human-Robot Interaction using Skeleton Information and Hand Gestures, Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, 2018.
[56]
S. Das, M. Koperski, F. Bremond, G. Francesca, Action recognition based on a mixture of RGB and depth based skeleton, Proceedings of the 14th IEEE International Conference on Advanced Video and Signal Based Surveillance, 2017, pp. 1–6.
[57]
C. Zimmermann, T. Welschehold, C. Dornhege, T. Brox, W. Burgard, 3D human pose estimation in RGBD images for robotic task learning, Proceedings of the IEEE International Conference on Robotics and Automation, 2018.
[58]
T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, C.L. Zitnick, Microsoft COCO: Common Objects in Context, in: D. Fleet, T. Pajdla, B. Schiele, T. Tuytelaars (Eds.), Computer Vision – ECCV 2014, Springer International Publishing, Cham, 2014, pp. 740–755.
[59]
R.C. Luo, Y.C. Wu, Hand Gesture Recognition for Human-Robot Interaction for Service Robot, Proceedings of the IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems, 2012, pp. 318–323.
[60]
M.V. den Bergh, D. Carton, R.D. Nijs, N. Mitsou, C. Landsiedel, K. Kuehnlenz, D. Wollherr, L.V. Gool, M. Buss, Real-time 3D hand gesture interaction with a robot for understanding directions from humans, Proceedings of the RO-MAN, 2011, pp. 357–362.
[61]
G. Cicirelli, C. Attolico, C. Guaragnella, T. D’Orazio, A kinect-based gesture recognition approach for a natural human robot interface, Int. J. Adv. Robot. Syst. 12 (2015) 22.
[62]
M. Munaro, F. Basso, E. Menegatti, Tracking people within groups with RGB-D data, Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, 2012, pp. 2101–2107.
[63]
O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A.C. Berg, L. Fei-Fei, Imagenet large scale visual recognition challenge, Int. J. Comput. Vis. 115 (2015) pp.211–252.
[64]
C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, Z. Wojna, Rethinking the inception architecture for computer vision, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 2818–2826.
[65]
F. Chollet, et al., Keras, 2015.
[66]
R. Battison, Lexical Borrowing in American Sign Language, 20901, Linstok Press, Inc., 9306 Mintwood Street, Silver Spring, Maryland, 1978.
[67]
A. Cherubini, R. Passama, B. Navarro, M. Sorour, A. Khelloufi, O. Mazhar, S. Tarbouriech, J. Zhu, O. Tempier, A. Crosnier, P. Fraisse, S. Ramdani, A collaborative robot for the factory of the future: BAZAR, Int. J. Adv. Manuf. Technol. (2019) 1–17. Special Issue Design and Management of Digital Manufacturing & Assembly Systems in the Industry 4.0 era.

Cited By

View all
  • (2024)Learning by doingRobotics and Computer-Integrated Manufacturing10.1016/j.rcim.2023.10267386:COnline publication date: 1-Apr-2024
  • (2024)Sensor-based human–robot collaboration for industrial tasksRobotics and Computer-Integrated Manufacturing10.1016/j.rcim.2023.10266386:COnline publication date: 1-Apr-2024
  • (2023)Robotic Vision for Human-Robot Interaction and Collaboration: A Survey and Systematic ReviewACM Transactions on Human-Robot Interaction10.1145/357073112:1(1-66)Online publication date: 16-Feb-2023
  • Show More Cited By

Index Terms

  1. A real-time human-robot interaction framework with robust background invariant hand gesture detection
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Information & Contributors

          Information

          Published In

          cover image Robotics and Computer-Integrated Manufacturing
          Robotics and Computer-Integrated Manufacturing  Volume 60, Issue C
          Dec 2019
          77 pages

          Publisher

          Pergamon Press, Inc.

          United States

          Publication History

          Published: 01 December 2019

          Author Tags

          1. Physical human-robot interaction
          2. Safe collaborative robotics
          3. Convolutional neural networks
          4. Real-time vision
          5. Transfer learning

          Qualifiers

          • Research-article

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • Downloads (Last 12 months)0
          • Downloads (Last 6 weeks)0
          Reflects downloads up to 13 Sep 2024

          Other Metrics

          Citations

          Cited By

          View all
          • (2024)Learning by doingRobotics and Computer-Integrated Manufacturing10.1016/j.rcim.2023.10267386:COnline publication date: 1-Apr-2024
          • (2024)Sensor-based human–robot collaboration for industrial tasksRobotics and Computer-Integrated Manufacturing10.1016/j.rcim.2023.10266386:COnline publication date: 1-Apr-2024
          • (2023)Robotic Vision for Human-Robot Interaction and Collaboration: A Survey and Systematic ReviewACM Transactions on Human-Robot Interaction10.1145/357073112:1(1-66)Online publication date: 16-Feb-2023
          • (2023)A versatile interaction framework for robot programming based on hand gestures and posesRobotics and Computer-Integrated Manufacturing10.1016/j.rcim.2023.10259684:COnline publication date: 1-Dec-2023
          • (2023)Thermal image-based hand gesture recognition for worker-robot collaboration in the construction industryAdvanced Engineering Informatics10.1016/j.aei.2023.10193956:COnline publication date: 1-Apr-2023
          • (2023)Attention Guided In-hand Mechanical Tools Recognition in Human-Robot Collaborative ProcessInteractive Collaborative Robotics10.1007/978-3-031-43111-1_1(1-12)Online publication date: 25-Oct-2023
          • (2022)Vision-based holistic scene understanding towards proactive human–robot collaborationRobotics and Computer-Integrated Manufacturing10.1016/j.rcim.2021.10230475:COnline publication date: 1-Jun-2022
          • (2022)A Survey of Robot Learning Strategies for Human-Robot Collaboration in Industrial SettingsRobotics and Computer-Integrated Manufacturing10.1016/j.rcim.2021.10223173:COnline publication date: 1-Feb-2022
          • (2022)Human–robot collaboration in industrial environmentsRobotics and Computer-Integrated Manufacturing10.1016/j.rcim.2021.10220873:COnline publication date: 1-Feb-2022
          • (2022)Dynamic Speed and Separation Monitoring Based on Scene Semantic InformationJournal of Intelligent and Robotic Systems10.1007/s10846-022-01607-2106:2Online publication date: 1-Oct-2022
          • Show More Cited By

          View Options

          View options

          Get Access

          Login options

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media