Abstract
Human-robot collaboration has gained a notable prominence in Industry 4.0, as the use of collaborative robots increases efficiency and productivity in the automation process. However, it is necessary to consider the use of mechanisms that increase security in these environments, as the literature reports that risk situations may exist in the context of human-robot collaboration. One of the strategies that can be adopted is the visual recognition of the collaboration environment using machine learning techniques, which can automatically identify what is happening in the scene and what may happen in the future. In this work, we are proposing a new framework that is capable of detecting robotic arm keypoints commonly used in Industry 4.0. In addition to detecting, the proposed framework is able to predict the future movement of these robotic arms, thus providing relevant information that can be considered in the recognition of the human-robot collaboration scenario. The proposed framework has two main modules. The first one contains a convolutional neural network based on self-calibrated convolutions enabling better discriminative feature extraction and the support of extreme learning machine neural networks with different kernels for predicting robotic arm keypoints. The second module is composed of deep recurrent learning models, such as long short-term memory and gated recurrent unit. These models are able to predict future robotic arm keypoints. All experiments were evaluated using the mean squared error metric. Results show that the proposed framework is capable of detecting and predicting with low error, contributing to the mitigation of risks in human-robot collaboration. In addition, it was possible to verify that the use of convolutional neural networks in conjunction with extreme learning machines can offer a lower detection error in a regression task (e.g., keypoint detection), something that, as far as the authors are aware of, is not yet known, nor had been evaluated previously in the literature.
Similar content being viewed by others
Data availability
The datasets generated during and/or analyzed during the current study are available from the corresponding author on reasonable request.
References
Ruchi Goel, Pooja Gupta (2020) Robotics and industry 4.0: Sharp Business and Sustainable Development. A roadmap to industry 4.0: smart production. Springer, pp 157–169
Eloise M, Riccardo M, Zampieri Emanuele GG, Maurizio F, Giulio R (2019) Human-robot collaboration in manufacturing applications: a review. Robotics 8(4):100
Semeraro F, Griffiths A, Cangelosi A (2023) Human-robot collaboration and machine learning: a systematic review of recent research. Robotics Comput-Integr Manufact 79:102432
Arash A, Maria ZA, Serena I, Alin A-S, Kazuhiro K, Oussama K (2018) Progress and prospects of the human-robot collaboration. Auton Robot 42(5):957–975
Bauer A, Wollherr D, Buss M (2008) Human-robot collaboration: a survey. Int J Humanoid Rob 5(01):47–66
Ehsan H-PS, Simon T, Sergey K, Alexandre D (2020) Operations management issues in design and control of hybrid human-robot collaborative manufacturing systems: a survey. Annu Rev Control 49:264–276
Lakomkin E, Zamani MA, Weber C, Magg S, Wermter S (2018) On the robustness of speech emotion recognition for human-robot interaction with deep neural networks. In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp 854–860. IEEE
Sridharan M, Meadows B (2019) Towards a theory of explanations for human-robot collaboration. KI-Künstliche Intell 33(4):331–342
Fragapane Giuseppe, Hvolby Hans-Henrik, Sgarbossa Fabio, Strandhagen Jan Ola (2020) Autonomous mobile robots in hospital logistics. In: IFIP International Conference on Advances in Production Management Systems, Springer, pp 672–679
Microsoft (2019) Microsoft dynamics 365 manufacturing trends report, 2019. Accessed: 2019-09-09
Reis G, Dantas M, Bezerra D, Nunes G, Dreyer P, Ledebour C, Kelner J, Sadok D, Souza R, Lins S et al (2021) Gripper design for radio base station autonomous maintenance system. Int J Autom Comput 18:1–9
Thors B, Furuskär A, Colombi D, Törnevik C (2017) Time-averaged realistic maximum power levels for the assessment of radio frequency exposure for 5g radio base stations using massive mimo. IEEE Access 5:19711–19719
Vasic M, Billard A(2013) Safety issues in human-robot interactions. In: 2013 IEEE International Conference on Robotics and Automation, pp 197–204. IEEE
Rodrigues IR, Barbosa G, Oliveira Filho A, Cani C, Dantas M, Sadok DH, Kelner J, Souza RS, Marquezini MV, Lins S (2021) Modeling and assessing an intelligent system for safety in human-robot collaboration using deep and machine learning techniques. Multi Tools Appl 81:2213–2239
Jianjing Zhang, Hongyi Liu, Qing Chang, Lihui Wang, Gao Robert X (2020) Recurrent neural network for motion trajectory prediction in human-robot collaborative assembly. CIRP Annals 69(1):9–12
Anvaripour M, Saif M (2019) Collision detection for human-robot interaction in an industrial setting using force myography and a deep learning approach. In: 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), pp 2149–2154. IEEE
Maceira M, Olivares-Alarcos A, Alenyà G (2020) Recurrent neural networks for inferring intentions in shared tasks for industrial collaborative robots. In: 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), pp 665–670. IEEE
Miseikis J, Knobelreiter P, Brijacak I, Yahyanejad S, Glette K, Elle OJ, Torresen J (2018) Robot localisation and 3d position estimation using a free-moving camera and cascaded convolutional neural networks. In: 2018 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM), pages 181–187. IEEE
Zhou F, Chi Z, Zhuang C, Ding H (2019) 3D pose estimation of robot arm with rgb images based on deep learning. In: International Conference on Intelligent Robotics and Applications, pp 541–553. Springer
Lee TE, Tremblay J, To T, Cheng J, Mosier T, Kroemer O, Fox D, Birchfield S (2020) Camera-to-robot pose estimation from a single image. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp 9426–9432. IEEE
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444
Liu J-J, Hou Q, Cheng M-M, Wang C, Feng J (2020) Improving convolutional networks with self-calibrated convolutions. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 10096–10105
Huang G-B, Zhu Q-Y, Siew C-K (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1–3):489–501
Rodrigues IR, da Silva Neto SR, Kelner J, Sadok D, Endo PT (2011) Convolutional extreme learning machines: a systematic review. Informatics 8:33
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Cho K, van Merriënboer B , Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, October. Association for Computational Linguistics, pp 1724–1734
Silva IR, Barbosa GB, Ledebour CC, Oliveira Filho AT, Kelner J, Sadok D, Lins S, Souza R (2020) Assessing deep learning models for human-robot collaboration collision detection in industrial environments. In: Brazilian Conference on Intelligent Systems, Springer, pp 240–255
Robla-Gòmez S, Becerra VM, Lltata JR, Gonzalez-Sarabia E, Torre-Ferrero C, Juan P-O (2017) Working together: a review on safe human-robot collaboration in industrial environments. IEEE Access 5:26754–26773
Lasota PA, Fong T, Shah JA et al (2017) A survey of methods for safe human-robot interaction. Found Trends Robot 5(4):261–349
Deng J, Dong W, Socher R, Li L-J, Li K, F-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp 248–255. IEEE
Tan M, Le Q (2019) Efficientnet: Rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning, pp 6105–6114. PMLR
Liu W, Wang Z, Liu X, Zeng N, Liu Y, Alsaadi FE (2017) A survey of deep neural network architectures and their applications. Neurocomputing 234:11–26
Lipton ZC, Berkowitz J, Elkan C (2015) A critical review of recurrent neural networks for sequence learning. arXiv preprint, arXiv:1506.00019
Abdel-Nasser S, Koustoumpardis Panagiotis N, Nikos A (2020) Human-robot collisions detection for safe human-robot interaction using one multi-input-output neural network. Soft Comput 24(9):6687–6719
Min PK, Jihwan K, Jinhyuk P, Park Frank C (2021) Learning-based real-time detection of robot collisions without joint torque sensors. IEEE Robot Autom Lett 6(1):103–110
Alex K, Ilya S, Hinton Geoffrey E (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097–1105
Mišeikis J, Brijacak I, Yahyanejad S, Glette K, Elle OJ, Torresen J (2018) Transfer learning for unseen robot detection and joint estimation on a multi-objective convolutional neural network. In: 2018 IEEE International Conference on Intelligence and Safety for Robotics (ISR), pp 337–342. IEEE
Heindl C, Zambal S, Scharinger J (2019) Learning to predict robot keypoints using artificially generated images. In: 2019 24th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), pp 1536–1539. IEEE
Ballas N, Li Y, Pal C, Courville A (2015) Delving deeper into convolutional networks for learning video representations. arXiv preprintarXiv:1511.06432, 2015
Heindl C, Zambal S, Ponitz T, Pichler A, Scharinger J (2019) 3D robot pose estimation from 2D images. arXiv preprint, arXiv:1902.04987
da Silva Neto SR, Tabosa Oliveira T, Teixeira IV, Aguiar de Oliveira SB, Souza Sampaio V, Lynn T, Endo PT (2022) Machine learning and deep learning techniques to support clinical diagnosis of arboviral diseases: A systematic review. PLoS Negl Trop Dis 16(1):e0010061
Jiuxiang G, Wang Z, Kuen J, Ma L, Shahroudy A, Shuai B, Liu T, Wang X, Wang G, Cai J et al (2018) Recent advances in convolutional neural networks. Pattern Recogn 77:354–377
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 770–778
Cui D, Zhang G, Han W, Lekamalage Chamara Kasun L, Hu K, Huang G-B (2017) Compact feature representation for image classification using elms. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp 1015–1022
Guang-Bin H, Hui WD, Yuan L (2011) Extreme learning machines: a survey. Int J Mach Learn Cybern 2(2):107–122
filters combination and error model validation (2019) Michel M dos Santos, Abel G da Silva Filho, and Wellington P dos Santos. Deep convolutional extreme learning machines. Neurocomputing 329:359–369
Huang F, Jun L, Tao J, Li L, Tan X, Liu P (2019) Research on optimization methods of elm classification algorithm for hyperspectral remote sensing images. IEEE Access 7:108070–108089
Li D, Qiu X, Zhu Z, Liu Y (2018) Criminal investigation image classification based on spatial cnn features and elm. In: 2018 10th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), Vol 2, pp 294–298. IEEE
Pu H, Zhai J-H, Zhang S-F (2017) A simple and effective method for image classification. In: 2017 International Conference on Machine Learning and Cybernetics (ICMLC), vol 1, pp 230–235. IEEE
Khellal A, Ma H, Fei Q (2018) Convolutional neural network features comparison between back-propagation and extreme learning machine. In: 2018 37th Chinese Control Conference (CCC), pp 9629–9634. IEEE
Lu S, Xia K, Wang S-H (2020) Diagnosis of cerebral microbleed via vgg and extreme learning machine trained by gaussian map bat algorithm. J Ambient Intell Humanized Computi, pp 1–12
Ijjina EP (2017) Human action recognition in rgb-d videos using motion sequence information and deep learning. Pattern Recogn 72:504–516
Zaki Hasan FM, Faisal S, Ajmal M (2019) Viewpoint invariant semantic object and scene categorization with rgb-d sensors. Auton Robot 43(4):1005–1022
Huang Jinghong Yu, Liang Z, Cai Zhaoquan G, Zhenghui CZ, Gao Wei Yu, Qianyun SD (2017) Extreme learning machine with multi-scale local receptive fields for texture classification. Multi Syst Signal Process 28(3):995–1011
Rezaeenour J, Ahmadi M, Jelodar H and Shahrooei R (2022) Systematic review of content analysis algorithms based on deep neural networks. Multimedia Tools and Applications
Rumelhart DE, Hinton GE, Williams RJ (1986) Learning representations by back-propagating errors. Nature 323(6088):533–536
Kanagachidambaresan GR, Ruwali A, Debrup B, Prakash KB (2021) Recurrent neural network. Springer International Publishing, Cham, pp 53–61
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Rengasamy D, Jafari M, Rothwell B, Chen X, Figueredo GP (2020) Deep learning with dynamically weighted loss function for sensor-based prognostics and health management. Sensors 20(3):723
Chung J, Gulcehre C, Kyunghyun C and Yoshua B (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. In: NIPS 2014 Workshop on Deep Learning, December 2014
He T, Zhang Z, Zhang H, Zhang Z, Xie J, Li M (2019) Bag of tricks for image classification with convolutional neural networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 558–567
Huang G, Huang G-B, Song S, You K (2015) Trends in extreme learning machines: a review. Neural Netw 61:32–48
Ribeiro AMNC, do Carmo PRX, Rodrigues IR, Sadok D, Lynn T, Endo PT (2020) Short-term firm-level energy-consumption forecasting for energy-intensive manufacturing: a comparison of machine learning and deep learning models. Algorithms 13(11):274
Cho K, Van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv preprint, arXiv:1406.1078
Universal Robots (2021) Universial robots. https://www.universal-robots.com/. accessed in november
Dutta A, Zisserman A (2019) The via annotation software for images, audio and video. In: Proceedings of the 27th ACM International Conference on Multimedia, pp 2276–2279
Iandola FN, Han S, Moskewicz MW, Ashraf K, Dally WJ, Keutzer K (2016) Squeezenet: Alexnet-level accuracy with 50x fewer parameters and< 0.5 mb model size. arXiv preprint, arXiv:1602.07360
Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings
Huang G, Liu Z, Van Der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern recognition, pp 4700–4708
Huang G-B, Siew C-K (2005) Extreme learning machine with randomly assigned rbf kernels. Int J Inf Technol 11(1):16–24
Baraha S, Biswal PK (2017) Implementation of activation functions for elm based classifiers. In: 2017 International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET), pp 1038–1042. IEEE
Zhang C, Benz P, Argaw DM, Lee S, Kim J, Rameau F, Bazin J-C, Kweon IS (2021) Resnet or densenet? Introducing dense shortcuts to resnet. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp 3550–3559
Yang S, Yu X, Zhou Y (2020) Lstm and gru neural network performance comparison study: Taking yelp review dataset as an example. In: 2020 International workshop on electronic communication and artificial intelligence (IWECAI), pp 98–101. IEEE
Patel MM, Tanwar S, Gupta R, Kumar N (2020) A deep learning-based cryptocurrency price prediction scheme for financial institutions. J Inform Security Appl 55:102583
Acknowledgements
This work was financed in part by the Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq), Fundação de Amparo a Ciência e Tecnologia de Pernambuco (FACEPE), Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES), and Research, Development and Innovation Center, Ericsson Telecommunications Inc., Brazil.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Rodrigues, I.R., Dantas, M., de Oliveira Filho, A.T. et al. A framework for robotic arm pose estimation and movement prediction based on deep and extreme learning models. J Supercomput 79, 7176–7205 (2023). https://doi.org/10.1007/s11227-022-04936-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11227-022-04936-z