{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,17]],"date-time":"2026-04-17T16:40:39Z","timestamp":1776444039458,"version":"3.51.2"},"reference-count":29,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100008530","name":"European Regional Development Fund","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100008530","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100011929","name":"Programa Operacional Tem\u00e1tico Factores de Competitividade","doi-asserted-by":"publisher","award":["POCI-01-0247-FEDER-046095"],"award-info":[{"award-number":["POCI-01-0247-FEDER-046095"]}],"id":[{"id":"10.13039\/501100011929","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100011929","name":"Programa Operacional Tem\u00e1tico Factores de Competitividade","doi-asserted-by":"publisher","award":["POCI-01-0247-FEDER-046102"],"award-info":[{"award-number":["POCI-01-0247-FEDER-046102"]}],"id":[{"id":"10.13039\/501100011929","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Robotics and Computer-Integrated Manufacturing"],"published-print":{"date-parts":[[2023,4]]},"DOI":"10.1016\/j.rcim.2022.102449","type":"journal-article","created":{"date-parts":[[2022,10,14]],"date-time":"2022-10-14T20:49:46Z","timestamp":1665780586000},"page":"102449","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":48,"special_numbering":"C","title":["Deep learning-based human action recognition to leverage context awareness in collaborative assembly"],"prefix":"10.1016","volume":"80","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3344-3620","authenticated-orcid":false,"given":"Duarte","family":"Moutinho","sequence":"first","affiliation":[]},{"given":"Lu\u00eds","family":"F. Rocha","sequence":"additional","affiliation":[]},{"given":"Carlos M.","family":"Costa","sequence":"additional","affiliation":[]},{"given":"Lu\u00eds F.","family":"Teixeira","sequence":"additional","affiliation":[]},{"given":"Germano","family":"Veiga","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.rcim.2022.102449_b1","doi-asserted-by":"crossref","first-page":"903","DOI":"10.17973\/MMSJ.2016_06_201611","article-title":"Human - robot collaboration in industry","volume":"2016-June","author":"Vysocky","year":"2016","journal-title":"MM Sci. J."},{"key":"10.1016\/j.rcim.2022.102449_b2","first-page":"48","article-title":"Industry 5.0 - Towards a sustainable, human- centric and resilient European industry","author":"Breque","year":"2021","journal-title":"Eur. Commission"},{"issue":"5","key":"10.1016\/j.rcim.2022.102449_b3","doi-asserted-by":"crossref","first-page":"957","DOI":"10.1007\/s10514-017-9677-2","article-title":"Progress and prospects of the human\u2013robot collaboration","volume":"42","author":"Ajoudani","year":"2018","journal-title":"Auton. Robots"},{"issue":"1","key":"10.1016\/j.rcim.2022.102449_b4","doi-asserted-by":"crossref","first-page":"47","DOI":"10.1142\/S0219843608001303","article-title":"Human-robot collaboration: A survey","volume":"5","author":"Bauer","year":"2008","journal-title":"Int. J. Humanoid Robot."},{"key":"10.1016\/j.rcim.2022.102449_b5","series-title":"ACM\/IEEE International Conference on Human-Robot Interaction, Vol. 2016-April","first-page":"83","article-title":"Anticipatory robot control for efficient human-robot collaboration","author":"Huang","year":"2016"},{"key":"10.1016\/j.rcim.2022.102449_b6","doi-asserted-by":"crossref","DOI":"10.1016\/j.rcim.2021.102304","article-title":"Vision-based holistic scene understanding towards proactive human\u2013robot collaboration","volume":"75","author":"Fan","year":"2022","journal-title":"Robot. Comput.-Integr. Manuf."},{"key":"10.1016\/j.rcim.2022.102449_b7","series-title":"Conference Proceedings - IEEE SOUTHEASTCON","first-page":"1","article-title":"Human-robot collaboration: A survey","author":"Chandrasekaran","year":"2015"},{"key":"10.1016\/j.rcim.2022.102449_b8","doi-asserted-by":"crossref","DOI":"10.1016\/j.rcim.2019.101830","article-title":"Bi-directional navigation intent communication using spatial augmented reality and eye-tracking glasses for improved safety in human\u2013robot interaction","volume":"61","author":"Chadalavada","year":"2020","journal-title":"Robot. Comput.-Integr. Manuf."},{"key":"10.1016\/j.rcim.2022.102449_b9","series-title":"IEEE International Conference on Intelligent Robots and Systems, Vol. 2016-Novem","first-page":"5048","article-title":"A robot reading human gaze: Why eye tracking is better than head tracking for human-robot collaboration","author":"Palinko","year":"2016"},{"key":"10.1016\/j.rcim.2022.102449_b10","series-title":"A survey of robot learning strategies for human-robot collaboration in industrial settings","author":"Mukherjee","year":"2022"},{"key":"10.1016\/j.rcim.2022.102449_b11","doi-asserted-by":"crossref","DOI":"10.1007\/s11042-020-08806-9","article-title":"Human action recognition using fusion of multiview and deep features: an application to video surveillance","author":"Khan","year":"2020","journal-title":"Multimedia Tools Appl."},{"issue":"8","key":"10.1016\/j.rcim.2022.102449_b12","doi-asserted-by":"crossref","first-page":"623","DOI":"10.1049\/iet-cvi.2016.0355","article-title":"Survey on deep learning methods in human action recognition","volume":"11","author":"Koohzadi","year":"2017","journal-title":"IET Comput. Vis."},{"key":"10.1016\/j.rcim.2022.102449_b13","first-page":"612","article-title":"Bio-intelligent selective laser melting system based on convolutional neural networks for in-process fault identification","volume":"88","author":"Angelone","year":"2020"},{"key":"10.1016\/j.rcim.2022.102449_b14","series-title":"Procedia CIRP, Vol. 83","first-page":"272","article-title":"Deep learning-based human motion prediction considering context awareness for human-robot collaboration in manufacturing","author":"Liu","year":"2019"},{"key":"10.1016\/j.rcim.2022.102449_b15","series-title":"31st AAAI Conference on Artificial Intelligence, AAAI 2017","first-page":"4263","article-title":"An end-to-end spatio-temporal attention model for human action recognition from skeleton data","author":"Song","year":"2017"},{"key":"10.1016\/j.rcim.2022.102449_b16","series-title":"Proceedings of the IEEE International Conference on Computer Vision, Vol. 2017-Octob","first-page":"2659","article-title":"A simple yet effective baseline for 3d human pose estimation","author":"Martinez","year":"2017"},{"issue":"1","key":"10.1016\/j.rcim.2022.102449_b17","doi-asserted-by":"crossref","first-page":"172","DOI":"10.1109\/TPAMI.2019.2929257","article-title":"OpenPose: Realtime multi-person 2D pose estimation using part affinity fields","volume":"43","author":"Cao","year":"2021","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.rcim.2022.102449_b18","series-title":"Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Vol. 2017-Janua","first-page":"1302","article-title":"Realtime multi-person 2D pose estimation using part affinity fields","author":"Cao","year":"2017"},{"key":"10.1016\/j.rcim.2022.102449_b19","series-title":"Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol. 2019-June","first-page":"11969","article-title":"PifPaf: Composite fields for human pose estimation","author":"Kreiss","year":"2019"},{"key":"10.1016\/j.rcim.2022.102449_b20","first-page":"1","article-title":"OpenPifPaf: Composite fields for semantic keypoint detection and spatio-temporal association","author":"Kreiss","year":"2021","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"10.1016\/j.rcim.2022.102449_b21","series-title":"Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol. 2019-June","first-page":"7745","article-title":"3D human pose estimation in video with temporal convolutions and semi-supervised training","author":"Pavllo","year":"2019"},{"issue":"1","key":"10.1016\/j.rcim.2022.102449_b22","doi-asserted-by":"crossref","first-page":"17","DOI":"10.1016\/j.cirp.2018.04.066","article-title":"Deep learning-based human motion recognition for predictive context-aware human-robot collaboration","volume":"67","author":"Wang","year":"2018","journal-title":"CIRP Annals"},{"key":"10.1016\/j.rcim.2022.102449_b23","doi-asserted-by":"crossref","DOI":"10.1016\/j.rcim.2021.102184","article-title":"Hybrid machine learning for human action recognition and prediction in assembly","volume":"72","author":"Zhang","year":"2021","journal-title":"Robot. Comput.-Integr. Manuf."},{"issue":"21","key":"10.1016\/j.rcim.2022.102449_b24","first-page":"1","article-title":"A mixed-perception approach for safe human\u2013robot collaboration in industrial automation","volume":"20","author":"Amin","year":"2020","journal-title":"Sensors (Switzerland)"},{"issue":"1","key":"10.1016\/j.rcim.2022.102449_b25","doi-asserted-by":"crossref","first-page":"9","DOI":"10.1016\/j.cirp.2020.04.077","article-title":"Recurrent neural network for motion trajectory prediction in human-robot collaborative assembly","volume":"69","author":"Zhang","year":"2020","journal-title":"CIRP Annals"},{"issue":"10","key":"10.1016\/j.rcim.2022.102449_b26","doi-asserted-by":"crossref","DOI":"10.3390\/s20102886","article-title":"Real-time human action recognition with a low-cost RGB camera and mobile robot platform","volume":"20","author":"Lee","year":"2020","journal-title":"Sensors (Switzerland)"},{"key":"10.1016\/j.rcim.2022.102449_b27","series-title":"2019 IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC)","first-page":"1","article-title":"Modeling of video projectors in OpenGL for implementing a spatial augmented reality teaching system for assembly operations","author":"Costa","year":"2019"},{"key":"10.1016\/j.rcim.2022.102449_b28","series-title":"Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition","first-page":"770","article-title":"Deep residual learning for image recognition","volume":"2016-Decem","author":"He","year":"2016"},{"key":"10.1016\/j.rcim.2022.102449_b29","series-title":"Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition","first-page":"6450","article-title":"A closer look at spatiotemporal convolutions for action recognition","author":"Tran","year":"2018"}],"container-title":["Robotics and Computer-Integrated Manufacturing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0736584522001314?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0736584522001314?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T17:01:28Z","timestamp":1760029288000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0736584522001314"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4]]},"references-count":29,"alternative-id":["S0736584522001314"],"URL":"https:\/\/doi.org\/10.1016\/j.rcim.2022.102449","relation":{},"ISSN":["0736-5845"],"issn-type":[{"value":"0736-5845","type":"print"}],"subject":[],"published":{"date-parts":[[2023,4]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Deep learning-based human action recognition to leverage context awareness in collaborative assembly","name":"articletitle","label":"Article Title"},{"value":"Robotics and Computer-Integrated Manufacturing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.rcim.2022.102449","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2022 Published by Elsevier Ltd.","name":"copyright","label":"Copyright"}],"article-number":"102449"}}