{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T01:21:44Z","timestamp":1775611304144,"version":"3.50.1"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"7","license":[{"start":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T00:00:00Z","timestamp":1721606400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T00:00:00Z","timestamp":1721606400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the Open Project of Henan Key Laboratory of Intelligent Manufacturing of Mechanical Equipment, Zhengzhou University of Light Industry","award":["No. IM202303"],"award-info":[{"award-number":["No. IM202303"]}]},{"DOI":"10.13039\/100014718","name":"Innovative Research Group Project of the National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No. 52005099"],"award-info":[{"award-number":["No. 52005099"]}],"id":[{"id":"10.13039\/100014718","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shanghai Frontier Science Research Center for Modern Textiles"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Intell Manuf"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s10845-024-02462-8","type":"journal-article","created":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T06:01:58Z","timestamp":1721628118000},"page":"4593-4612","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Dynamic scenario-enhanced diverse human motion prediction network for proactive human\u2013robot collaboration in customized assembly tasks"],"prefix":"10.1007","volume":"36","author":[{"given":"Pengfei","family":"Ding","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6215-0237","authenticated-orcid":false,"given":"Jie","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Pai","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Peng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Fei","sequence":"additional","affiliation":[]},{"given":"Ziqi","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,22]]},"reference":[{"key":"2462_CR1","doi-asserted-by":"crossref","unstructured":"Aksan, E., Kaufmann, M., & Hilliges, O. (2019). Structured prediction helps 3d human motion modelling. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 7144\u20137153). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content_ICCV_2019\/html\/Aksan_Structured_Prediction_Helps_3D_Human_Motion_Modelling_ICCV_2019_paper.html","DOI":"10.1109\/ICCV.2019.00724"},{"key":"2462_CR2","doi-asserted-by":"crossref","unstructured":"Aliakbarian, S., Saleh, F. S., Salzmann, M., Petersson, L., & Gould, S. (2020). A stochastic conditioning scheme for diverse human motion prediction. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 5223\u20135232). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content_CVPR_2020\/html\/Aliakbarian_A_Stochastic_Conditioning_Scheme_for_Diverse_Human_Motion_Prediction_CVPR_2020_paper.html","DOI":"10.1109\/CVPR42600.2020.00527"},{"key":"2462_CR3","doi-asserted-by":"crossref","unstructured":"Aliakbarian, S., Saleh, F., Petersson, L., Gould, S., & Salzmann, M. (2021). Contextually plausible and diverse 3d human motion prediction. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 11333\u201311342). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content\/ICCV2021\/html\/Aliakbarian_Contextually_Plausible_and_Diverse_3D_Human_Motion_Prediction_ICCV_2021_paper.html","DOI":"10.1109\/ICCV48922.2021.01114"},{"issue":"7","key":"2462_CR4","doi-asserted-by":"publisher","first-page":"1635","DOI":"10.1007\/s10845-018-1411-1","volume":"31","author":"T Baenziger","year":"2020","unstructured":"Baenziger, T., Kunz, A., & Wegener, K. (2020). Optimizing human-robot task allocation using a simulation tool based on standardized work descriptions. Journal of Intelligent Manufacuring, 31(7), 1635\u20131648. https:\/\/doi.org\/10.1007\/s10845-018-1411-1","journal-title":"Journal of Intelligent Manufacuring"},{"key":"2462_CR5","doi-asserted-by":"crossref","unstructured":"Barsoum, E., Kender, J., & Liu, Z. (2018). Hp-gan: Probabilistic 3d human motion prediction via gan. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops (pp. 1418\u20131427). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content_cvpr_2018_workshops\/w29\/html\/Barsoum_HP-GAN_Probabilistic_3D_CVPR_2018_paper.html","DOI":"10.1109\/CVPRW.2018.00191"},{"issue":"7","key":"2462_CR6","doi-asserted-by":"publisher","first-page":"3053","DOI":"10.1007\/s10845-022-01989-y","volume":"34","author":"J Cacace","year":"2023","unstructured":"Cacace, J., Caccavale, R., Finzi, A., & Grieco, R. (2023). Combining human guidance and structured task execution during physical human-robot collaboration. Journal of Intelligent Manufacuring, 34(7), 3053\u20133067. https:\/\/doi.org\/10.1007\/s10845-022-01989-y","journal-title":"Journal of Intelligent Manufacuring"},{"key":"2462_CR7","doi-asserted-by":"publisher","first-page":"226","DOI":"10.1007\/978-3-030-58571-6_14","volume-title":"Computer vision\u2014ECCV 2020","author":"Y Cai","year":"2020","unstructured":"Cai, Y., Huang, L., Wang, Y., Cham, T. J., Cai, J., Yuan, J., et al. (2020). Learning progressive joint propagation for human motion prediction. In A. Vedaldi, H. Bischof, T. Brox, & J.-M. Frahm (Eds.), Computer vision\u2014ECCV 2020 (Vol. 12352, pp. 226\u2013242). Springer International Publishing. https:\/\/doi.org\/10.1007\/978-3-030-58571-6_14"},{"key":"2462_CR8","doi-asserted-by":"crossref","unstructured":"Cui, Q., Sun, H., & Yang, F. (2020). Learning dynamic relationships for 3d human motion prediction. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 6519\u20136527). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content_CVPR_2020\/html\/Cui_Learning_Dynamic_Relationships_for_3D_Human_Motion_Prediction_CVPR_2020_paper.html","DOI":"10.1109\/CVPR42600.2020.00655"},{"issue":"2","key":"2462_CR9","doi-asserted-by":"publisher","first-page":"696","DOI":"10.1109\/TCDS.2021.3064280","volume":"14","author":"L Dai","year":"2021","unstructured":"Dai, L., Liu, J., Ju, Z., & Gao, Y. (2021). Attention-mechanism-based real-time gaze tracking in natural scenes with residual blocks. IEEE Transactions on Cognitive and Developmental Systems, 14(2), 696\u2013707.","journal-title":"IEEE Transactions on Cognitive and Developmental Systems"},{"key":"2462_CR10","doi-asserted-by":"crossref","unstructured":"Dang, L., Nie, Y., Long, C., Zhang, Q., & Li, G. (2021). Msr-gcn: Multi-scale residual graph convolution networks for human motion prediction. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 11467\u201311476). Retrieved January 24, 2024, from https:\/\/openaccess.thecvf.com\/content\/ICCV2021\/html\/Dang_MSR-GCN_Multi-Scale_Residual_Graph_Convolution_Networks_for_Human_Motion_Prediction_ICCV_2021_paper.html?ref=https:\/\/githubhelp.com","DOI":"10.1109\/ICCV48922.2021.01127"},{"key":"2462_CR11","doi-asserted-by":"publisher","unstructured":"Dang, L., Nie, Y., Long, C., Zhang, Q., & Li, G. (2022). Diverse human motion prediction via gumbel-softmax sampling from an auxiliary space. In Proceedings of the 30th ACM international conference on multimedia (pp. 5162\u20135171). https:\/\/doi.org\/10.1145\/3503161.3547956","DOI":"10.1145\/3503161.3547956"},{"key":"2462_CR12","doi-asserted-by":"publisher","unstructured":"Dong, X., Long, C., Xu, W., & Xiao, C. (2021). Dual graph convolutional networks with transformer and curriculum learning for image captioning. In Proceedings of the 29th ACM international conference on multimedia (pp. 2615\u20132624). Presented at the MM \u201921: ACM Multimedia Conference, Virtual Event China: ACM. https:\/\/doi.org\/10.1145\/3474085.3475439","DOI":"10.1145\/3474085.3475439"},{"key":"2462_CR13","doi-asserted-by":"crossref","unstructured":"Duan, J., Wang, L., Long, C., Zhou, S., Zheng, F., Shi, L., & Hua, G. (2022). Complementary attention gated network for pedestrian trajectory prediction. In Proceedings of the AAAI conference on artificial intelligence (Vol. 36, pp. 542\u2013550). Retrieved January 24, 2024, from https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/19933","DOI":"10.1609\/aaai.v36i1.19933"},{"issue":"1","key":"2462_CR14","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1007\/s10845-022-01953-w","volume":"34","author":"M Faccio","year":"2023","unstructured":"Faccio, M., Granata, I., Menini, A., Milanese, M., Rossato, C., Bottin, M., Minto, R., Pluchino, P., Gamberini, L., Boschetti, G., & Rosati, G. (2023). Human factors in cobot era: A review of modern production systems features. Journal of Intelligent Manufacuring, 34(1), 85\u2013106. https:\/\/doi.org\/10.1007\/s10845-022-01953-w","journal-title":"Journal of Intelligent Manufacuring"},{"key":"2462_CR15","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2021.102304","volume":"75","author":"J Fan","year":"2022","unstructured":"Fan, J., Zheng, P., & Li, S. (2022). Vision-based holistic scene understanding towards proactive human\u2013robot collaboration. Robotics and Computer-Integrated Manufacturing, 75, 102304.","journal-title":"Robotics and Computer-Integrated Manufacturing"},{"issue":"12","key":"2462_CR16","doi-asserted-by":"publisher","first-page":"9663","DOI":"10.1109\/TIE.2019.2898624","volume":"66","author":"Q Gao","year":"2019","unstructured":"Gao, Q., Liu, J., Ju, Z., & Zhang, X. (2019). Dual-hand detection for human\u2013robot interaction by a parallel network based on hand detection and body pose estimation. IEEE Transactions on Industrial Electronics, 66(12), 9663\u20139672.","journal-title":"IEEE Transactions on Industrial Electronics"},{"issue":"11","key":"2462_CR17","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., et al. (2020). Generative adversarial networks. Communications of the ACM, 63(11), 139\u2013144. https:\/\/doi.org\/10.1145\/3422622","journal-title":"Communications of the ACM"},{"key":"2462_CR18","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770\u2013778). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content_cvpr_2016\/html\/He_Deep_Residual_Learning_CVPR_2016_paper.html","DOI":"10.1109\/CVPR.2016.90"},{"key":"2462_CR19","unstructured":"Kingma, D. P., & Welling, M. (2022). Auto-encoding variational bayes. Preprint retrieved from http:\/\/arxiv.org\/abs\/1312.6114"},{"key":"2462_CR20","doi-asserted-by":"crossref","unstructured":"Kundu, J. N., Gor, M., & Babu, R. V. (2019). Bihmp-gan: Bidirectional 3d human motion prediction gan. In Proceedings of the AAAI conference on artificial intelligence (Vol. 33, pp. 8553\u20138560). Retrieved January 24, 2024 from https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/4874","DOI":"10.1609\/aaai.v33i01.33018553"},{"key":"2462_CR21","doi-asserted-by":"crossref","unstructured":"Li, M., Chen, S., Zhao, Y., Zhang, Y., Wang, Y., & Tian, Q. (2020). Dynamic multiscale graph neural networks for 3d skeleton based human motion prediction. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 214\u2013223). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content_CVPR_2020\/html\/Li_Dynamic_Multiscale_Graph_Neural_Networks_for_3D_Skeleton_Based_Human_CVPR_2020_paper.html","DOI":"10.1109\/CVPR42600.2020.00029"},{"key":"2462_CR22","doi-asserted-by":"publisher","first-page":"547","DOI":"10.1016\/j.jmsy.2021.07.017","volume":"60","author":"S Li","year":"2021","unstructured":"Li, S., Wang, R., Zheng, P., & Wang, L. (2021). Towards proactive human\u2013robot collaboration: A foreseeable cognitive manufacturing paradigm. Journal of Manufacturing Systems, 60, 547\u2013552.","journal-title":"Journal of Manufacturing Systems"},{"key":"2462_CR23","doi-asserted-by":"crossref","unstructured":"Lin, T. Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., & Belongie, S. (2017). Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2117\u20132125). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content_cvpr_2017\/html\/Lin_Feature_Pyramid_Networks_CVPR_2017_paper.html","DOI":"10.1109\/CVPR.2017.106"},{"key":"2462_CR24","doi-asserted-by":"publisher","first-page":"317","DOI":"10.1016\/j.jmsy.2021.12.001","volume":"62","author":"CH Lin","year":"2022","unstructured":"Lin, C. H., Wang, K. J., Tadesse, A. A., & Woldegiorgis, B. H. (2022). Human-robot collaboration empowered by hidden semi-Markov model for operator behaviour prediction in a smart assembly system. Journal of Manufacturing Systems, 62, 317\u2013333.","journal-title":"Journal of Manufacturing Systems"},{"key":"2462_CR25","doi-asserted-by":"publisher","first-page":"287","DOI":"10.1016\/j.jmsy.2017.04.009","volume":"44","author":"H Liu","year":"2017","unstructured":"Liu, H., & Wang, L. (2017). Human motion prediction for human-robot collaboration. Journal of Manufacturing Systems, 44, 287\u2013294.","journal-title":"Journal of Manufacturing Systems"},{"issue":"11\u201312","key":"2462_CR26","doi-asserted-by":"publisher","first-page":"3533","DOI":"10.1007\/s00170-021-07265-2","volume":"115","author":"Z Liu","year":"2021","unstructured":"Liu, Z., Liu, Q., Wang, L., Xu, W., & Zhou, Z. (2021). Task-level decision-making for dynamic and stochastic human-robot collaboration based on dual agents deep reinforcement learning. The International Journal of Advanced Manufacturing Technology, 115(11\u201312), 3533\u20133552.","journal-title":"The International Journal of Advanced Manufacturing Technology"},{"issue":"4","key":"2462_CR27","doi-asserted-by":"publisher","first-page":"2056","DOI":"10.1109\/TMECH.2023.3275316","volume":"28","author":"W Liu","year":"2023","unstructured":"Liu, W., Liang, X., & Zheng, M. (2023a). Task-constrained motion planning considering uncertainty-informed human motion prediction for human-robot collaborative disassembly. IEEE-ASME Transactions on Mechatronics, 28(4), 2056\u20132063. https:\/\/doi.org\/10.1109\/TMECH.2023.3275316","journal-title":"IEEE-ASME Transactions on Mechatronics"},{"key":"2462_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.aei.2023.102129","volume":"58","author":"Z Liu","year":"2023","unstructured":"Liu, Z., Liu, Q., Xu, W., Wang, L., & Ji, Z. (2023b). Adaptive real-time similar repetitive manual procedure prediction and robotic procedure generation for human-robot collaboration. Advanced Engineering Informatics, 58, 102129. https:\/\/doi.org\/10.1016\/j.aei.2023.102129","journal-title":"Advanced Engineering Informatics"},{"key":"2462_CR29","doi-asserted-by":"crossref","unstructured":"Mao, W., Liu, M., Salzmann, M., & Li, H. (2019). Learning trajectory dependencies for human motion prediction. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 9489\u20139497). Retrieved January 24, 2024, form http:\/\/openaccess.thecvf.com\/content_ICCV_2019\/html\/Mao_Learning_Trajectory_Dependencies_for_Human_Motion_Prediction_ICCV_2019_paper.html","DOI":"10.1109\/ICCV.2019.00958"},{"key":"2462_CR30","doi-asserted-by":"crossref","unstructured":"Mao, W., Liu, M., & Salzmann, M. (2021). Generating smooth pose sequences for diverse human motion prediction. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 13309\u201313318). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content\/ICCV2021\/html\/Mao_Generating_Smooth_Pose_Sequences_for_Diverse_Human_Motion_Prediction_ICCV_2021_paper.html","DOI":"10.1109\/ICCV48922.2021.01306"},{"issue":"12","key":"2462_CR31","doi-asserted-by":"publisher","first-page":"9372","DOI":"10.1109\/TIE.2019.2891449","volume":"66","author":"MA Simao","year":"2019","unstructured":"Simao, M. A., Gibaru, O., & Neto, P. (2019). Online recognition of incomplete gesture data to interface collaborative robots. IEEE Transactions on Industrial Electronics, 66(12), 9372\u20139382.","journal-title":"IEEE Transactions on Industrial Electronics"},{"key":"2462_CR32","unstructured":"Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. Preprint retrieved form http:\/\/arxiv.org\/abs\/1409.1556"},{"issue":"2","key":"2462_CR33","doi-asserted-by":"publisher","first-page":"4702","DOI":"10.1109\/LRA.2022.3151614","volume":"7","author":"W Wan","year":"2022","unstructured":"Wan, W., Yang, L., Liu, L., Zhang, Z., Jia, R., Choi, Y. K., et al. (2022). Learn to predict how humans manipulate large-sized objects from interactive motions. IEEE Robotics and Automation Letters, 7(2), 4702\u20134709.","journal-title":"IEEE Robotics and Automation Letters"},{"key":"2462_CR34","doi-asserted-by":"publisher","first-page":"373","DOI":"10.1016\/j.jmsy.2020.06.020","volume":"56","author":"B Wang","year":"2020","unstructured":"Wang, B., Hu, S. J., Sun, L., & Freiheit, T. (2020a). Intelligent welding system technologies: State-of-the-art review and perspectives. Journal of Manufacturing Systems, 56, 373\u2013391.","journal-title":"Journal of Manufacturing Systems"},{"issue":"2","key":"2462_CR35","doi-asserted-by":"publisher","first-page":"701","DOI":"10.1016\/j.cirp.2019.05.002","volume":"68","author":"L Wang","year":"2019","unstructured":"Wang, L., Gao, R., V\u00e1ncza, J., Kr\u00fcger, J., Wang, X. V., Makris, S., & Chryssolouris, G. (2019). Symbiotic human-robot collaborative assembly. CIRP Annals, 68(2), 701\u2013726.","journal-title":"CIRP Annals"},{"issue":"1","key":"2462_CR36","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1016\/j.cirp.2018.04.066","volume":"67","author":"P Wang","year":"2018","unstructured":"Wang, P., Liu, H., Wang, L., & Gao, R. X. (2018). Deep learning-based human motion recognition for predictive context-aware human-robot collaboration. CIRP Annals, 67(1), 17\u201320.","journal-title":"CIRP Annals"},{"key":"2462_CR37","first-page":"1","volume":"72","author":"Q Wang","year":"2023","unstructured":"Wang, Q., Liu, D., Liu, Z., Xu, J., Liu, H., & Tan, J. (2023). A geometry-enhanced 6D pose estimation network with incomplete shape recovery for industrial parts. IEEE Transactions on Instrumentation and Measurement, 72, 1\u201311.","journal-title":"IEEE Transactions on Instrumentation and Measurement"},{"issue":"1","key":"2462_CR38","doi-asserted-by":"publisher","first-page":"425","DOI":"10.1016\/j.cirp.2020.03.014","volume":"69","author":"XV Wang","year":"2020","unstructured":"Wang, X. V., Wang, L., Lei, M., & Zhao, Y. (2020b). Closed-loop augmented reality towards accurate human-robot collaboration. CIRP Annals, 69(1), 425\u2013428.","journal-title":"CIRP Annals"},{"key":"2462_CR39","doi-asserted-by":"publisher","DOI":"10.1016\/j.autcon.2022.104497","volume":"142","author":"X Xia","year":"2022","unstructured":"Xia, X., Zhou, T., Du, J., & Li, N. (2022). Human motion prediction for intelligent construction: A review. Automation in Construction, 142, 104497.","journal-title":"Automation in Construction"},{"key":"2462_CR40","doi-asserted-by":"crossref","unstructured":"Yan, X., Rastogi, A., Villegas, R., Sunkavalli, K., Shechtman, E., Hadap, S., et al. (2018). Mt-vae: Learning motion transformations to generate multimodal human dynamics. In Proceedings of the European conference on computer vision (ECCV) (pp. 265\u2013281). Retrieved January 24, 2024, from http:\/\/openaccess.thecvf.com\/content_ECCV_2018\/html\/Xinchen_Yan_Generating_Multimodal_Human_ECCV_2018_paper.html","DOI":"10.1007\/978-3-030-01228-1_17"},{"key":"2462_CR41","doi-asserted-by":"crossref","unstructured":"Yuan, W., Khot, T., Held, D., Mertz, C., & Hebert, M. (2018). Pcn: Point completion network. In 2018 international conference on 3D vision (3DV) (pp. 728\u2013737). IEEE. Retrieved January 24, 2024, from https:\/\/ieeexplore.ieee.org\/abstract\/document\/8491026\/","DOI":"10.1109\/3DV.2018.00088"},{"key":"2462_CR42","doi-asserted-by":"publisher","first-page":"346","DOI":"10.1007\/978-3-030-58545-7_20","volume-title":"Computer vision\u2014ECCV 2020","author":"Y Yuan","year":"2020","unstructured":"Yuan, Y., & Kitani, K. (2020). DLow: Diversifying latent flows for diverse human motion prediction. In A. Vedaldi, H. Bischof, T. Brox, & J. M. Frahm (Eds.), Computer vision\u2014ECCV 2020 (Vol. 12354, pp. 346\u2013364). Springer International Publishing. https:\/\/doi.org\/10.1007\/978-3-030-58545-7_20"},{"issue":"1","key":"2462_CR43","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1016\/j.cirp.2020.04.077","volume":"69","author":"J Zhang","year":"2020","unstructured":"Zhang, J., Liu, H., Chang, Q., Wang, L., & Gao, R. X. (2020). Recurrent neural network for motion trajectory prediction in human-robot collaborative assembly. CIRP Annals, 69(1), 9\u201312.","journal-title":"CIRP Annals"},{"key":"2462_CR44","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2021.102184","volume":"72","author":"J Zhang","year":"2021","unstructured":"Zhang, J., Wang, P., & Gao, R. X. (2021). Hybrid machine learning for human action recognition and prediction in assembly. Robotics and Computer-Integrated Manufacturing, 72, 102184.","journal-title":"Robotics and Computer-Integrated Manufacturing"},{"issue":"2","key":"2462_CR45","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1631\/FITEE.1700053","volume":"18","author":"N Zheng","year":"2017","unstructured":"Zheng, N., Liu, Z., Ren, P., Ma, Y., Chen, S., Yu, S., et al. (2017). Hybrid-augmented intelligence: Collaboration and cognition. Frontiers of Information Technology & Electronic Engineering, 18(2), 153\u2013179. https:\/\/doi.org\/10.1631\/FITEE.1700053","journal-title":"Frontiers of Information Technology & Electronic Engineering"}],"container-title":["Journal of Intelligent Manufacturing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10845-024-02462-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10845-024-02462-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10845-024-02462-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,20]],"date-time":"2025-09-20T08:06:32Z","timestamp":1758355592000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10845-024-02462-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,22]]},"references-count":45,"journal-issue":{"issue":"7","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["2462"],"URL":"https:\/\/doi.org\/10.1007\/s10845-024-02462-8","relation":{},"ISSN":["0956-5515","1572-8145"],"issn-type":[{"value":"0956-5515","type":"print"},{"value":"1572-8145","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,22]]},"assertion":[{"value":"24 January 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 July 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}