{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T22:33:07Z","timestamp":1761172387335,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":40,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032032805","type":"print"},{"value":"9783032032812","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,23]],"date-time":"2025-10-23T00:00:00Z","timestamp":1761177600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,23]],"date-time":"2025-10-23T00:00:00Z","timestamp":1761177600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-03281-2_28","type":"book-chapter","created":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T04:57:24Z","timestamp":1761109044000},"page":"409-418","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhancing Dynamic Scene Understanding in\u00a0Manual Assembly Processes"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-5473-8711","authenticated-orcid":false,"given":"Chenxi","family":"Guo","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,23]]},"reference":[{"issue":"4","key":"28_CR1","doi-asserted-by":"publisher","first-page":"3308","DOI":"10.1109\/LRA.2018.2852786","volume":"3","author":"H Ahn","year":"2018","unstructured":"Ahn, H., Choi, S., Kim, N., Cha, G., Oh, S.: Interactive text2pickup networks for natural language-based human-robot collaboration. IEEE Robot. Automation Lett. 3(4), 3308\u20133315 (2018)","journal-title":"IEEE Robot. Automation Lett."},{"key":"28_CR2","doi-asserted-by":"crossref","unstructured":"Br\u00fcggemann, D., Sakaridis, C., Br\u00f6dermann, T., Van\u00a0Gool, L.: Contrastive model adaptation for cross-condition robustness in semantic segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 11378\u201311387 (2023)","DOI":"10.1109\/ICCV51070.2023.01045"},{"key":"28_CR3","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? a new model and the kinetics dataset. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6299\u20136308 (2017)","DOI":"10.1109\/CVPR.2017.502"},{"key":"28_CR4","doi-asserted-by":"crossref","unstructured":"Castro, P., Kim, T.K.: Crt-6d: fast 6d object pose estimation with cascaded refinement transformers. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 5746\u20135755 (2023)","DOI":"10.1109\/WACV56688.2023.00570"},{"key":"28_CR5","unstructured":"Chan, C.M., et al.: Chateval: towards better LLM-based evaluators through multi-agent debate. arXiv preprint arXiv:2308.07201 (2023)"},{"key":"28_CR6","doi-asserted-by":"crossref","unstructured":"Diwan, T., Anirudh, G., Tembhurne, J.V.: Object detection using yolo: Challenges, architectural successors, datasets and applications. Multimedia Tools and Applications 82(6), 9243\u20139275 (2023)","DOI":"10.1007\/s11042-022-13644-y"},{"issue":"2","key":"28_CR7","doi-asserted-by":"publisher","first-page":"457","DOI":"10.4271\/2016-01-0337","volume":"9","author":"AM Djuric","year":"2016","unstructured":"Djuric, A.M., Urbanic, R., Rickli, J.: A framework for collaborative robot (cobot) integration in advanced manufacturing systems. SAE Int. J. Mater. Manufacturing 9(2), 457\u2013464 (2016)","journal-title":"SAE Int. J. Mater. Manufacturing"},{"key":"28_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2019.101888","volume":"63","author":"S D\u2019Avella","year":"2020","unstructured":"D\u2019Avella, S., Tripicchio, P., Avizzano, C.A.: A study on picking objects in cluttered environments: exploiting depth features for a custom low-cost universal jamming gripper. Robot. Comput. Integrated Manufacturing 63, 101888 (2020)","journal-title":"Robot. Comput. Integrated Manufacturing"},{"issue":"2","key":"28_CR9","doi-asserted-by":"publisher","first-page":"753","DOI":"10.1109\/LRA.2019.2893432","volume":"4","author":"Z Hu","year":"2019","unstructured":"Hu, Z., Pan, J., Fan, T., Yang, R., Manocha, D.: Safe navigation with human instructions in complex scenes. IEEE Robot. Autom. Lett. 4(2), 753\u2013760 (2019)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"28_CR10","doi-asserted-by":"crossref","unstructured":"Ji, J., Desai, R., Niebles, J.C.: Detecting human-object relationships in videos. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8106\u20138116 (2021)","DOI":"10.1109\/ICCV48922.2021.00800"},{"key":"28_CR11","doi-asserted-by":"publisher","first-page":"154","DOI":"10.3389\/frobt.2019.00154","volume":"6","author":"I Keller","year":"2020","unstructured":"Keller, I., Lohan, K.S.: On the illumination influence for object learning on robot companions. Front. Robot. AI 6, 154 (2020)","journal-title":"Front. Robot. AI"},{"key":"28_CR12","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1016\/j.jmsy.2020.02.010","volume":"55","author":"ZH Lai","year":"2020","unstructured":"Lai, Z.H., Tao, W., Leu, M.C., Yin, Z.: Smart augmented reality instructional system for mechanical assembly towards worker-centered intelligent manufacturing. J. Manuf. Syst. 55, 69\u201381 (2020)","journal-title":"J. Manuf. Syst."},{"key":"28_CR13","doi-asserted-by":"publisher","first-page":"307","DOI":"10.1016\/j.jmsy.2024.02.009","volume":"73","author":"W Li","year":"2024","unstructured":"Li, W., Xu, A., Wei, M., Zuo, W., Li, R.: Deep learning-based augmented reality work instruction assistance system for complex manual assembly. J. Manuf. Syst. 73, 307\u2013319 (2024)","journal-title":"J. Manuf. Syst."},{"key":"28_CR14","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1016\/j.cviu.2017.10.011","volume":"166","author":"Z Li","year":"2018","unstructured":"Li, Z., Gavrilyuk, K., Gavves, E., Jain, M., Snoek, C.G.: Videolstm convolves, attends and flows for action recognition. Comput. Vis. Image Underst. 166, 41\u201350 (2018)","journal-title":"Comput. Vis. Image Underst."},{"key":"28_CR15","doi-asserted-by":"crossref","unstructured":"Liu, M., Tang, S., Li, Y., Rehg, J.M.: Forecasting human-object interaction: joint prediction of motor attention and actions in first person video. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part I 16, pp. 704\u2013721. Springer (2020)","DOI":"10.1007\/978-3-030-58452-8_41"},{"key":"28_CR16","doi-asserted-by":"crossref","unstructured":"Lu, Y., Liu, Y.: Egocentric hand-object interaction detection. In: 2022 IEEE Smartworld, Ubiquitous Intelligence & Computing, Scalable Computing & Communications, Digital Twin, Privacy Computing, Metaverse, Autonomous & Trusted Vehicles (SmartWorld\/UIC\/ScalCom\/DigitalTwin\/PriComp\/Meta), pp. 25\u201332. IEEE (2022)","DOI":"10.1109\/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00031"},{"key":"28_CR17","doi-asserted-by":"publisher","first-page":"4703","DOI":"10.1109\/ACCESS.2023.3235368","volume":"11","author":"ASM Miah","year":"2023","unstructured":"Miah, A.S.M., Hasan, M.A.M., Shin, J.: Dynamic hand gesture recognition using multi-branch attention based graph and general deep learning model. IEEE Access 11, 4703\u20134716 (2023)","journal-title":"IEEE Access"},{"issue":"1","key":"28_CR18","doi-asserted-by":"publisher","first-page":"13","DOI":"10.3390\/computers12010013","volume":"12","author":"ASM Miah","year":"2023","unstructured":"Miah, A.S.M., Hasan, M.A.M., Shin, J., Okuyama, Y., Tomioka, Y.: Multistage spatial attention-based neural network for hand gesture recognition. Computers 12(1), 13 (2023)","journal-title":"Computers"},{"key":"28_CR19","unstructured":"Nauen, T.C., Palacio, S., Raue, F., Dengel, A.: Which transformer to favor: a comparative analysis of efficiency in vision transformers. arXiv preprint arXiv:2308.09372 (2023)"},{"key":"28_CR20","unstructured":"Oquab, M., et\u00a0al.: Dinov2: learning robust visual features without supervision. arXiv preprint arXiv:2304.07193 (2023)"},{"key":"28_CR21","doi-asserted-by":"publisher","first-page":"477","DOI":"10.1016\/j.jmsy.2023.05.008","volume":"68","author":"J Pang","year":"2023","unstructured":"Pang, J., Zheng, P., Li, S., Liu, S.: A verification-oriented and part-focused assembly monitoring system based on multi-layered digital twin. J. Manuf. Syst. 68, 477\u2013492 (2023)","journal-title":"J. Manuf. Syst."},{"issue":"1","key":"28_CR22","doi-asserted-by":"publisher","first-page":"2100099","DOI":"10.1002\/aisy.202100099","volume":"4","author":"V Patel","year":"2022","unstructured":"Patel, V., Chesmore, A., Legner, C.M., Pandey, S.: Trends in workplace wearable technologies and connected-worker solutions for next-generation occupational safety, health, and productivity. Adv. Intell. Syst. 4(1), 2100099 (2022)","journal-title":"Adv. Intell. Syst."},{"key":"28_CR23","doi-asserted-by":"crossref","unstructured":"Piergiovanni, A., Ryoo, M.S.: Representation flow for action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9945\u20139953 (2019)","DOI":"10.1109\/CVPR.2019.01018"},{"key":"28_CR24","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1016\/j.promfg.2020.04.093","volume":"45","author":"F Pilati","year":"2020","unstructured":"Pilati, F., Faccio, M., Gamberi, M., Regattieri, A.: Learning manual assembly through real-time motion capture for operator training with augmented reality. Procedia Manufacturing 45, 189\u2013195 (2020)","journal-title":"Procedia Manufacturing"},{"key":"28_CR25","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"28_CR26","unstructured":"Raji\u010d, F., Ke, L., Tai, Y.W., Tang, C.K., Danelljan, M., Yu, F.: Segment anything meets point tracking. arXiv preprint arXiv:2307.01197 (2023)"},{"key":"28_CR27","unstructured":"Ren, A., Veer, S., Majumdar, A.: Generalization guarantees for imitation learning. In: Conference on Robot Learning, pp. 1426\u20131442. PMLR (2021)"},{"issue":"1","key":"28_CR28","doi-asserted-by":"publisher","first-page":"2014191","DOI":"10.1080\/08839514.2021.2014191","volume":"36","author":"A Riedel","year":"2022","unstructured":"Riedel, A., Brehm, N., Pfeifroth, T.: Hand gesture recognition of methods-time measurement-1 motions in manual assembly tasks using graph convolutional networks. Appl. Artif. Intell. 36(1), 2014191 (2022)","journal-title":"Appl. Artif. Intell."},{"key":"28_CR29","unstructured":"Roy, D., Rajendiran, R., Fernando, B.: Interaction visual transformer for egocentric action anticipation. arXiv preprint arXiv:2211.14154 (2022)"},{"key":"28_CR30","doi-asserted-by":"crossref","unstructured":"Roy, D., Rajendiran, R., Fernando, B.: Interaction region visual transformer for egocentric action anticipation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 6740\u20136750 (2024)","DOI":"10.1109\/WACV57701.2024.00660"},{"key":"28_CR31","doi-asserted-by":"crossref","unstructured":"Shan, D., Geng, J., Shu, M., Fouhey, D.F.: Understanding human hands in contact at internet scale. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9869\u20139878 (2020)","DOI":"10.1109\/CVPR42600.2020.00989"},{"key":"28_CR32","doi-asserted-by":"crossref","unstructured":"Shao, Z., Yu, Z., Wang, M., Yu, J.: Prompting large language models with answer heuristics for knowledge-based visual question answering. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14974\u201314983 (2023)","DOI":"10.1109\/CVPR52729.2023.01438"},{"key":"28_CR33","doi-asserted-by":"crossref","unstructured":"Shugurov, I., Li, F., Busam, B., Ilic, S.: Osop: A multi-stage one shot object pose estimation framework. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6835\u20136844 (2022)","DOI":"10.1109\/CVPR52688.2022.00671"},{"key":"28_CR34","doi-asserted-by":"publisher","unstructured":"Singh, A.: Exploring language models: a comprehensive survey and analysis. In: Proceedings 2023 International Conference Research Methodologies in Knowledge Management, Artificial Intelligence and Telecommunication Engineering (RMKMATE), pp. 1\u20134. (2023). https:\/\/doi.org\/10.1109\/RMKMATE59243.2023.10369423","DOI":"10.1109\/RMKMATE59243.2023.10369423"},{"key":"28_CR35","doi-asserted-by":"crossref","unstructured":"Tan, H.L., et al.: Task-oriented multi-modal question answering for collaborative applications. In: 2020 IEEE International Conference on Image Processing (ICIP), pp. 1426\u20131430. IEEE (2020)","DOI":"10.1109\/ICIP40778.2020.9190659"},{"key":"28_CR36","doi-asserted-by":"publisher","first-page":"89699","DOI":"10.1109\/ACCESS.2021.3090471","volume":"9","author":"S Thermos","year":"2021","unstructured":"Thermos, S., Potamianos, G., Daras, P.: Joint object affordance reasoning and segmentation in RGB-d videos. IEEE Access 9, 89699\u201389713 (2021)","journal-title":"IEEE Access"},{"key":"28_CR37","doi-asserted-by":"crossref","unstructured":"Venkatesh, S.G., Biswas, A., Upadrashta, R., Srinivasan, V., Talukdar, P., Amrutur, B.: Spatial reasoning from natural language instructions for robot manipulation. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 11196\u201311202. IEEE (2021)","DOI":"10.1109\/ICRA48506.2021.9560895"},{"key":"28_CR38","doi-asserted-by":"crossref","unstructured":"Wang, G., Manhardt, F., Tombari, F., Ji, X.: Gdr-net: geometry-guided direct regression network for monocular 6d object pose estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16611\u201316621 (2021)","DOI":"10.1109\/CVPR46437.2021.01634"},{"key":"28_CR39","doi-asserted-by":"crossref","unstructured":"Xing, J., Bauersfeld, L., Song, Y., Xing, C., Scaramuzza, D.: Contrastive learning for enhancing robust scene transfer in vision-based agile flight. arXiv preprint arXiv:2309.09865 (2023)","DOI":"10.1109\/ICRA57147.2024.10610095"},{"issue":"1","key":"28_CR40","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1109\/TSMC.2022.3166397","volume":"53","author":"L Zhou","year":"2022","unstructured":"Zhou, L., Zhang, L., Konz, N.: Computer vision techniques in manufacturing. IEEE Trans. Syst. Man Cybern. Syst. 53(1), 105\u2013117 (2022)","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."}],"container-title":["Lecture Notes in Computer Science","Architecture of Computing Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-03281-2_28","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T04:57:33Z","timestamp":1761109053000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-03281-2_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,23]]},"ISBN":["9783032032805","9783032032812"],"references-count":40,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-03281-2_28","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,23]]},"assertion":[{"value":"23 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ARCS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Architecture of Computing Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kiel","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Germany","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 April 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 April 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"38","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"arcs2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/arcs-conference.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}