{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T02:54:57Z","timestamp":1743044097594,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":31,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819607761"},{"type":"electronic","value":"9789819607778"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-0777-8_18","type":"book-chapter","created":{"date-parts":[[2025,2,4]],"date-time":"2025-02-04T13:57:02Z","timestamp":1738677422000},"page":"247-262","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DSGCN: Dual-Stream Graph Convolutional Network for Skeleton-Based Action Recognition Under Noise Interference"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-6276-7706","authenticated-orcid":false,"given":"Yanxin","family":"Cui","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0007-1937-1185","authenticated-orcid":false,"given":"Yufeng","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-1405-0057","authenticated-orcid":false,"given":"Weiming","family":"Fan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5099-482X","authenticated-orcid":false,"given":"Xuna","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1314-2481","authenticated-orcid":false,"given":"Jiahui","family":"Yu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9524-7609","authenticated-orcid":false,"given":"Zhaojie","family":"Ju","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,5]]},"reference":[{"issue":"3","key":"18_CR1","first-page":"3200","volume":"45","author":"Z Sun","year":"2022","unstructured":"Sun, Z., Ke, Q., Rahmani, H., Bennamoun, M., Wang, G., Liu, J.: Human action recognition from various data modalities: a review. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3200\u20133225 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"18_CR2","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? a new model and the kinetics dataset. In:\u00a0proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6299\u20136308 (2017)","DOI":"10.1109\/CVPR.2017.502"},{"key":"18_CR3","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos.\u00a0Adv. Neural Inf. Process. Syst.\u00a027 (2014)"},{"key":"18_CR4","unstructured":"Xiao, F., Lee, Y.J., Grauman, K., Malik, J., Feichtenhofer, C.: Audiovisual slowfast networks for video recognition (2020).\u00a0arXiv preprint arXiv:2001.08740"},{"key":"18_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.neucom.2020.10.037","volume":"423","author":"H Wang","year":"2021","unstructured":"Wang, H., Yu, B., Xia, K., Li, J., Zuo, X.: Skeleton edge motion networks for human action recognition. Neurocomputing 423, 1\u201312 (2021)","journal-title":"Neurocomputing"},{"key":"18_CR6","unstructured":"Sharma, S., Kiros, R., Salakhutdinov, R.: Action recognition using visual attention (2015).\u00a0arXiv preprint arXiv:1511.04119"},{"key":"18_CR7","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3d convolutional networks. In:\u00a0Proceedings of the IEEE International Conference on Computer Vision, pp. 4489\u2013449 (2015)","DOI":"10.1109\/ICCV.2015.510"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Duan, H., Zhao, Y., Chen, K., Lin, D., Dai, B.: Revisiting skeleton-based action recognition. In:\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2969\u20132978 (2022)","DOI":"10.1109\/CVPR52688.2022.00298"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Ling, Y., Ma, Z., Xie, B., Zhang, Q., Weng, X.: SA-BiGCN: bi-stream graph convolution networks with spatial attentions for the eye contact detection in the wild.\u00a0IEEE Trans. Intell. Transp. Syst. (2023)","DOI":"10.1109\/TITS.2023.3310194"},{"key":"18_CR10","doi-asserted-by":"crossref","unstructured":"Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In:\u00a0Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, no. 1 (2018)","DOI":"10.1609\/aaai.v32i1.12328"},{"issue":"5","key":"18_CR11","doi-asserted-by":"publisher","first-page":"1915","DOI":"10.1109\/TCSVT.2020.3015051","volume":"31","author":"YF Song","year":"2020","unstructured":"Song, Y.F., Zhang, Z., Shan, C., Wang, L.: Richly activated graph convolutional network for robust skeleton-based action recognition. IEEE Trans. Circuits Syst. Video Technol. 31(5), 1915\u20131925 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"18_CR12","doi-asserted-by":"crossref","unstructured":"Shahroudy, A., Liu, J., Ng, T.T., Wang, G.: Ntu rgb+ d: a large scale dataset for 3d human activity analysis. In:\u00a0Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1010\u20131019 (2016)","DOI":"10.1109\/CVPR.2016.115"},{"issue":"10","key":"18_CR13","doi-asserted-by":"publisher","first-page":"2684","DOI":"10.1109\/TPAMI.2019.2916873","volume":"42","author":"J Liu","year":"2019","unstructured":"Liu, J., Shahroudy, A., Perez, M., Wang, G., Duan, L.Y., Kot, A.C.: Ntu rgb+ d 120: a large-scale benchmark for 3d human activity understanding. IEEE Trans. Pattern Anal. Mach. Intell. 42(10), 2684\u20132701 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"18_CR14","doi-asserted-by":"crossref","unstructured":"Liu, J., Shahroudy, A., Xu, D., Wang, G.: Spatio-temporal lstm with trust gates for 3d human action recognition. In:\u00a0Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part III 14, pp. 816\u2013833. Springer International Publishing (2016)","DOI":"10.1007\/978-3-319-46487-9_50"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"Li, C., Zhong, Q., Xie, D., Pu, S.: Co-occurrence feature learning from skeleton data for action recognition and detection with hierarchical aggregation (2018).\u00a0arXiv preprint arXiv:1804.06055","DOI":"10.24963\/ijcai.2018\/109"},{"key":"18_CR16","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., Lu, H.: Skeleton-based action recognition with directed graph neural networks. In:\u00a0Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 7912\u20137921 (2019)","DOI":"10.1109\/CVPR.2019.00810"},{"key":"18_CR17","unstructured":"Yu, J., Xu, Y., Chen, H., Ju, Z.: Versatile graph neural networks toward intuitive human activity understanding.\u00a0IEEE Trans. Neural Netw. Learn. Syst. (2022)"},{"key":"18_CR18","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., Lu, H.: Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In:\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12026\u201312035 (2019)","DOI":"10.1109\/CVPR.2019.01230"},{"key":"18_CR19","doi-asserted-by":"crossref","unstructured":"Li, M., Chen, S., Chen, X., Zhang, Y., Wang, Y., Tian, Q.: Actional-structural graph convolutional networks for skeleton-based action recognition. In:\u00a0Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 3595\u20133603 (2019)","DOI":"10.1109\/CVPR.2019.00371"},{"key":"18_CR20","doi-asserted-by":"crossref","unstructured":"Yu, J., Cheng, X., Chen, H., Xu, Y.: Pose-guided robust action recognition for outdoor internet of things.\u00a0IEEE Trans. Consum. Electr. (2024)","DOI":"10.1109\/TCE.2024.3384974"},{"issue":"19","key":"18_CR21","doi-asserted-by":"publisher","first-page":"7117","DOI":"10.3390\/s22197117","volume":"22","author":"M Nan","year":"2022","unstructured":"Nan, M., Florea, A.M.: Fast temporal graph convolutional model for skeleton-based action recognition. Sensors 22(19), 7117 (2022)","journal-title":"Sensors"},{"key":"18_CR22","doi-asserted-by":"crossref","unstructured":"Cheng, K., Zhang, Y., He, X., Chen, W., Cheng, J., Lu, H.: Skeleton-based action recognition with shift graph convolutional network. In:\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 183\u2013192 (2020)","DOI":"10.1109\/CVPR42600.2020.00026"},{"key":"18_CR23","doi-asserted-by":"crossref","unstructured":"Trivedi, N., Sarvadevabhatla, R.K.: Psumnet: unified modality part streams are all you need for efficient pose-based action recognition. In:\u00a0European Conference on Computer Vision, pp. 211\u201322). Cham: Springer Nature Switzerland (2022)","DOI":"10.1007\/978-3-031-25072-9_14"},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"Demisse, G.G., Papadopoulos, K., Aouada, D., Ottersten, B.: Pose encoding for robust skeleton-based action recognition. In:\u00a0Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 188\u2013194 (2018)","DOI":"10.1109\/CVPRW.2018.00056"},{"key":"18_CR25","doi-asserted-by":"crossref","unstructured":"Weinland, D., \u00d6zuysal, M., Fua, P.: Making action recognition robust to occlusions and viewpoint changes. In:\u00a0Computer Vision\u2013ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, September 5\u201311, 2010, Proceedings, Part III 11, pp. 635\u2013648. Springer Berlin Heidelberg (2010)","DOI":"10.1007\/978-3-642-15558-1_46"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Chen, Y., Fan, W., Zheng, W., Huang, R., Yu, J.: Predicting bird's-eye-view semantic representations using correlated context learning.\u00a0IEEE Robot. Autom. Lett. (2024)","DOI":"10.1109\/LRA.2024.3384078"},{"key":"18_CR27","unstructured":"Yoon, Y., Yu, J., Jeon, M.: Predictively encoded graph convolutional network for noise-robust skeleton-based action recognition.\u00a0Appl. Intell. 1\u201315 (2022)"},{"key":"18_CR28","unstructured":"Oord, A.V.D., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding (2018).\u00a0arXiv preprint arXiv:1807.03748"},{"key":"18_CR29","doi-asserted-by":"crossref","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In:\u00a0Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2921\u20132929 (2016)","DOI":"10.1109\/CVPR.2016.319"},{"issue":"4","key":"18_CR30","doi-asserted-by":"publisher","first-page":"1654","DOI":"10.1109\/TCDS.2021.3131253","volume":"14","author":"J Yu","year":"2021","unstructured":"Yu, J., Gao, H., Chen, Y., Zhou, D., Liu, J., Ju, Z.: Adaptive spatiotemporal representation learning for skeleton-based human action recognition. IEEE Trans. Cogn. Dev. Syst. 14(4), 1654\u20131665 (2021)","journal-title":"IEEE Trans. Cogn. Dev. Syst."},{"issue":"4","key":"18_CR31","doi-asserted-by":"publisher","first-page":"784","DOI":"10.1109\/THMS.2022.3144951","volume":"52","author":"J Yu","year":"2022","unstructured":"Yu, J., Gao, H., Chen, Y., Zhou, D., Liu, J., Ju, Z.: Deep object detector with attentional spatiotemporal LSTM for space human\u2013robot interaction. IEEE Trans. Hum. Mach. Syst. 52(4), 784\u2013793 (2022)","journal-title":"IEEE Trans. Hum. Mach. Syst."}],"container-title":["Lecture Notes in Computer Science","Intelligent Robotics and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-0777-8_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,4]],"date-time":"2025-02-04T13:57:29Z","timestamp":1738677449000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-0777-8_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819607761","9789819607778"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-0777-8_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"5 February 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIRA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Robotics and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xi'an","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31 July 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 August 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icira2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.icira2024.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}