{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T09:24:22Z","timestamp":1748683462904,"version":"3.37.3"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"7","license":[{"start":{"date-parts":[[2023,1,18]],"date-time":"2023-01-18T00:00:00Z","timestamp":1674000000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,18]],"date-time":"2023-01-18T00:00:00Z","timestamp":1674000000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62276061"],"award-info":[{"award-number":["62276061"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int. J. Mach. Learn. &amp; Cyber."],"published-print":{"date-parts":[[2023,7]]},"DOI":"10.1007\/s13042-023-01774-0","type":"journal-article","created":{"date-parts":[[2023,1,18]],"date-time":"2023-01-18T07:02:50Z","timestamp":1674025370000},"page":"2439-2453","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Multi-receptive field spatiotemporal network for action recognition"],"prefix":"10.1007","volume":"14","author":[{"given":"Mu","family":"Nie","sequence":"first","affiliation":[]},{"given":"Sen","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Zhenhua","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Baochang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Huimin","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Wankou","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,1,18]]},"reference":[{"issue":"8","key":"1774_CR1","first-page":"2752","volume":"43","author":"DC Luvizon","year":"2021","unstructured":"Luvizon DC, Picard D, Tabia H (2021) Multi-task deep learning for real-time 3d human pose estimation and action recognition. IEEE Trans Pattern Anal Mach Intell 43(8):2752\u20132764","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1774_CR2","doi-asserted-by":"publisher","first-page":"4104","DOI":"10.1109\/TIP.2022.3180585","volume":"31","author":"Y Liu","year":"2022","unstructured":"Liu Y, Yuan J, Tu Z (2022) Motion-driven visual tempo learning for video-based action recognition. IEEE Trans Image Process 31:4104\u20134116","journal-title":"IEEE Trans Image Process"},{"issue":"4","key":"1774_CR3","doi-asserted-by":"publisher","first-page":"779","DOI":"10.1007\/s13042-019-01024-2","volume":"11","author":"X Jin","year":"2020","unstructured":"Jin X, Sun W, Jin Z (2020) A discriminative deep association learning for facial expression recognition. Int J Mach Learn Cybern 11(4):779\u2013793","journal-title":"Int J Mach Learn Cybern"},{"issue":"1","key":"1774_CR4","doi-asserted-by":"publisher","first-page":"166","DOI":"10.1109\/TFUZZ.2020.2984991","volume":"29","author":"H Lu","year":"2020","unstructured":"Lu H, Zhang M, Xu X, Li Y, Shen HT (2020) Deep fuzzy hashing network for efficient image retrieval. IEEE Trans Fuzzy Syst 29(1):166\u2013176","journal-title":"IEEE Trans Fuzzy Syst"},{"key":"1774_CR5","doi-asserted-by":"publisher","first-page":"287","DOI":"10.1016\/j.neucom.2022.09.071","volume":"512","author":"R Yue","year":"2022","unstructured":"Yue R, Tian Z, Du S (2022) Action recognition based on rgb and skeleton data sets: a survey. Neurocomputing 512:287\u2013306","journal-title":"Neurocomputing"},{"issue":"4","key":"1774_CR6","doi-asserted-by":"publisher","first-page":"1157","DOI":"10.1007\/s13042-021-01441-2","volume":"13","author":"MH Javed","year":"2022","unstructured":"Javed MH, Yu Z, Li T, Rajeh TM, Rafique F, Waqar S (2022) Hybrid two-stream dynamic CNN for view adaptive human action recognition using ensemble learning. Int J Mach Learn Cybern 13(4):1157\u20131166","journal-title":"Int J Mach Learn Cybern"},{"key":"1774_CR7","doi-asserted-by":"crossref","unstructured":"Wu W, He D, Tan X, Chen S, Wen S (2019) Multi-agent reinforcement learning based frame sampling for effective untrimmed video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 6222\u20136231","DOI":"10.1109\/ICCV.2019.00632"},{"key":"1774_CR8","doi-asserted-by":"crossref","unstructured":"Du Y, Yuan C, Li B, Zhao L, Li Y, Hu W (2018) Interaction-aware spatio-temporal pyramid attention networks for action classification. In: Proceedings of the European Conference on Computer Vision, pp 373\u2013389","DOI":"10.1007\/978-3-030-01270-0_23"},{"key":"1774_CR9","first-page":"1","volume":"2","author":"MH Javed","year":"2021","unstructured":"Javed MH, Yu Z, Li T, Rajeh TM, Rafique F, Waqar S (2021) Hybrid two-stream dynamic cnn for view adaptive human action recognition using ensemble learning. Int J Mach Learn Cybern 2:1\u201310","journal-title":"Int J Mach Learn Cybern"},{"issue":"8","key":"1774_CR10","doi-asserted-by":"publisher","first-page":"2329","DOI":"10.1016\/j.patcog.2015.03.006","volume":"48","author":"M Ziaeefard","year":"2015","unstructured":"Ziaeefard M, Bergevin R (2015) Semantic human activity recognition: a literature review. Pattern Recogn 48(8):2329\u20132345","journal-title":"Pattern Recogn"},{"key":"1774_CR11","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1016\/j.patcog.2018.08.016","volume":"86","author":"L Chen","year":"2019","unstructured":"Chen L, Song Z, Lu J, Zhou J (2019) Learning principal orientations and residual descriptor for action recognition. Pattern Recogn 86:14\u201326","journal-title":"Pattern Recogn"},{"key":"1774_CR12","doi-asserted-by":"crossref","unstructured":"Tran D, Wang H, Torresani L, Ray J, LeCun Y, Paluri M (2018) A closer look at spatiotemporal convolutions for action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 6450\u20136459","DOI":"10.1109\/CVPR.2018.00675"},{"key":"1774_CR13","doi-asserted-by":"crossref","unstructured":"Qiu Z, Yao T, Mei T (2017) Learning spatio-temporal representation with pseudo-3d residual networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp 5533\u20135541","DOI":"10.1109\/ICCV.2017.590"},{"key":"1774_CR14","doi-asserted-by":"crossref","unstructured":"Zhu Y, Newsam S (2018) Random temporal skipping for multirate video analysis. In: Asian Conference on Computer Vision, pp 542\u2013557","DOI":"10.1007\/978-3-030-20893-6_34"},{"key":"1774_CR15","doi-asserted-by":"crossref","unstructured":"Zhang D, Dai X, Wang Y-F (2018) Dynamic temporal pyramid network: A closer look at multi-scale modeling for activity detection. In: Asian Conference on Computer Vision, pp 712\u2013728","DOI":"10.1007\/978-3-030-20870-7_44"},{"key":"1774_CR16","doi-asserted-by":"crossref","unstructured":"Feichtenhofer C, Fan H, Malik J, He K (2019) Slowfast networks for video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 6202\u20136211","DOI":"10.1109\/ICCV.2019.00630"},{"key":"1774_CR17","doi-asserted-by":"publisher","first-page":"446","DOI":"10.1016\/j.neucom.2019.05.058","volume":"358","author":"Z Zheng","year":"2019","unstructured":"Zheng Z, An G, Wu D, Ruan Q (2019) Spatial-temporal pyramid based convolutional neural network for action recognition. Neurocomputing 358:446\u2013455","journal-title":"Neurocomputing"},{"key":"1774_CR18","unstructured":"Yu F, Koltun V (2015) Multi-scale context aggregation by dilated convolutions. CoRR arXiv:1511.07122"},{"key":"1774_CR19","doi-asserted-by":"crossref","unstructured":"Kuehne H, Jhuang H, Garrote E, Poggio TA, Serre T (2011) HMDB: A large video database for human motion recognition. In: Metaxas DN, Quan L, Sanfeliu A, Gool LV (eds) IEEE International Conference on Computer Vision, pp 2556\u20132563","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"1774_CR20","unstructured":"Soomro K, Zamir AR, Shah M (2012) UCF101: A dataset of 101 human actions classes from videos in the wild. CoRR arXiv:1212.0402"},{"key":"1774_CR21","doi-asserted-by":"crossref","unstructured":"Li Y, Li Y, Vasconcelos N (2018) Resound: Towards action recognition without representation bias. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 513\u2013528","DOI":"10.1007\/978-3-030-01231-1_32"},{"key":"1774_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107321","volume":"103","author":"Y Chen","year":"2020","unstructured":"Chen Y, Ma G, Yuan C, Li B, Zhang H, Wang F, Hu W (2020) Graph convolutional network with structure pooling and joint-wise channel attention for action recognition. Pattern Recogn 103:107321","journal-title":"Pattern Recogn"},{"key":"1774_CR23","doi-asserted-by":"crossref","unstructured":"Tran D, Bourdev L, Fergus R, Torresani L, Paluri M (2015) Learning spatiotemporal features with 3d convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp 4489\u20134497","DOI":"10.1109\/ICCV.2015.510"},{"key":"1774_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.107037","volume":"98","author":"J Li","year":"2020","unstructured":"Li J, Liu X, Zhang M, Wang D (2020) Spatio-temporal deformable 3d convnets with attention for action recognition. Pattern Recogn 98:107037","journal-title":"Pattern Recogn"},{"key":"1774_CR25","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos. In: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, pp. 568\u2013576 (2014)"},{"key":"1774_CR26","doi-asserted-by":"crossref","unstructured":"Hara K, Kataoka H, Satoh Y (2018) Can spatiotemporal 3d cnns retrace the history of 2d cnns and imagenet? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 6546\u20136555","DOI":"10.1109\/CVPR.2018.00685"},{"issue":"3","key":"1774_CR27","doi-asserted-by":"publisher","first-page":"823","DOI":"10.1007\/s13042-020-01204-5","volume":"12","author":"D Zhuang","year":"2021","unstructured":"Zhuang D, Jiang M, Kong J, Liu T (2021) Spatiotemporal attention enhanced features fusion network for action recognition. Int J Mach Learn Cybern 12(3):823\u2013841","journal-title":"Int J Mach Learn Cybern"},{"key":"1774_CR28","doi-asserted-by":"crossref","unstructured":"Feichtenhofer C, Pinz A, Zisserman A (2016) Convolutional two-stream network fusion for video action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1933\u20131941","DOI":"10.1109\/CVPR.2016.213"},{"key":"1774_CR29","doi-asserted-by":"crossref","unstructured":"Feichtenhofer C, Pinz A, Wildes RP (2017) Spatiotemporal multiplier networks for video action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4768\u20134777","DOI":"10.1109\/CVPR.2017.787"},{"key":"1774_CR30","doi-asserted-by":"crossref","unstructured":"Zolfaghari M, Singh K, Brox T (2018) Eco: Efficient convolutional network for online video understanding. In: Proceedings of the European Conference on Computer Vision, pp 695\u2013712","DOI":"10.1007\/978-3-030-01216-8_43"},{"issue":"3","key":"1774_CR31","doi-asserted-by":"publisher","first-page":"1347","DOI":"10.1109\/TIP.2017.2778563","volume":"27","author":"W Du","year":"2017","unstructured":"Du W, Wang Y, Qiao Y (2017) Recurrent spatial-temporal attention network for action recognition in videos. IEEE Trans Image Process 27(3):1347\u20131360","journal-title":"IEEE Trans Image Process"},{"issue":"9","key":"1774_CR32","doi-asserted-by":"publisher","first-page":"4646","DOI":"10.1109\/TIP.2019.2912357","volume":"28","author":"C Li","year":"2019","unstructured":"Li C, Zhang B, Chen C, Ye Q, Han J, Guo G, Ji R (2019) Deep manifold structure transfer for action recognition. IEEE Trans Image Process 28(9):4646\u20134658","journal-title":"IEEE Trans Image Process"},{"key":"1774_CR33","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.patcog.2018.07.028","volume":"85","author":"H Yang","year":"2019","unstructured":"Yang H, Yuan C, Li B, Du Y, Xing J, Hu W, Maybank SJ (2019) Asymmetric 3d convolutional neural networks for action recognition. Pattern Recogn 85:1\u201312","journal-title":"Pattern Recogn"},{"key":"1774_CR34","doi-asserted-by":"crossref","unstructured":"Zhou B, Andonian A, Oliva A, Torralba A (2018) Temporal relational reasoning in videos. In: Proceedings of the European Conference on Computer Vision, vol. 11205, pp 831\u2013846","DOI":"10.1007\/978-3-030-01246-5_49"},{"key":"1774_CR35","doi-asserted-by":"crossref","unstructured":"Shi Y, Tian Y, Huang T, Wang Y (2018) Temporal attentive network for action recognition. In: 2018 IEEE International Conference on Multimedia and Expo (ICME), pp 1\u20136","DOI":"10.1109\/ICME.2018.8486452"},{"key":"1774_CR36","doi-asserted-by":"crossref","unstructured":"Lin J, Gan C, Han S (2019) Tsm: Temporal shift module for efficient video understanding. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 7083\u20137093","DOI":"10.1109\/ICCV.2019.00718"},{"key":"1774_CR37","doi-asserted-by":"crossref","unstructured":"Luo C, Yuille AL (2019) Grouped spatial-temporal aggregation for efficient action recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 5512\u20135521","DOI":"10.1109\/ICCV.2019.00561"},{"key":"1774_CR38","first-page":"3468","volume":"2","author":"C Feichtenhofer","year":"2016","unstructured":"Feichtenhofer C, Pinz A, Wildes RP (2016) Spatiotemporal residual networks for video action recognition. Adv Neural Inf Process Syst 2:3468\u20133476","journal-title":"Adv Neural Inf Process Syst"},{"key":"1774_CR39","doi-asserted-by":"crossref","unstructured":"Xie S, Sun C, Huang J, Tu Z, Murphy K (2018) Rethinking spatiotemporal feature learning: Speed-accuracy trade-offs in video classification. In: Proceedings of the European Conference on Computer Vision, pp. 305\u2013321","DOI":"10.1007\/978-3-030-01267-0_19"},{"key":"1774_CR40","doi-asserted-by":"crossref","unstructured":"Carreira J, Zisserman A (2017) Quo vadis, action recognition? a new model and the kinetics dataset. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 4724\u20134733","DOI":"10.1109\/CVPR.2017.502"},{"key":"1774_CR41","doi-asserted-by":"crossref","unstructured":"Li C, Zhong Q, Xie D, Pu S (2019) Collaborative spatiotemporal feature learning for video action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 7872\u20137881","DOI":"10.1109\/CVPR.2019.00806"},{"issue":"10","key":"1774_CR42","doi-asserted-by":"publisher","first-page":"4941","DOI":"10.1109\/TIP.2019.2917283","volume":"28","author":"B Xu","year":"2019","unstructured":"Xu B, Ye H, Zheng Y, Wang H, Luwang T, Jiang Y-G (2019) Dense dilated network for video action recognition. IEEE Trans Image Process 28(10):4941\u20134953","journal-title":"IEEE Trans Image Process"},{"key":"1774_CR43","first-page":"2","volume":"2","author":"J Fu","year":"2020","unstructured":"Fu J, Liu J, Jiang J, Li Y, Bao Y, Lu H (2020) Scene segmentation with dual relation-aware attention network. IEEE Trans Neural Netw Learn Syst 2:2","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"1774_CR44","doi-asserted-by":"crossref","unstructured":"Li Y, Ji B, Shi X, Zhang J, Kang B, Wang L (2020) Tea: Temporal excitation and aggregation for action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 909\u2013918","DOI":"10.1109\/CVPR42600.2020.00099"},{"key":"1774_CR45","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1016\/j.patrec.2019.02.007","volume":"122","author":"Z Wang","year":"2019","unstructured":"Wang Z, Chen K, Zhang M, He P, Wang Y, Zhu P, Yang Y (2019) Multi-scale aggregation network for temporal action proposals. Pattern Recogn Lett 122:60\u201365","journal-title":"Pattern Recogn Lett"},{"key":"1774_CR46","doi-asserted-by":"crossref","unstructured":"Yang C, Xu Y, Shi J, Dai B, Zhou B (2020) Temporal pyramid network for action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 591\u2013600","DOI":"10.1109\/CVPR42600.2020.00067"},{"key":"1774_CR47","doi-asserted-by":"crossref","unstructured":"Li X, Shuai B, Tighe J (2020) Directional temporal modeling for action recognition. In: Proceedings of the European Conference on Computer Vision, pp 275\u2013291","DOI":"10.1007\/978-3-030-58539-6_17"},{"key":"1774_CR48","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, pp 5998\u20136008"},{"key":"1774_CR49","doi-asserted-by":"crossref","unstructured":"Wang X, Girshick R, Gupta A, He K (2018) Non-local neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 7794\u20137803","DOI":"10.1109\/CVPR.2018.00813"},{"key":"1774_CR50","unstructured":"Kay W, Carreira J, Simonyan K, Zhang B, Hillier C, Vijayanarasimhan S, Viola F, Green T, Back T, Natsev P, Suleyman M, Zisserman A (2017) The kinetics human action video dataset. CoRR arXiv:1705.06950"},{"issue":"11","key":"1774_CR51","doi-asserted-by":"publisher","first-page":"2740","DOI":"10.1109\/TPAMI.2018.2868668","volume":"41","author":"L Wang","year":"2018","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y, Lin D, Tang X, Van Gool L (2018) Temporal segment networks for action recognition in videos. IEEE Trans Pattern Anal Mach Intell 41(11):2740\u20132755","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"11","key":"1774_CR52","doi-asserted-by":"publisher","first-page":"2990","DOI":"10.1109\/TMM.2020.2965434","volume":"22","author":"J Li","year":"2020","unstructured":"Li J, Liu X, Zhang W, Zhang M, Song J, Sebe N (2020) Spatio-temporal attention networks for action recognition and detection. IEEE Trans Multimedia 22(11):2990\u20133001","journal-title":"IEEE Trans Multimedia"},{"key":"1774_CR53","doi-asserted-by":"crossref","unstructured":"Jiang B, Wang M, Gan W, Wu W, Yan J (2019) Stm: Spatiotemporal and motion encoding for action recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 2000\u20132009","DOI":"10.1109\/ICCV.2019.00209"},{"key":"1774_CR54","doi-asserted-by":"crossref","unstructured":"Diba A, Fayyaz M, Sharma V, Arzani MM, Yousefzadeh R, Gall J, Van\u00a0Gool L (2018) Spatio-temporal channel correlation networks for action classification. In: Proceedings of the European Conference on Computer Vision, pp 284\u2013299","DOI":"10.1007\/978-3-030-01225-0_18"},{"key":"1774_CR55","doi-asserted-by":"crossref","unstructured":"Zhou Y, Sun X, Zha Z-J, Zeng W (2018) Mict: Mixed 3d\/2d convolutional tube for human action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 449\u2013458","DOI":"10.1109\/CVPR.2018.00054"},{"key":"1774_CR56","doi-asserted-by":"publisher","first-page":"5783","DOI":"10.1109\/TIP.2020.2984904","volume":"29","author":"H Yang","year":"2020","unstructured":"Yang H, Yuan C, Zhang L, Sun Y, Hu W, Maybank SJ (2020) STA-CNN: convolutional spatial-temporal attention learning for action recognition. IEEE Trans Image Process 29:5783\u20135793","journal-title":"IEEE Trans Image Process"},{"key":"1774_CR57","doi-asserted-by":"crossref","unstructured":"Kanojia G, Kumawat S, Raman S (2019) Attentive spatio-temporal representation learning for diving classification. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp 2467\u20132476","DOI":"10.1109\/CVPRW.2019.00302"},{"key":"1774_CR58","unstructured":"Bertasius G, Wang H, Torresani L (2021) Is space-time attention all you need for video understanding? In: Proceedings of the International Conference on Machine Learning (ICML), vol. 139, pp 813\u2013824"},{"issue":"2","key":"1774_CR59","doi-asserted-by":"publisher","first-page":"336","DOI":"10.1007\/s11263-019-01228-7","volume":"128","author":"RR Selvaraju","year":"2020","unstructured":"Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D (2020) Grad-cam: Visual explanations from deep networks via gradient-based localization. Int J Comput Vis 128(2):336\u2013359","journal-title":"Int J Comput Vis"}],"container-title":["International Journal of Machine Learning and Cybernetics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-023-01774-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13042-023-01774-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13042-023-01774-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,5,15]],"date-time":"2023-05-15T14:57:11Z","timestamp":1684162631000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13042-023-01774-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,18]]},"references-count":59,"journal-issue":{"issue":"7","published-print":{"date-parts":[[2023,7]]}},"alternative-id":["1774"],"URL":"https:\/\/doi.org\/10.1007\/s13042-023-01774-0","relation":{},"ISSN":["1868-8071","1868-808X"],"issn-type":[{"type":"print","value":"1868-8071"},{"type":"electronic","value":"1868-808X"}],"subject":[],"published":{"date-parts":[[2023,1,18]]},"assertion":[{"value":"19 January 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 January 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 January 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}