{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,6]],"date-time":"2025-05-06T16:38:46Z","timestamp":1746549526916},"reference-count":57,"publisher":"Springer Science and Business Media LLC","issue":"20","license":[{"start":{"date-parts":[[2023,7,5]],"date-time":"2023-07-05T00:00:00Z","timestamp":1688515200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,7,5]],"date-time":"2023-07-05T00:00:00Z","timestamp":1688515200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"National Key R &D Program of China under Grant","award":["2018YFE0205503"],"award-info":[{"award-number":["2018YFE0205503"]}]},{"name":"International Cooperation and Exchange of NSFC under Grant","award":["61720106007"],"award-info":[{"award-number":["61720106007"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2023,10]]},"DOI":"10.1007\/s10489-023-04687-1","type":"journal-article","created":{"date-parts":[[2023,7,5]],"date-time":"2023-07-05T09:02:32Z","timestamp":1688547752000},"page":"23067-23082","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Graph attention network-optimized dynamic monocular visual odometry"],"prefix":"10.1007","volume":"53","author":[{"given":"Zhao","family":"Hongru","sequence":"first","affiliation":[]},{"given":"Qiao","family":"Xiuquan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,5]]},"reference":[{"issue":"4","key":"4687_CR1","doi-asserted-by":"publisher","first-page":"651","DOI":"10.1109\/JPROC.2019.2895105","volume":"107","author":"X Qiao","year":"2019","unstructured":"Qiao, X., Ren, P., Dustdar, S., Liu, L., Ma, H., Chen, J.: Web ar: A promising future for mobile augmented reality-state of the art, challenges, and insights. Proceedings of the IEEE 107(4), 651\u2013666 (2019)","journal-title":"Proceedings of the IEEE"},{"key":"4687_CR2","doi-asserted-by":"crossref","unstructured":"Yadav, R., Kala, R.: Fusion of visual odometry and place recognition for slam in extreme conditions. Applied Intelligence, 1\u201320 (2022)","DOI":"10.1007\/s10489-021-03050-6"},{"key":"4687_CR3","doi-asserted-by":"publisher","first-page":"2449","DOI":"10.1109\/TMM.2021.3081873","volume":"24","author":"H Liu","year":"2021","unstructured":"Liu, H., Fang, S., Zhang, Z., Li, D., Lin, K., Wang, J.: Mfdnet: Collaborative poses perception and matrix fisher distribution for head pose estimation. IEEE Transactions on Multimedia 24, 2449\u20132460 (2021)","journal-title":"IEEE Transactions on Multimedia"},{"issue":"10","key":"4687_CR4","doi-asserted-by":"publisher","first-page":"7107","DOI":"10.1109\/TII.2022.3143605","volume":"18","author":"H Liu","year":"2022","unstructured":"Liu, H., Liu, T., Zhang, Z., Sangaiah, A.K., Yang, B., Li, Y.: Arhpe: Asymmetric relation-aware representation learning for head pose estimation in industrial human-computer interaction. IEEE Transactions on Industrial Informatics 18(10), 7107\u20137117 (2022)","journal-title":"IEEE Transactions on Industrial Informatics"},{"issue":"7","key":"4687_CR5","doi-asserted-by":"publisher","first-page":"4361","DOI":"10.1109\/TII.2021.3128240","volume":"18","author":"H Liu","year":"2021","unstructured":"Liu, H., Zheng, C., Li, D., Shen, X., Lin, K., Wang, J., Zhang, Z., Zhang, Z., Xiong, N.N.: Edmf: Efficient deep matrix factorization with review feature learning for industrial recommender system. IEEE Transactions on Industrial Informatics 18(7), 4361\u20134371 (2021)","journal-title":"IEEE Transactions on Industrial Informatics"},{"key":"4687_CR6","doi-asserted-by":"crossref","unstructured":"Liu, H., Liu, T., Chen, Y., Zhang, Z., Li, Y.-F.: Ehpe: skeleton cues-based gaussian coordinate encoding for efficient human pose estimation. IEEE Transactions on Multimedia (2022)","DOI":"10.1109\/TMM.2022.3197364"},{"key":"4687_CR7","doi-asserted-by":"crossref","unstructured":"Wang, S., Clark, R., Wen, H., Trigoni, N.: Deepvo: Towards end-to-end visual odometry with deep recurrent convolutional neural networks. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 2043\u20132050 (2017). IEEE","DOI":"10.1109\/ICRA.2017.7989236"},{"issue":"4\u20135","key":"4687_CR8","doi-asserted-by":"publisher","first-page":"513","DOI":"10.1177\/0278364917734298","volume":"37","author":"S Wang","year":"2018","unstructured":"Wang, S., Clark, R., Wen, H., Trigoni, N.: End-to-end, sequence-to-sequence probabilistic visual odometry through deep neural networks. The International Journal of Robotics Research 37(4\u20135), 513\u2013542 (2018)","journal-title":"The International Journal of Robotics Research"},{"key":"4687_CR9","unstructured":"Sun, T., Sun, Y., Liu, M., Yeung, D.-Y.: Movable-object-aware visual slam via weakly supervised semantic segmentation. arXiv preprint arXiv:1906.03629 (2019)"},{"key":"4687_CR10","doi-asserted-by":"crossref","unstructured":"Kuo, X.-Y., Liu, C., Lin, K.-C., Lee, C.-Y.: Dynamic attention-based visual odometry. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 36\u201337 (2020)","DOI":"10.1109\/CVPRW50498.2020.00026"},{"key":"4687_CR11","unstructured":"Damirchi, H., Khorrambakht, R., Taghirad, H.D.: Exploring self-attention for visual odometry. arXiv preprint arXiv:2011.08634 (2020)"},{"issue":"4","key":"4687_CR12","doi-asserted-by":"publisher","first-page":"4076","DOI":"10.1109\/LRA.2018.2860039","volume":"3","author":"B Bescos","year":"2018","unstructured":"Bescos, B., F\u00e1cil, J.M., Civera, J., Neira, J.: Dynaslam: Tracking, mapping, and inpainting in dynamic scenes. IEEE Robotics and Automation Letters 3(4), 4076\u20134083 (2018)","journal-title":"IEEE Robotics and Automation Letters"},{"key":"4687_CR13","doi-asserted-by":"publisher","first-page":"166528","DOI":"10.1109\/ACCESS.2019.2952161","volume":"7","author":"L Cui","year":"2019","unstructured":"Cui, L., Ma, C.: Sof-slam: A semantic visual slam for dynamic environments. IEEE Access 7, 166528\u2013166539 (2019)","journal-title":"IEEE Access"},{"key":"4687_CR14","doi-asserted-by":"crossref","unstructured":"Wang, K., Lin, Y., Wang, L., Han, L., Hua, M., Wang, X., Lian, S., Huang, B.: A unified framework for mutual improvement of slam and semantic segmentation. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 5224\u20135230 (2019). IEEE","DOI":"10.1109\/ICRA.2019.8793499"},{"key":"4687_CR15","unstructured":"Lipton, Z.C., Berkowitz, J., Elkan, C.: A critical review of recurrent neural networks for sequence learning. arXiv preprint arXiv:1506.00019 (2015)"},{"key":"4687_CR16","doi-asserted-by":"crossref","unstructured":"Xue, F., Wang, X., Li, S., Wang, Q., Wang, J., Zha, H.: Beyond tracking: Selecting memory and refining poses for deep visual odometry. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8575\u20138583 (2019)","DOI":"10.1109\/CVPR.2019.00877"},{"key":"4687_CR17","doi-asserted-by":"crossref","unstructured":"Li, S., Xue, F., Wang, X., Yan, Z., Zha, H.: Sequential adversarial learning for self-supervised deep visual odometry. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2851\u20132860 (2019)","DOI":"10.1109\/ICCV.2019.00294"},{"key":"4687_CR18","doi-asserted-by":"crossref","unstructured":"Zou, Y., Ji, P., Tran, Q.-H., Huang, J.-B., Chandraker, M.: Learning monocular visual odometry via self-supervised long-term modeling. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XIV 16, pp. 710\u2013727 (2020). Springer","DOI":"10.1007\/978-3-030-58568-6_42"},{"key":"4687_CR19","doi-asserted-by":"crossref","unstructured":"Xue, F., Wang, Q., Wang, X., Dong, W., Wang, J., Zha, H.: Guided feature selection for deep visual odometry. In: Asian Conference on Computer Vision, pp. 293\u2013308 (2018). Springer","DOI":"10.1007\/978-3-030-20876-9_19"},{"key":"4687_CR20","doi-asserted-by":"crossref","unstructured":"Saputra, M.R.U., de Gusmao, P.P., Wang, S., Markham, A., Trigoni, N.: Learning monocular visual odometry through geometry-aware curriculum learning. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 3549\u20133555 (2019). IEEE","DOI":"10.1109\/ICRA.2019.8793581"},{"key":"4687_CR21","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1016\/j.robot.2016.11.012","volume":"89","author":"Y Sun","year":"2017","unstructured":"Sun, Y., Liu, M., Meng, M.Q.-H.: Improving rgb-d slam in dynamic environments: A motion removal approach. Robotics and Autonomous Systems 89, 110\u2013122 (2017)","journal-title":"Robotics and Autonomous Systems"},{"key":"4687_CR22","unstructured":"Dai, W., Zhang, Y., Li, P., Fang, Z., Scherer, S.: Rgb-d slam in dynamic environments using point correlations. IEEE Transactions on Pattern Analysis and Machine Intelligence (2020)"},{"key":"4687_CR23","doi-asserted-by":"crossref","unstructured":"Yu, C., Liu, Z., Liu, X.-J., Xie, F., Yang, Y., Wei, Q., Fei, Q.: Ds-slam: A semantic visual slam towards dynamic environments. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1168\u20131174 (2018). IEEE","DOI":"10.1109\/IROS.2018.8593691"},{"issue":"12","key":"4687_CR24","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE transactions on pattern analysis and machine intelligence 39(12), 2481\u20132495 (2017)","journal-title":"IEEE transactions on pattern analysis and machine intelligence"},{"key":"4687_CR25","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"4687_CR26","doi-asserted-by":"crossref","unstructured":"Ji, T., Wang, C., Xie, L.: Towards real-time semantic rgb-d slam in dynamic environments. arXiv preprint arXiv:2104.01316 (2021)","DOI":"10.1109\/ICRA48506.2021.9561743"},{"issue":"1","key":"4687_CR27","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1109\/TNN.2008.2005605","volume":"20","author":"F Scarselli","year":"2008","unstructured":"Scarselli, F., Gori, M., Tsoi, A.C., Hagenbuchner, M., Monfardini, G.: The graph neural network model. IEEE transactions on neural networks 20(1), 61\u201380 (2008)","journal-title":"IEEE transactions on neural networks"},{"key":"4687_CR28","unstructured":"Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)"},{"key":"4687_CR29","unstructured":"Hamilton, W., Ying, Z., Leskovec, J.: Inductive representation learning on large graphs. Advances in neural information processing systems 30 (2017)"},{"key":"4687_CR30","unstructured":"Veli\u010dkovi\u0107, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. arXiv preprint arXiv:1710.10903 (2017)"},{"key":"4687_CR31","doi-asserted-by":"crossref","unstructured":"Xue, F., Wu, X., Cai, S., Wang, J.: Learning multi-view camera relocalization with graph neural networks. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11372\u201311381 (2020). IEEE","DOI":"10.1109\/CVPR42600.2020.01139"},{"key":"4687_CR32","doi-asserted-by":"crossref","unstructured":"Turkoglu, M.O., Brachmann, E., Schindler, K., Brostow, G., Monszpart, A.: Visual camera re-localization using graph neural networks and relative pose supervision. arXiv preprint arXiv:2104.02538 (2021)","DOI":"10.1109\/3DV53792.2021.00025"},{"key":"4687_CR33","doi-asserted-by":"crossref","unstructured":"Jiao, J., Cao, Y., Song, Y., Lau, R.: Look deeper into depth: Monocular depth estimation with semantic booster and attention-driven loss. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 53\u201369 (2018)","DOI":"10.1007\/978-3-030-01267-0_4"},{"key":"4687_CR34","unstructured":"Gao, T., Wei, W., Cai, Z., Fan, Z., Xie, S., Wang, X., Yu, Q.: Ci-net: Contextual information for joint semantic segmentation and depth estimation. arXiv preprint arXiv:2107.13800 (2021)"},{"key":"4687_CR35","doi-asserted-by":"crossref","unstructured":"Nekrasov, V., Dharmasiri, T., Spek, A., Drummond, T., Shen, C., Reid, I.: Real-time joint semantic segmentation and depth estimation using asymmetric annotations. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 7101\u20137107 (2019). IEEE","DOI":"10.1109\/ICRA.2019.8794220"},{"key":"4687_CR36","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"4687_CR37","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"4687_CR38","unstructured":"Nekrasov, V., Shen, C., Reid, I.: Light-weight refinenet for real-time semantic segmentation. arXiv preprint arXiv:1810.03272 (2018)"},{"issue":"5","key":"4687_CR39","first-page":"1228","volume":"42","author":"G Lin","year":"2019","unstructured":"Lin, G., Liu, F., Milan, A., Shen, C., Reid, I.: Refinenet: Multi-path refinement networks for dense prediction. IEEE transactions on pattern analysis and machine intelligence 42(5), 1228\u20131242 (2019)","journal-title":"IEEE transactions on pattern analysis and machine intelligence"},{"issue":"9","key":"4687_CR40","doi-asserted-by":"publisher","first-page":"838","DOI":"10.1016\/j.bjoms.2014.07.253","volume":"52","author":"NL Gerlach","year":"2014","unstructured":"Gerlach, N.L., Meijer, G.J., Kroon, D.-J., Bronkhorst, E.M., Berg\u00e9, S.J., Maal, T.J.J.: Evaluation of the potential of automatic segmentation of the mandibular canal using cone-beam computed tomography. British journal of oral and maxillofacial surgery 52(9), 838\u2013844 (2014)","journal-title":"British journal of oral and maxillofacial surgery"},{"key":"4687_CR41","doi-asserted-by":"crossref","unstructured":"Sun, D., Yang, X., Liu, M.-Y., Kautz, J.: Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8934\u20138943 (2018)","DOI":"10.1109\/CVPR.2018.00931"},{"key":"4687_CR42","unstructured":"Rong, Y., Huang, W., Xu, T., Huang, J.: Dropedge: Towards deep graph convolutional networks on node classification. arXiv preprint arXiv:1907.10903 (2019)"},{"key":"4687_CR43","doi-asserted-by":"crossref","unstructured":"Wan, Y., Gao, W., Wu, Y.: Optical flow assisted monocular visual odometry. In: Asian Conference on Pattern Recognition, pp. 366\u2013377 (2019). Springer","DOI":"10.1007\/978-3-030-41404-7_26"},{"key":"4687_CR44","doi-asserted-by":"crossref","unstructured":"Cordts, M., Omran, M., Ramos, S., Rehfeld, T., Enzweiler, M., Benenson, R., Franke, U., Roth, S., Schiele, B.: The cityscapes dataset for semantic urban scene understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3213\u20133223 (2016)","DOI":"10.1109\/CVPR.2016.350"},{"key":"4687_CR45","doi-asserted-by":"crossref","unstructured":"Campos, C., Elvira, R., Rodr\u00edguez, J.J.G., Montiel, J.M., Tard\u00f3s, J.D.: Orb-slam3: An accurate open-source library for visual, visual-inertial, and multimap slam. IEEE Transactions on Robotics (2021)","DOI":"10.1109\/TRO.2021.3075644"},{"key":"4687_CR46","doi-asserted-by":"crossref","unstructured":"Geiger, A., Ziegler, J., Stiller, C.: Stereoscan: Dense 3d reconstruction in real-time. In: 2011 IEEE Intelligent Vehicles Symposium (IV), pp. 963\u2013968 (2011). Ieee","DOI":"10.1109\/IVS.2011.5940405"},{"issue":"9","key":"4687_CR47","doi-asserted-by":"publisher","first-page":"2265","DOI":"10.1007\/s11263-022-01641-5","volume":"130","author":"S Lee","year":"2022","unstructured":"Lee, S., Rameau, F., Im, S., Kweon, I.S.: Self-supervised monocular depth and motion learning in dynamic scenes: Semantic prior to rescue. International Journal of Computer Vision 130(9), 2265\u20132285 (2022)","journal-title":"International Journal of Computer Vision"},{"key":"4687_CR48","unstructured":"Kazerouni, A., Heydarian, A., Soltany, M., Mohammadshahi, A., Omidi, A., Ebadollahi, S.: An intelligent modular real-time vision-based system for environment perception"},{"key":"4687_CR49","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Sapra, K., Reda, F.A., Shih, K.J., Newsam, S., Tao, A., Catanzaro, B.: Improving semantic segmentation via video propagation and label relaxation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8856\u20138865 (2019)","DOI":"10.1109\/CVPR.2019.00906"},{"key":"4687_CR50","unstructured":"Kreso, I., Segvic, S., Krapac, J.: Ladder-style densenets for semantic segmentation of large natural images. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 238\u2013245 (2017)"},{"issue":"2","key":"4687_CR51","doi-asserted-by":"publisher","first-page":"3580","DOI":"10.1109\/LRA.2020.2978666","volume":"5","author":"\u00d6 Erkent","year":"2020","unstructured":"Erkent, \u00d6., Laugier, C.: Semantic segmentation with unsupervised domain adaptation under varying weather conditions for autonomous vehicles. IEEE Robotics and Automation Letters 5(2), 3580\u20133587 (2020)","journal-title":"IEEE Robotics and Automation Letters"},{"key":"4687_CR52","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.700"},{"issue":"11","key":"4687_CR53","doi-asserted-by":"publisher","first-page":"3174","DOI":"10.1109\/TCSVT.2017.2740321","volume":"28","author":"Y Cao","year":"2017","unstructured":"Cao, Y., Wu, Z., Shen, C.: Estimating depth from monocular images as classification using deep fully convolutional residual networks. IEEE Transactions on Circuits and Systems for Video Technology 28(11), 3174\u20133182 (2017)","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"key":"4687_CR54","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac\u00a0Aodha, O., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"issue":"6","key":"4687_CR55","doi-asserted-by":"publisher","first-page":"381","DOI":"10.1145\/358669.358692","volume":"24","author":"MA Fischler","year":"1981","unstructured":"Fischler, M.A., Bolles, R.C.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Communications of the ACM 24(6), 381\u2013395 (1981)","journal-title":"Communications of the ACM"},{"issue":"1","key":"4687_CR56","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/LRA.2015.2505717","volume":"1","author":"G Costante","year":"2015","unstructured":"Costante, G., Mancini, M., Valigi, P., Ciarfuglia, T.A.: Exploring representation learning with cnns for frame-to-frame ego-motion estimation. IEEE robotics and automation letters 1(1), 18\u201325 (2015)","journal-title":"IEEE robotics and automation letters"},{"key":"4687_CR57","doi-asserted-by":"crossref","unstructured":"Zhong, F., Wang, S., Zhang, Z., Wang, Y.: Detect-slam: Making object detection and slam mutually beneficial. In: 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1001\u20131010 (2018). IEEE","DOI":"10.1109\/WACV.2018.00115"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-023-04687-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-023-04687-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-023-04687-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,21]],"date-time":"2023-10-21T16:05:25Z","timestamp":1697904325000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-023-04687-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,5]]},"references-count":57,"journal-issue":{"issue":"20","published-print":{"date-parts":[[2023,10]]}},"alternative-id":["4687"],"URL":"https:\/\/doi.org\/10.1007\/s10489-023-04687-1","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,5]]},"assertion":[{"value":"4 May 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 July 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}]}}