{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T21:18:32Z","timestamp":1769635112321,"version":"3.49.0"},"reference-count":53,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2023,8,7]],"date-time":"2023-08-07T00:00:00Z","timestamp":1691366400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,8,7]],"date-time":"2023-08-07T00:00:00Z","timestamp":1691366400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U1813222"],"award-info":[{"award-number":["U1813222"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003787","name":"Natural Science Foundation of Hebei Province","doi-asserted-by":"publisher","award":["F2021202038"],"award-info":[{"award-number":["F2021202038"]}],"id":[{"id":"10.13039\/501100003787","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-16421-7","type":"journal-article","created":{"date-parts":[[2023,8,7]],"date-time":"2023-08-07T03:09:07Z","timestamp":1691377747000},"page":"22409-22431","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["GAF-Net: Global view guided attribute fusion network for remote sensing image captioning"],"prefix":"10.1007","volume":"83","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0499-9010","authenticated-orcid":false,"given":"Yuqing","family":"Peng","sequence":"first","affiliation":[]},{"given":"Yamin","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Jiao","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Xinhao","family":"Ji","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,7]]},"reference":[{"key":"16421_CR1","doi-asserted-by":"crossref","unstructured":"Anderson P (2018) Bottom-up and top-down attention for image captioning and visual question answering. In: Proc IEEE\/CVF Conf Comput Vis Pattern Recognit(CVPR), Boston, USA,\u00a0 pp 6077\u20136086","DOI":"10.1109\/CVPR.2018.00636"},{"key":"16421_CR2","unstructured":"Aswani AV, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Proc Adv Neural Inf Process Syst(NIPS), USA,\u00a0 pp 5998\u20136008"},{"key":"16421_CR3","doi-asserted-by":"publisher","first-page":"1309","DOI":"10.1007\/s00371-020-01867-9","volume":"37","author":"G Barlas","year":"2021","unstructured":"Barlas G, Veinidis C, Arampatzis A (2021) What we see in a photograph: content selection for image captioning. Vis Comput 37:1309\u20131326. https:\/\/doi.org\/10.1007\/s00371-020-01867-9","journal-title":"Vis Comput"},{"key":"16421_CR4","doi-asserted-by":"publisher","first-page":"35329","DOI":"10.1007\/s11042-019-08116-9","volume":"78","author":"D Cao","year":"2019","unstructured":"Cao D, Zhu M, Gao L (2019) An image caption method based on object detection. Multimed Tools Appl 78:35329\u201335350. https:\/\/doi.org\/10.1007\/s11042-019-08116-9","journal-title":"Multimed Tools Appl"},{"key":"16421_CR5","doi-asserted-by":"publisher","first-page":"2959","DOI":"10.1007\/s11042-017-4593-1","volume":"77","author":"YS Chang","year":"2018","unstructured":"Chang YS (2018) Fine-grained attention for image caption generation. Multimed Tools Appl 77:2959\u20132971. https:\/\/doi.org\/10.1007\/s11042-017-4593-1","journal-title":"Multimed Tools Appl"},{"key":"16421_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/LGRS.2022.3192062","volume":"19","author":"Z Chen","year":"2022","unstructured":"Chen Z, Wang J, Ma A, Zhang Y (2022) TypeFormer: Multiscale transformer with type controller for remote sensing image caption. IEEE Geosci Remote Sens Lett 19:1\u20135. https:\/\/doi.org\/10.1109\/LGRS.2022.3192062","journal-title":"IEEE Geosci Remote Sens Lett"},{"issue":"10","key":"16421_CR7","doi-asserted-by":"publisher","first-page":"1865","DOI":"10.1109\/JPROC.2017.2675998","volume":"105","author":"G Cheng","year":"2017","unstructured":"Cheng G, Han J, Lu X (2017) Remote sensing image scene classification: Benchmark and state of the art. Proc IEEE 105(10):1865\u20131883","journal-title":"Proc IEEE"},{"key":"16421_CR8","doi-asserted-by":"publisher","first-page":"1044","DOI":"10.3390\/rs11091044","volume":"11","author":"W Cui","year":"2019","unstructured":"Cui W, Wang F, He X, Zhang D, Xu X, Yao M, Wang Z, Huang J (2019) Multi-scale semantic segmentation and spatial relationship recognition of remote sensing images based on an attention model. Remote Sens 11:1044","journal-title":"Remote Sens"},{"key":"16421_CR9","doi-asserted-by":"crossref","unstructured":"Denkowski M, Lavie A (2014) Meteor universal: Language specific translation evaluation for any target language. In: Proc of the Ninth Workshop on Statistical Machine Translation(WMT@ACL), USA,\u00a0 pp 376\u2013380","DOI":"10.3115\/v1\/W14-3348"},{"key":"16421_CR10","doi-asserted-by":"publisher","first-page":"1313","DOI":"10.1007\/s11042-021-11293-1","volume":"81","author":"K Deorukhkar","year":"2022","unstructured":"Deorukhkar K, Ket S (2022) A detailed review of prevailing image captioning methods using deep learning techniques. Multimed Tools Appl 81:1313\u20131336. https:\/\/doi.org\/10.1007\/s11042-021-11293-1","journal-title":"Multimed Tools Appl"},{"key":"16421_CR11","doi-asserted-by":"publisher","unstructured":"Farhadi A, Hejrati M, Sadeghi MA, Young P, Rashtchian C, Hockenmaier J, Forsyth D (2010) Every picture tells a story: generating sentences from images. In: Computer vision \u2013 ECCV 2010, vol 6314. Springer, Berlin, pp 15\u201329. https:\/\/doi.org\/10.1007\/978-3-642-15561-12","DOI":"10.1007\/978-3-642-15561-12"},{"key":"16421_CR12","doi-asserted-by":"publisher","unstructured":"Gong Y, Wang L, Hodosh M, Hockenmaier J, Lazebnik S (2014) Improving image-sentence embeddings using large weakly annotated photo collections. In: Computer Vision \u2013 ECCV 2014, vol 8692. Springer International Publishing, Cham, pp 529\u2013545\u00a0https:\/\/doi.org\/10.1007\/978-3-319-10593-2","DOI":"10.1007\/978-3-319-10593-2"},{"issue":"6","key":"16421_CR13","doi-asserted-by":"publisher","first-page":"3325","DOI":"10.1109\/TGRS.2014.2374218","volume":"53","author":"J Han","year":"2015","unstructured":"Han J, Zhang D, Cheng G, Guo L, Ren J (2015) Object detection in optical remote sensing images based on weakly supervised learning and high-level feature learning. IEEE Trans Geosci Remote Sens 53(6):3325\u20133337","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"16421_CR14","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proc IEEE Conf Comput Vis Pattern Recognit (CVPR), NV, USA,\u00a0 pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"16421_CR15","doi-asserted-by":"publisher","first-page":"853","DOI":"10.1613\/jair.3994","volume":"47","author":"M Hodosh","year":"2013","unstructured":"Hodosh M, Young P, Hockenmaier J (2013) Framing image description as a ranking task: data, models and evaluation metrics. J Artif Intell Res 47:853\u2013899. https:\/\/doi.org\/10.1613\/jair.3994","journal-title":"J Artif Intell Res"},{"key":"16421_CR16","doi-asserted-by":"crossref","unstructured":"Huang L, Wang W, Chen J, Wei XY (2019) Attention on attention for image captioning. In: Proc IEEE Int Conf Comput Vis(ICCV), Korea (South), pp 4633\u20134642","DOI":"10.1109\/ICCV.2019.00473"},{"issue":"3","key":"16421_CR17","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1109\/LGRS.2020.2980933","volume":"18","author":"W Huang","year":"2021","unstructured":"Huang W, Wang Q, Li X (2021) Denoising-based multiscale feature fusion for remote sensing image captioning. IEEE Geosci Remote Sens Lett 18(3):436\u2013440","journal-title":"IEEE Geosci Remote Sens Lett"},{"key":"16421_CR18","doi-asserted-by":"publisher","unstructured":"Kulkarni G, Premraj V, Dhar S, Li S, Choi Y, Berg AC, Berg TL (2011) Baby talk: understanding and generating simple image descriptions. In: Proc IEEE Conf Comput Vis Pattern Recognit (CVPR), CO, USA,\u00a0 pp 1601-1608. https:\/\/doi.org\/10.1109\/cvpr.2011.5995466","DOI":"10.1109\/cvpr.2011.5995466"},{"issue":"6","key":"16421_CR19","doi-asserted-by":"publisher","first-page":"939","DOI":"10.3390\/rs12060939","volume":"12","author":"Y Li","year":"2020","unstructured":"Li Y, Fang S, Jiao L, Liu R, Shang R (2020) A multi-level attention model for remote sensing image captions. Remote Sens 12(6):939","journal-title":"Remote Sens"},{"issue":"6","key":"16421_CR20","doi-asserted-by":"publisher","first-page":"5246","DOI":"10.1109\/TGRS.2020.3010106","volume":"59","author":"X Li","year":"2020","unstructured":"Li X, Zhang X, Huang W, Wang Q (2020) Truncation cross entropy loss for remote sensing image captioning. IEEE Trans Geosci Remote Sens 59(6):5246\u20135257","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"16421_CR21","unstructured":"Lin CY (2004) ROUGE: A package for automatic evaluation of summaries. Proceedings of the ACL Workshop on Text Summarization Branches Out. Baltimore, USA, pp 74\u201381"},{"key":"16421_CR22","doi-asserted-by":"publisher","first-page":"2183","DOI":"10.1109\/TGRS.2017.2776321","volume":"56","author":"X Lu","year":"2018","unstructured":"Lu X, Wang B, Zheng X, Li X (2018) Exploring models and data for remote sensing image caption generation. IEEE Trans Geosci Remote Sens 56:2183\u20132195","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"16421_CR23","doi-asserted-by":"crossref","unstructured":"Manning C, Surdeanu M, Bauer J, Finkel J, Bethard S, McClosky D (2014) The stanford coreNLP natural language processing toolkit. In: 52nd Annual Meeting of the Association for Computational Linguistics. Baltimore 2014,\u00a0 pp 55\u201360","DOI":"10.3115\/v1\/P14-5010"},{"key":"16421_CR24","unstructured":"Mao J, Xu W, Yang Y, Wang J, Y uille AL (2015) Deep captioning with multimodal recurrent neural networks (m-RNN). In: Bengio Y, LeCun Y (eds) 3rd International conference on learning representations, ICLR 2015, San Diego, CA, USA,\u00a0 pp 7\u20139"},{"key":"16421_CR25","doi-asserted-by":"crossref","unstructured":"Meng Y, Gu Y, Ye X, Tian J, Wang S, Zhang H, Hou B, Jiao L (2021) Multi-view attention network for remote sensing image captioning. International Geoscience and Remote Sensing SymposiumI(GARSS), pp 2349\u20132352","DOI":"10.1109\/IGARSS47720.2021.9555083"},{"key":"16421_CR26","doi-asserted-by":"publisher","first-page":"30615","DOI":"10.1007\/s11042-020-09539-5","volume":"79","author":"C Nogueira","year":"2020","unstructured":"Nogueira C, Vinhal T, J\u00fanior C (2020) Reference-based model using multimodal gated recurrent units for image captioning. Multimed Tools Appl 79:30615\u201330635. https:\/\/doi.org\/10.1007\/s11042-020-09539-5","journal-title":"Multimed Tools Appl"},{"key":"16421_CR27","doi-asserted-by":"crossref","unstructured":"Papineni K, Roukos S, Ward T, Zhu WJ (2002) BLEU: A method for automatic evaluation of machine translation. In: Proceedings of the 40th Annual Meeting on Association for Computational Linguistics. pp 311-318","DOI":"10.3115\/1073083.1073135"},{"key":"16421_CR28","doi-asserted-by":"crossref","unstructured":"Qu B, Li X, Tao D, Lu X (2016) Deep semantic understanding of high resolution remote sensing image. In: Proc Int Conf Comput Inf Telecommun Syst (CITS), Kunming, China,\u00a0 pp 1\u20135","DOI":"10.1109\/CITS.2016.7546397"},{"key":"16421_CR29","doi-asserted-by":"crossref","unstructured":"Ramos R, Martins B (2021) Remote sensing image captioning with continuous output neural models. SIGSPATIAL\/GIS, Beijing, China, pp 29\u201332","DOI":"10.1145\/3474717.3483631"},{"issue":"6","key":"16421_CR30","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2017","unstructured":"Ren S, He K, Girshick R, Sun J (2017) Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):1137\u20131149","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"16421_CR31","doi-asserted-by":"crossref","unstructured":"Rennie SJ, Marcheret E, Mroueh Y, Ross J, Goel V (2017) Self-critical sequence training for image captioning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu: IEEE Computer Society 2017, pp 1179-1195","DOI":"10.1109\/CVPR.2017.131"},{"key":"#cr-split#-16421_CR32.1","doi-asserted-by":"crossref","unstructured":"Rush AM, Chopra S, Weston J (2015) A neural attention model for abstractive sentence summarization. In","DOI":"10.18653\/v1\/D15-1044"},{"key":"#cr-split#-16421_CR32.2","unstructured":"Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing(EMNLP),\u00a0 pp 379-389"},{"issue":"3","key":"16421_CR33","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky O (2015) ImageNet large scale visual recognition challenge. Int J Comput Vis(IJCV) 115(3):211\u2013252","journal-title":"Int J Comput Vis(IJCV)"},{"key":"16421_CR34","doi-asserted-by":"crossref","unstructured":"See A, Liu P, Manning C (2017) Get to the point: Summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics,\u00a0vol 1. ACL 2017, pp 1073-1083","DOI":"10.18653\/v1\/P17-1099"},{"key":"16421_CR35","doi-asserted-by":"publisher","first-page":"26661","DOI":"10.1007\/s11042-020-09294-7","volume":"79","author":"X Shen","year":"2020","unstructured":"Shen X, Liu B, Zhou Y (2020) Remote sensing image caption generation via transformer and reinforcement learning. Multimed Tools Appl 79:26661\u201326682. https:\/\/doi.org\/10.1007\/s11042-020-09294-7","journal-title":"Multimed Tools Appl"},{"issue":"6","key":"16421_CR36","doi-asserted-by":"publisher","first-page":"3623","DOI":"10.1109\/TGRS.2017.2677464","volume":"55","author":"Z Shi","year":"2017","unstructured":"Shi Z, Zou Z (2017) Can a machine generate humanlike language descriptions for a remote sensing image. IEEE Trans Geosci Remote Sens 55(6):3623\u20133634","journal-title":"IEEE Trans Geosci Remote Sens"},{"issue":"8","key":"16421_CR37","doi-asserted-by":"publisher","first-page":"6922","DOI":"10.1109\/TGRS.2020.3031111","volume":"59","author":"G Sumbul","year":"2020","unstructured":"Sumbul G, Nayak S, Demir B (2020) SD-RSIC: Summarization driven deep remote sensing image captioning. Trans Geosci Remote Sens 59(8):6922\u20136934","journal-title":"Trans Geosci Remote Sens"},{"key":"16421_CR38","doi-asserted-by":"crossref","unstructured":"Vedantam R, Zitnick CL, Parikh D (2015) CIDEr: Consensus-based image description evaluation. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp 4566\u20134575","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"16421_CR39","doi-asserted-by":"publisher","first-page":"652","DOI":"10.1109\/Tpami.2016.2587640","volume":"39","author":"O Vinyals","year":"2017","unstructured":"Vinyals O, Toshev A, Bengio S, Erhan D (2017) Show and tell: lessons learned from the 2015 MSCOCO image captioning challenge. IEEE Trans Pattern Anal Mach Intell 39:652\u2013663. https:\/\/doi.org\/10.1109\/Tpami.2016.2587640","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"12","key":"16421_CR40","doi-asserted-by":"publisher","first-page":"10532","DOI":"10.1109\/TGRS.2020.3044054","volume":"59","author":"Q Wang","year":"2021","unstructured":"Wang Q, Huang W, Zhang X, Li X (2021) Word-sentence framework for remote sensing image captioning. IEEE Trans Geosci Remote Sens 59(12):10532\u201310543","journal-title":"IEEE Trans Geosci Remote Sens"},{"issue":"8","key":"16421_CR41","doi-asserted-by":"publisher","first-page":"1274","DOI":"10.1109\/LGRS.2019.2893772","volume":"16","author":"B Wang","year":"2019","unstructured":"Wang B, Lu X, Zheng X, Li X (2019) Semantic descriptions of high-resolution remote sensing images. IEEE Geosci Remote Sens Lett 16(8):1274\u20131278","journal-title":"IEEE Geosci Remote Sens Lett"},{"key":"16421_CR42","doi-asserted-by":"publisher","first-page":"256","DOI":"10.1109\/JSTARS.2019.2959208","volume":"13","author":"B Wang","year":"2020","unstructured":"Wang B, Zheng X, Qu B, Lu X (2020) Retrieval topic recurrent memory network for remote sensing image captioning. IEEE J Sel Topics Appl Earth Observ Remote Sens 13:256\u2013270","journal-title":"IEEE J Sel Topics Appl Earth Observ Remote Sens"},{"key":"16421_CR43","doi-asserted-by":"crossref","unstructured":"Wu S, Zhang X, Wang X, Li C, Jiao L (2020) Scene attention mechanism for remote sensing image caption generation. International Joint Conference on Neural Networks(IJCNN), UK,\u00a0 pp 1\u20137","DOI":"10.1109\/IJCNN48605.2020.9207381"},{"key":"16421_CR44","doi-asserted-by":"publisher","first-page":"24225","DOI":"10.1007\/s11042-020-09110-2","volume":"79","author":"P Xia","year":"2020","unstructured":"Xia P, He J, Yin J (2020) Boosting image caption generation with feature fusion module. Multimed Tools Appl 79:24225\u201324239. https:\/\/doi.org\/10.1007\/s11042-020-09110-2","journal-title":"Multimed Tools Appl"},{"key":"16421_CR45","unstructured":"Xu K (2015) Show, attend and tell: Neural image caption generation with visual attention. In: Proc. Int. Conf. Mach. Learn.,Lille, France,\u00a0 pp 2048\u20132057"},{"key":"16421_CR46","doi-asserted-by":"publisher","first-page":"25557","DOI":"10.1007\/s11042-021-10632-6","volume":"80","author":"Z Ye","year":"2021","unstructured":"Ye Z, Khan R, Naqvi N (2021) A novel automatic image caption generation using bidirectional long-short term memory framework. Multim Tools Appl 80:25557\u201325582. https:\/\/doi.org\/10.1007\/s11042-021-10632-6","journal-title":"Multim Tools Appl"},{"issue":"20","key":"16421_CR47","doi-asserted-by":"publisher","first-page":"2349","DOI":"10.3390\/rs11202349","volume":"11","author":"Z Zhang","year":"2019","unstructured":"Zhang Z, Diao W, Zhang W, Yan M, Gao X, Sun X (2019) LAM: Remote sensing image captioning with label-attention mechanism. Remote Sens 11(20):2349","journal-title":"Remote Sens"},{"key":"16421_CR48","doi-asserted-by":"crossref","unstructured":"Zhang X, Wang Q, Chen S, Li X (2019) Multi\u2013scale cropping mechanism for remote sensing image captioning. In: Proc IEEE Int Geosci Remote Sens Symp (IGARSS), Yokohama, Japan,\u00a0 pp 10039\u201310042","DOI":"10.1109\/IGARSS.2019.8900503"},{"issue":"6","key":"16421_CR49","doi-asserted-by":"publisher","first-page":"612","DOI":"10.3390\/rs11060612","volume":"11","author":"X Zhang","year":"2019","unstructured":"Zhang X, Wang X, Tang X, Zhou H, Li C (2019) Description generation for remote sensing images using attribute attention mechanism. Remote Sens 11(6):612","journal-title":"Remote Sens"},{"key":"16421_CR50","first-page":"1","volume":"60","author":"Z Zhang","year":"2022","unstructured":"Zhang Z, Zhang W, Yan M, Gao X, Fu K, Sun X (2022) Global visual feature and linguistic state guided attention for remote sensing image captioning. IEEE Trans. Geosci Remote Sens 60:1\u201316","journal-title":"IEEE Trans. Geosci Remote Sens"},{"key":"16421_CR51","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2022.3230378","volume":"60","author":"R Zhao","year":"2022","unstructured":"Zhao R, Shi Z, Zou Z (2022) High-resolution remote sensing image captioning based on structured attention. IEEE Trans Geosci Remote Sens 60:1\u201314","journal-title":"IEEE Trans Geosci Remote Sens"},{"issue":"2","key":"16421_CR52","doi-asserted-by":"publisher","first-page":"275","DOI":"10.3390\/rs12020275","volume":"12","author":"Z Zou","year":"2020","unstructured":"Zou Z, Shi T, Li W, Zhang Z, Shi Z (2020) Do game data generalize well for remote sensing image segmentation? Remote Sens 12(2):275","journal-title":"Remote Sens"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16421-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-16421-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16421-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,22]],"date-time":"2024-02-22T13:18:32Z","timestamp":1708607912000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-16421-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,7]]},"references-count":53,"journal-issue":{"issue":"8","published-online":{"date-parts":[[2024,3]]}},"alternative-id":["16421"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-16421-7","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,8,7]]},"assertion":[{"value":"10 May 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 May 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 July 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 August 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interests"}}]}}