{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T21:44:38Z","timestamp":1772315078890,"version":"3.50.1"},"reference-count":44,"publisher":"Springer Science and Business Media LLC","issue":"15","license":[{"start":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T00:00:00Z","timestamp":1761868800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T00:00:00Z","timestamp":1761868800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62301330"],"award-info":[{"award-number":["62301330"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s00371-025-04187-y","type":"journal-article","created":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T09:37:54Z","timestamp":1761903474000},"page":"12839-12851","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Dsf-net: a dual-stream fusion network integrating structural and detailed features for fundus-based diabetic retinopathy classification"],"prefix":"10.1007","volume":"41","author":[{"given":"Yang","family":"Wen","sequence":"first","affiliation":[]},{"given":"Ying","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Shuang","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Lijiao","family":"Xiong","sequence":"additional","affiliation":[]},{"given":"Huating","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Weiping","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Congrong","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Pengju","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Zhen","family":"Liang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,31]]},"reference":[{"issue":"3","key":"4187_CR1","doi-asserted-by":"publisher","first-page":"345","DOI":"10.3390\/diagnostics13030345","volume":"13","author":"A Sebastian","year":"2023","unstructured":"Sebastian, A., Elharrouss, O., Al-Maadeed, S., Almaadeed, N.: A survey on deep-learning-based diabetic retinopathy classification. Diagnostics 13(3), 345 (2023)","journal-title":"Diagnostics"},{"issue":"3","key":"4187_CR2","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1177\/09691413221144382","volume":"30","author":"Z Zhelev","year":"2023","unstructured":"Zhelev, Z., Peters, J., Rogers, M., Allen, M., Kijauskaite, G., Seedat, F., Wilkinson, E., Hyde, C.: Test accuracy of artificial intelligence-based grading of fundus images in diabetic retinopathy screening: a systematic review. J. Med. Screen. 30(3), 97\u2013112 (2023)","journal-title":"J. Med. Screen."},{"key":"4187_CR3","doi-asserted-by":"crossref","unstructured":"AbdelMaksoud, E., Barakat, S., Elmogy, M.: \u201cDiabetic retinopathy grading system based on transfer learning,\u201d arXiv preprint arXiv:2012.12515, (2020)","DOI":"10.1109\/ICDABI51230.2020.9325672"},{"key":"4187_CR4","doi-asserted-by":"crossref","unstructured":"Wen, Y., Luo, B., Shi, W., Ji, J., Cao, W., Yang, X., Sheng, B.: Sat-net: structure-aware transformer-based attention fusion network for low-quality retinal fundus images enhancement. IEEE Transactions on Multimedia (2025)","DOI":"10.1109\/TMM.2025.3565935"},{"issue":"12","key":"4187_CR5","doi-asserted-by":"publisher","first-page":"5726","DOI":"10.3390\/s23125726","volume":"23","author":"C Mohanty","year":"2023","unstructured":"Mohanty, C., Mahapatra, S., Acharya, B., Kokkoras, F., Gerogiannis, V.C., Karamitsos, I., Kanavos, A.: Using deep learning architectures for detection and classification of diabetic retinopathy. Sensors 23(12), 5726 (2023)","journal-title":"Sensors"},{"issue":"12","key":"4187_CR6","doi-asserted-by":"publisher","first-page":"2509","DOI":"10.2337\/dc18-0147","volume":"41","author":"Z Li","year":"2018","unstructured":"Li, Z., Keel, S., Liu, C., He, Y., Meng, W., Scheetz, J., Lee, P.Y., Shaw, J., Ting, D., Wong, T.Y., et al.: An automated grading system for detection of vision-threatening referable diabetic retinopathy on the basis of color fundus photographs. Diabetes Care 41(12), 2509\u20132516 (2018)","journal-title":"Diabetes Care"},{"key":"4187_CR7","doi-asserted-by":"crossref","unstructured":"Thakur, G.K., Thakur, A., Kulkarni, S., Khan, N., Khan, S.: Deep learning approaches for medical image analysis and diagnosis. Cureus 16(5), (2024)","DOI":"10.7759\/cureus.59507"},{"issue":"3","key":"4187_CR8","doi-asserted-by":"publisher","first-page":"312","DOI":"10.1016\/j.job.2022.03.003","volume":"64","author":"M Tsuneki","year":"2022","unstructured":"Tsuneki, M.: Deep learning models in medical image analysis. J. Oral Biosci. 64(3), 312\u2013320 (2022)","journal-title":"J. Oral Biosci."},{"issue":"1","key":"4187_CR9","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s41746-024-01204-7","volume":"7","author":"Z Qi","year":"2024","unstructured":"Qi, Z., Li, T., Chen, J., Yam, J.C., Wen, Y., Huang, G., Zhong, H., He, M., Zhu, D., Dai, R., et al.: A deep learning system for myopia onset prediction and intervention effectiveness evaluation in children. npj Dig. Med. 7(1), 206 (2024)","journal-title":"npj Dig. Med."},{"key":"4187_CR10","doi-asserted-by":"crossref","unstructured":"Lepetit-Aimon, G., Duval, R., Cheriet, F.: \u201cLarge receptive field fully convolutional network for semantic segmentation of retinal vasculature in fundus images,\u201d in International Workshop on Ophthalmic Medical Image Analysis, pp. 201\u2013209, Springer, (2018)","DOI":"10.1007\/978-3-030-00949-6_24"},{"issue":"18","key":"4187_CR11","doi-asserted-by":"publisher","first-page":"8446","DOI":"10.3390\/app14188446","volume":"14","author":"Y Yan","year":"2024","unstructured":"Yan, Y., Yang, L., Huang, W.: Fundus-danet: dilated convolution and fusion attention mechanism for multilabel retinal fundus image classification. Appl. Sci. 14(18), 8446 (2024)","journal-title":"Appl. Sci."},{"key":"4187_CR12","doi-asserted-by":"crossref","unstructured":"Yin, G., Sheng, L., Liu, B., Yu, N., Wang, X., Shao, J., Loy, C. C.: \u201cZoom-net: Mining deep feature interactions for visual relationship recognition,\u201d In: Proceedings of the European conference on computer vision (ECCV), pp. 322\u2013338, (2018)","DOI":"10.1007\/978-3-030-01219-9_20"},{"issue":"7","key":"4187_CR13","doi-asserted-by":"publisher","first-page":"2679","DOI":"10.1109\/TMI.2024.3371948","volume":"43","author":"Z Huang","year":"2024","unstructured":"Huang, Z., Zhao, R., Leung, F.H., Banerjee, S., Lam, K.-M., Zheng, Y.-P., Ling, S.H.: Landmark localization from medical images with generative distribution prior. IEEE Trans. Med. Imaging 43(7), 2679\u20132692 (2024)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"1","key":"4187_CR14","doi-asserted-by":"publisher","first-page":"30542","DOI":"10.1038\/s41598-024-81961-3","volume":"14","author":"A Ray","year":"2024","unstructured":"Ray, A., Sarkar, S., Schwenker, F., Sarkar, R.: Decoding skin cancer classification: perspectives, insights, and advances through researchers\u2019 lens. Sci. Rep. 14(1), 30542 (2024)","journal-title":"Sci. Rep."},{"key":"4187_CR15","doi-asserted-by":"crossref","unstructured":"Matsuzaka, Y., Yashiro, R.: The diagnostic classification of the pathological image using computer vision. Algorithms 18(2), (2025)","DOI":"10.3390\/a18020096"},{"key":"4187_CR16","first-page":"5325","volume":"39","author":"Y Lin","year":"2025","unstructured":"Lin, Y., Wang, W., Luo, X., Wu, Z., Liu, C., Wen, J., Xu, Y.: Deep hierarchies and invariant disease-indicative feature learning for computer aided diagnosis of multiple fundus diseases. Proc. AAAI Conf. Artif. Intell. 39, 5325\u20135333 (2025)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"4187_CR17","doi-asserted-by":"publisher","first-page":"6308","DOI":"10.1109\/JSTARS.2020.3026724","volume":"13","author":"M Sheykhmousa","year":"2020","unstructured":"Sheykhmousa, M., Mahdianpari, M., Ghanbari, H., Mohammadimanesh, F., Ghamisi, P., Homayouni, S.: Support vector machine versus random forest for remote sensing image classification: a meta-analysis and systematic review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 13, 6308\u20136325 (2020)","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"4187_CR18","doi-asserted-by":"crossref","unstructured":"Dinesh, P., Vickram, A., Kalyanasundaram, P.: \u201cMedical image prediction for diagnosis of breast cancer disease comparing the machine learning algorithms: Svm, knn, logistic regression, random forest and decision tree to measure accuracy,\u201d In: AIP Conference Proceedings, vol. 2853, p. 020140, AIP Publishing LLC, (2024)","DOI":"10.1063\/5.0203746"},{"issue":"4","key":"4187_CR19","doi-asserted-by":"publisher","first-page":"2793","DOI":"10.1007\/s11831-023-09898-w","volume":"30","author":"SS Kshatri","year":"2023","unstructured":"Kshatri, S.S., Singh, D.: Convolutional neural network in medical image analysis: a review. Arch. Comput. Method. Eng. 30(4), 2793\u20132810 (2023)","journal-title":"Arch. Comput. Method. Eng."},{"issue":"5","key":"4187_CR20","doi-asserted-by":"publisher","first-page":"551","DOI":"10.3390\/diagnostics15050551","volume":"15","author":"M Alruwaili","year":"2025","unstructured":"Alruwaili, M., Mohamed, M.: An integrated deep learning model with efficientnet and resnet for accurate multi-class skin disease classification. Diagnostics 15(5), 551 (2025)","journal-title":"Diagnostics"},{"key":"4187_CR21","doi-asserted-by":"crossref","unstructured":"Reshmy, A., Praveen, D., Senthilselvi, A., et al., \u201cEfficient tamil character recognition: A comparative study of learning-based algorithms with resnet-50, densenet121, and efficientnet architectures,\u201d In: 2024 International Conference on Recent Innovation in Smart and Sustainable Technology (ICRISST), pp. 1\u20135, IEEE, (2024)","DOI":"10.1109\/ICRISST59181.2024.10921908"},{"issue":"1","key":"4187_CR22","doi-asserted-by":"publisher","first-page":"259","DOI":"10.31154\/cogito.v10i1.706.680-691","volume":"10","author":"GA Sandag","year":"2024","unstructured":"Sandag, G.A., Kabo, D.T.: Comparative analysis of lung cancer classification models using efficientnet and resnet on ct-scan lung images. CogITo Smart J. 10(1), 259\u2013270 (2024)","journal-title":"CogITo Smart J."},{"issue":"1","key":"4187_CR23","doi-asserted-by":"publisher","first-page":"15013","DOI":"10.1038\/s41598-024-64982-w","volume":"14","author":"S Liu","year":"2024","unstructured":"Liu, S., Yue, W., Guo, Z., Wang, L.: Multi-branch cnn and grouping cascade attention for medical image classification. Sci. Rep. 14(1), 15013 (2024)","journal-title":"Sci. Rep."},{"key":"4187_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2025.105175","volume":"162","author":"Y Wang","year":"2025","unstructured":"Wang, Y., Wang, H., Zhang, F.: Medical image segmentation with an emphasis on prior convolution and channel multi-branch attention. Dig. Sig. Proc. 162, 105175 (2025)","journal-title":"Dig. Sig. Proc."},{"key":"4187_CR25","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1016\/B978-0-12-817440-1.00003-6","volume":"1","author":"RJ Chalakkal","year":"2020","unstructured":"Chalakkal, R.J., Abdulla, W.H., Hong, S.C., et al.: Fundus retinal image analyses for screening and diagnosing diabetic retinopathy, macular edema, and glaucoma disorders. Diab. Fundus OCT 1, 59\u2013111 (2020)","journal-title":"Diab. Fundus OCT"},{"issue":"2","key":"4187_CR26","doi-asserted-by":"publisher","first-page":"286","DOI":"10.1007\/s11390-024-3679-2","volume":"39","author":"Y Wen","year":"2024","unstructured":"Wen, Y., Wu, Y.-L., Bi, L., Shi, W.-Z., Liu, X.-X., Xu, Y.-P., Xu, X., Cao, W.-M., Feng, D.D.: A transformer-assisted cascade learning network for choroidal vessel segmentation. J. Comput. Sci. Technol. 39(2), 286\u2013304 (2024)","journal-title":"J. Comput. Sci. Technol."},{"key":"4187_CR27","doi-asserted-by":"crossref","unstructured":"Sheiba, S., Neelakantappa, M., Shaik, A.: \u201cSegmentation of retinal blood vessels and optic disc using deep neural networks: State-of-the-art review,\u201d Revolutionizing Healthcare 5.0: The Power of Generative AI: Advancements in Patient Care Through Generative AI Algorithms, pp. 139\u2013152, (2025)","DOI":"10.1007\/978-3-031-75771-6_9"},{"key":"4187_CR28","doi-asserted-by":"publisher","first-page":"54190","DOI":"10.1109\/ACCESS.2021.3070685","volume":"9","author":"MT Al-Antary","year":"2021","unstructured":"Al-Antary, M.T., Arafa, Y.: Multi-scale attention network for diabetic retinopathy classification. IEEE Access 9, 54190\u201354200 (2021)","journal-title":"IEEE Access"},{"key":"4187_CR29","doi-asserted-by":"publisher","DOI":"10.1016\/j.measurement.2023.113553","volume":"222","author":"G Gao","year":"2023","unstructured":"Gao, G., Li, J., Yang, L., Liu, Y.: A multi-scale global attention network for blood vessel segmentation from fundus images. Measurement 222, 113553 (2023)","journal-title":"Measurement"},{"issue":"5","key":"4187_CR30","doi-asserted-by":"publisher","first-page":"2093","DOI":"10.1364\/BOE.558532","volume":"16","author":"G Ni","year":"2025","unstructured":"Ni, G., Cao, K., Qin, X., Zeng, X., Wu, R., Wan, L., Zhong, J., Liu, Y.: Advanced 3d retinal lesion segmentation using channel-spatial attention-guided multi-scale feature aggregation. Biomed. Opt. Express 16(5), 2093\u20132110 (2025)","journal-title":"Biomed. Opt. Express"},{"key":"4187_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2023.105534","volume":"87","author":"X Huo","year":"2024","unstructured":"Huo, X., Sun, G., Tian, S., Wang, Y., Yu, L., Long, J., Zhang, W., Li, A.: Hifuse: hierarchical multi-scale feature fusion network for medical image classification. Biomed. Signal Process. Control 87, 105534 (2024)","journal-title":"Biomed. Signal Process. Control"},{"key":"4187_CR32","unstructured":"Fang, M., Wang, Z., Pan, S., Feng, X., Zhao, Y., Hou, D., Wu, L., Xie, X., Zhang, X.-Y., Tian, J., et al.: Large models in medical imaging: Advances and prospects. Chinese Medical Journal 10\u20131097 (2025)"},{"key":"4187_CR33","doi-asserted-by":"publisher","first-page":"1991","DOI":"10.1109\/TMM.2022.3141933","volume":"25","author":"Y Qiu","year":"2022","unstructured":"Qiu, Y., Liu, Y., Chen, Y., Zhang, J., Zhu, J., Xu, J.: A2s ppnet: attentive atrous spatial pyramid pooling network for salient object detection. IEEE Trans. Multimedia 25, 1991\u20132006 (2022)","journal-title":"IEEE Trans. Multimedia"},{"issue":"8","key":"4187_CR34","doi-asserted-by":"publisher","DOI":"10.1088\/1361-6501\/ac68d2","volume":"33","author":"Z Zheng","year":"2022","unstructured":"Zheng, Z., Hu, Y., Zhang, Y., Yang, H., Qiao, Y., Qu, Z., Huang, Y.: Casppnet: a chained atrous spatial pyramid pooling network for steel defect detection. Meas. Sci. Technol. 33(8), 085403 (2022)","journal-title":"Meas. Sci. Technol."},{"key":"4187_CR35","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107622","volume":"110","author":"X Lian","year":"2021","unstructured":"Lian, X., Pang, Y., Han, J., Pan, J.: Cascaded hierarchical atrous spatial pyramid pooling module for semantic segmentation. Pattern Recogn. 110, 107622 (2021)","journal-title":"Pattern Recogn."},{"issue":"1","key":"4187_CR36","doi-asserted-by":"publisher","first-page":"97","DOI":"10.3390\/electronics12010097","volume":"12","author":"MA Saleh","year":"2022","unstructured":"Saleh, M.A., Ali, A.A., Ahmed, K., Sarhan, A.M.: A brief analysis of multimodal medical image fusion techniques. Electronics 12(1), 97 (2022)","journal-title":"Electronics"},{"key":"4187_CR37","first-page":"21","volume":"2","author":"Y Li","year":"2021","unstructured":"Li, Y., Zhao, J., Lv, Z., Li, J.: Medical image fusion method by deep learning. Int. J. Cog. Comput. Eng. 2, 21\u201329 (2021)","journal-title":"Int. J. Cog. Comput. Eng."},{"issue":"8","key":"4187_CR38","doi-asserted-by":"publisher","first-page":"4111","DOI":"10.1109\/JBHI.2022.3171523","volume":"26","author":"W Wang","year":"2022","unstructured":"Wang, W., Li, X., Xu, Z., Yu, W., Zhao, J., Ding, D., Chen, Y.: Learning two-stream cnn for multi-modal age-related macular degeneration categorization. IEEE J. Biomed. Health Inform. 26(8), 4111\u20134122 (2022)","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"4187_CR39","doi-asserted-by":"crossref","unstructured":"Zhu, L., Wang, X., Ke, Z., Zhang, W., Lau, R. W.: \u201cBiformer: Vision transformer with bi-level routing attention,\u201d In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 10323\u201310333, (2023)","DOI":"10.1109\/CVPR52729.2023.00995"},{"key":"4187_CR40","doi-asserted-by":"crossref","unstructured":"Woo, S., Debnath, S., Hu, R., Chen, X., Liu, Z., Kweon, I. S., Xie, S.: \u201cConvnext v2: Co-designing and scaling convnets with masked autoencoders,\u201d In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 16133\u201316142, (2023)","DOI":"10.1109\/CVPR52729.2023.01548"},{"key":"4187_CR41","unstructured":"Yue, Y., Li, Z.: \u201cMedmamba: Vision mamba for medical image classification,\u201d arXiv preprint arXiv:2403.03849, (2024)"},{"key":"4187_CR42","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S. et al., \u201cAn image is worth 16x16 words: Transformers for image recognition at scale,\u201d arXiv preprint arXiv:2010.11929, (2020)"},{"key":"4187_CR43","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: \u201cSwin transformer: Hierarchical vision transformer using shifted windows,\u201d In: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 10012\u201310022, (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"4187_CR44","doi-asserted-by":"crossref","unstructured":"Liu, Z., Hu, H., Lin, Y., Yao, Z., Xie, Z., Wei, Y., Ning, J., Cao, Y., Zhang, Z., Dong, L., et al., \u201cSwin transformer v2: Scaling up capacity and resolution,\u201d In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 12009\u201312019, (2022)","DOI":"10.1109\/CVPR52688.2022.01170"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04187-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04187-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04187-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T13:17:10Z","timestamp":1763644630000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04187-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,31]]},"references-count":44,"journal-issue":{"issue":"15","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4187"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04187-y","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,31]]},"assertion":[{"value":"21 August 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 October 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}