{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T02:48:36Z","timestamp":1773802116095,"version":"3.50.1"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2024,8,18]],"date-time":"2024-08-18T00:00:00Z","timestamp":1723939200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,8,18]],"date-time":"2024-08-18T00:00:00Z","timestamp":1723939200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Key Areas Research and Development Program of Guangzhou Grant","award":["92267107"],"award-info":[{"award-number":["92267107"]}]},{"name":"Key Areas Research and Development Program of Guangzhou Grant","award":["92267107"],"award-info":[{"award-number":["92267107"]}]},{"DOI":"10.13039\/501100012245","name":"Science and Technology Planning Project of Guangdong Province","doi-asserted-by":"publisher","award":["2021B0101220006"],"award-info":[{"award-number":["2021B0101220006"]}],"id":[{"id":"10.13039\/501100012245","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012245","name":"Science and Technology Planning Project of Guangdong Province","doi-asserted-by":"publisher","award":["2021B0101220006"],"award-info":[{"award-number":["2021B0101220006"]}],"id":[{"id":"10.13039\/501100012245","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Science and Technology Projects in Guangzhou","award":["202201011706"],"award-info":[{"award-number":["202201011706"]}]},{"name":"Science and Technology Projects in Guangzhou","award":["202201011706"],"award-info":[{"award-number":["202201011706"]}]},{"name":"Key Areas Research and Development Program of Guangzhou","award":["2023B01J0029"],"award-info":[{"award-number":["2023B01J0029"]}]},{"name":"Key Areas Research and Development Program of Guangzhou","award":["2023B01J0029"],"award-info":[{"award-number":["2023B01J0029"]}]},{"name":"Science and Technology Research in key areas in Foshan","award":["2020001006832"],"award-info":[{"award-number":["2020001006832"]}]},{"name":"Science and Technology Research in key areas in Foshan","award":["2020001006832"],"award-info":[{"award-number":["2020001006832"]}]},{"name":"Key Area Research and Development Program of Guangdong","award":["2019B010153002"],"award-info":[{"award-number":["2019B010153002"]}]},{"name":"Key Area Research and Development Program of Guangdong","award":["2019B010153002"],"award-info":[{"award-number":["2019B010153002"]}]},{"name":"Science and Technology Projects of Guangzhou","award":["202007040006"],"award-info":[{"award-number":["202007040006"]}]},{"name":"Science and Technology Projects of Guangzhou","award":["202007040006"],"award-info":[{"award-number":["202007040006"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,11]]},"DOI":"10.1007\/s11760-024-03484-8","type":"journal-article","created":{"date-parts":[[2024,8,18]],"date-time":"2024-08-18T12:01:45Z","timestamp":1723982505000},"page":"8427-8443","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":29,"title":["RM-UNet: UNet-like Mamba with rotational SSM module for medical image segmentation"],"prefix":"10.1007","volume":"18","author":[{"given":"Hao","family":"Tang","sequence":"first","affiliation":[]},{"given":"Guoheng","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Lianglun","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Xiaochen","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Qi","family":"Tao","sequence":"additional","affiliation":[]},{"given":"Xuhang","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Guo","family":"Zhong","sequence":"additional","affiliation":[]},{"given":"Xiaohui","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,18]]},"reference":[{"key":"3484_CR1","doi-asserted-by":"crossref","unstructured":"Bray, F., Laversanne, M., Sung, H., Ferlay, J., Siegel, R.L., Soerjomataram, I., Jemal, A.: Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. (2024)","DOI":"10.3322\/caac.21834"},{"issue":"10","key":"3484_CR2","first-page":"1995","volume":"3361","author":"Y LeCun","year":"1995","unstructured":"LeCun, Y., Bengio, Y., et al.: Convolutional networks for images, speech, and time series. Handb. Brain Theory Neural Netw. 3361(10), 1995 (1995)","journal-title":"Handb. Brain Theory Neural Netw."},{"key":"3484_CR3","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention\u2014MICCAI 2015, pp. 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"3484_CR4","doi-asserted-by":"crossref","unstructured":"Chen, X., Cun, X., Pun, C., Wang, S.: ShaDocNet: learning patial-aware tokens in transformer for document shadow removal. In: International Conference on Acoustics, Speech, and Signal Processing, pp. 1\u20135 (2023)","DOI":"10.1109\/ICASSP49357.2023.10095403"},{"key":"3484_CR5","doi-asserted-by":"crossref","unstructured":"Luo, S., Chen, X., Chen, W., Li, Z., Wang, S., Pun, C-M.: Devignet: high-resolution vignetting removal via a dual aggregated fusion transformer with adaptive channel expansion. In: AAAI Conference on Artificial Intelligence, pp. 4000\u20134008 (2024)","DOI":"10.1609\/aaai.v38i5.28193"},{"key":"3484_CR6","unstructured":"Vaswani, A., Shazeer, N.M., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. In: Neural Information Processing Systems (2017)"},{"key":"3484_CR7","unstructured":"Luo, W., Li, Y., Urtasun, R., Zemel, R.: Understanding the effective receptive field in deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 29 (2016)"},{"key":"3484_CR8","doi-asserted-by":"publisher","first-page":"568","DOI":"10.1109\/JBHI.2019.2912935","volume":"24","author":"S Guan","year":"2018","unstructured":"Guan, S., Khan, A.A., Sikdar, S., Chitnis, P.V.: Fully dense unet for 2-d sparse photoacoustic tomography artifact removal. IEEE J. Biomed. Health Inform. 24, 568\u2013576 (2018)","journal-title":"IEEE J. Biomed. Health Inform."},{"key":"3484_CR9","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1016\/j.neunet.2019.08.025","volume":"121","author":"N Ibtehaz","year":"2020","unstructured":"Ibtehaz, N., Rahman, M.S.: MultiResUNet:Rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural Netw. 121, 74\u201387 (2020)","journal-title":"Neural Netw."},{"key":"3484_CR10","unstructured":"Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., Lu, L., Yuille, A.L., Zhou, Y.: Transunet: Transformers make strong encoders for medical image segmentation. arXiv:2102.04306 (2021)"},{"key":"3484_CR11","doi-asserted-by":"crossref","unstructured":"Valanarasu, J.M.J., Oza, P., Hacihaliloglu, I., Patel, V.M.: Medical transformer: Gated axial-attention for medical image segmentation. In: Medical Image Computing and Computer Assisted Intervention\u2014MICCAI 2021, pp. 36\u201346 (2021)","DOI":"10.1007\/978-3-030-87193-2_4"},{"key":"3484_CR12","doi-asserted-by":"crossref","unstructured":"Cao, H., Wang, Y., Chen, J., Jiang, D., Zhang, X., Tian, Q., Wang, M.: Swin-UNet: Unet-like pure transformer for medical image segmentation. In: Computer Vision\u2014ECCV 2022 Workshops, pp. 205\u2013218 (2023)","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"3484_CR13","doi-asserted-by":"crossref","unstructured":"Hatamizadeh, A., Tang, Y., Nath, V., Yang, D., Myronenko, A., Landman, B., Roth, H.R., Xu, D.: Unetr: Transformers for 3d medical image segmentation. In: 2022 IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 1748\u20131758 (2022)","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"3484_CR14","first-page":"1","volume":"71","author":"A Lin","year":"2022","unstructured":"Lin, A., Chen, B., Xu, J., Zhang, Z., Lu, G., Zhang, D.: Ds-transunet: Dual swin transformer u-net for medical image segmentation. IEEE Trans. Instrum. Meas. 71, 1\u201315 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"3484_CR15","doi-asserted-by":"publisher","first-page":"4036","DOI":"10.1109\/TIP.2023.3293771","volume":"32","author":"H-Y Zhou","year":"2023","unstructured":"Zhou, H.-Y., Guo, J., Zhang, Y., Han, X., Yu, L., Wang, L., Yu, Y.: nnformer: Volumetric medical image segmentation via a 3d transformer. IEEE Trans. Image Process. 32, 4036\u20134045 (2023)","journal-title":"IEEE Trans. Image Process."},{"key":"3484_CR16","unstructured":"Gu, A., Goel, K., R\u2019e, C.: Efficiently modeling long sequences with structured state spaces. arXiv:2111.00396 (2021)"},{"key":"3484_CR17","unstructured":"Gu, A., Dao, T.: Mamba: Linear-time sequence modeling with selective state spaces. arXiv:2312.00752 (2023)"},{"key":"3484_CR18","unstructured":"Ma, J., Li, F., Wang, B.: U-mamba: Enhancing long-range dependency for biomedical image segmentation. arXiv:2401.04722 (2024)"},{"key":"3484_CR19","unstructured":"Zhu, L., Liao, B., Zhang, Q., Wang, X., Liu, W., Wang, X.: Vision mamba: Efficient visual representation learning with bidirectional state space model. arXiv:2401.09417 (2024)"},{"key":"3484_CR20","unstructured":"Liu, Y., Tian, Y., Zhao, Y., Yu, H., Xie, L., Wang, Y., Ye, Q., Liu, Y.: Vmamba: Visual state space model. arXiv:2401.10166 (2024)"},{"key":"3484_CR21","unstructured":"Yang, C., Chen, Z., Espinosa, M., Ericsson, L., Wang, Z., Liu, J., Crowley, E.J.: Plainmamba: Improving non-hierarchical mamba in visual recognition. arXiv:2403.17695 (2024)"},{"key":"3484_CR22","doi-asserted-by":"crossref","unstructured":"Zhao, S., Chen, H., Zhang, X.-l., Xiao, P., Bai, L., Ouyang, W.: Rs-mamba for large remote sensing image dense prediction. arXiv:2404.02668 (2024)","DOI":"10.1109\/TGRS.2024.3425540"},{"key":"3484_CR23","doi-asserted-by":"crossref","unstructured":"Wang, Z., Zheng, J.-Q., Zhang, Y., Cui, G., Li, L.: Mamba-unet: Unet-like pure visual mamba for medical image segmentation. arXiv:2402.05079 (2024)","DOI":"10.1109\/AINIT61980.2024.10581519"},{"key":"3484_CR24","doi-asserted-by":"crossref","unstructured":"Ruan, J., Xiang, S.: Vm-unet: Vision mamba unet for medical image segmentation. arXiv:2402.02491 (2024)","DOI":"10.1109\/ISBI53787.2023.10230496"},{"key":"3484_CR25","doi-asserted-by":"crossref","unstructured":"Zhang, M., Yu, Y., Gu, L., Lin, T., Tao, X.: Vm-unet-v2 rethinking vision mamba unet for medical image segmentation. arXiv:2403.09157 (2024)","DOI":"10.1007\/978-981-97-5128-0_27"},{"key":"3484_CR26","doi-asserted-by":"crossref","unstructured":"Liu, J., Yang, H., Zhou, H.-Y., Xi, Y., Yu, L., Yu, Y., Liang, Y., Shi, G., Zhang, S., Zheng, H., Wang, S.: Swin-umamba: Mamba-based unet with imagenet-based pretraining. arXiv:2402.03302 (2024)","DOI":"10.1007\/978-3-031-72114-4_59"},{"key":"3484_CR27","doi-asserted-by":"crossref","unstructured":"Ye, Z., Chen, T., Wang, F., Zhang, H., Li, G., Zhang, L.: P-mamba: Marrying perona malik diffusion with mamba for efficient pediatric echocardiographic left ventricular segmentation. arXiv:2402.08506 (2024)","DOI":"10.21203\/rs.3.rs-4963037\/v1"},{"key":"3484_CR28","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"3484_CR29","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: Cbam: Convolutional block attention module. In: Computer Vision \u2013 ECCV 2018, pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"3484_CR30","doi-asserted-by":"crossref","unstructured":"Fu, J., Liu, J., Tian, H., Fang, Z., Lu, H.: Dual attention network for scene segmentation. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3141\u20133149 (2018)","DOI":"10.1109\/CVPR.2019.00326"},{"key":"3484_CR31","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13708\u201313717 (2021)","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"3484_CR32","unstructured":"Park, J., Woo, S., Lee, J.-Y., Kweon, I.-S.: Bam: Bottleneck attention module. arXiv:1807.06514 (2018)"},{"key":"3484_CR33","doi-asserted-by":"crossref","unstructured":"Misra, D., Nalamada, T., Arasanipalai, A.U., Hou, Q.: Rotate to attend: Convolutional triplet attention module. In: 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 3138\u20133147 (2021)","DOI":"10.1109\/WACV48630.2021.00318"},{"key":"3484_CR34","unstructured":"Lee, C.-Y., Xie, S., Gallagher, P., Zhang, Z., Tu, Z.: Deeply-Supervised Nets. In: Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 562\u2013570 (2015)"},{"key":"3484_CR35","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: Deep learning with depthwise separable convolutions. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 1800\u20131807 (2016)","DOI":"10.1109\/CVPR.2017.195"},{"key":"3484_CR36","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.neunet.2017.12.012","volume":"107","author":"S Elfwing","year":"2018","unstructured":"Elfwing, S., Uchibe, E., Doya, K.: Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 107, 3\u201311 (2018)","journal-title":"Neural Netw."},{"key":"3484_CR37","doi-asserted-by":"crossref","unstructured":"Codella, N.C.F., Gutman, D., Celebi, M.E., Helba, B., Marchetti, M.A., Dusza, S.W., Kalloo, A., Liopyris, K., Mishra, N., Kittler, H., Halpern, A.: Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (isbi), hosted by the international skin imaging collaboration (isic). In: 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI), pp. 168\u2013172 (2018)","DOI":"10.1109\/ISBI.2018.8363547"},{"key":"3484_CR38","unstructured":"Codella, N.C.F., Rotemberg, V.M., Tschandl, P., Celebi, M.E., Dusza, S.W., Gutman, D., Helba, B., Kalloo, A., Liopyris, K., Marchetti, M.A., Kittler, H., Halpern, A.C.: Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (isic). arXiv:1902.03368 (2019)"},{"key":"3484_CR39","doi-asserted-by":"crossref","unstructured":"Fan, D.-P., Ji, G.-P., Zhou, T., Chen, G., Fu, H., Shen, J., Shao, L.: Pranet: Parallel reverse attention network for polyp segmentation. In: Medical Image Computing and Computer Assisted Intervention - MICCAI 2020: 23rd International Conference, Lima, Peru, October 4-8, 2020, Proceedings, Part VI, pp. 263\u2013273 (2020)","DOI":"10.1007\/978-3-030-59725-2_26"},{"key":"3484_CR40","doi-asserted-by":"crossref","unstructured":"Jha, D., Smedsrud, P.H., Riegler, M.A., Halvorsen, P., Lange, T., Johansen, D., Johansen, H.D.: Kvasir-seg: A segmented polyp dataset. In: MultiMedia Modeling, pp. 451\u2013462 (2020)","DOI":"10.1007\/978-3-030-37734-2_37"},{"key":"3484_CR41","doi-asserted-by":"crossref","unstructured":"Bernal, J., S\u00e1nchez, F.J., Fern\u00e1ndez-Esparrach, G., Gil, D., Rodr\u00edguez, C., Vilari\u00f1o, F.: Wm-dova maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Computerized Medical Imaging and Graphics 43, 99\u2013111 (2015)","DOI":"10.1016\/j.compmedimag.2015.02.007"},{"issue":"2","key":"3484_CR42","doi-asserted-by":"publisher","first-page":"630","DOI":"10.1109\/TMI.2015.2487997","volume":"35","author":"N Tajbakhsh","year":"2016","unstructured":"Tajbakhsh, N., Gurudu, S.R., Liang, J.: Automated polyp detection in colonoscopy videos using shape and context information. IEEE Trans. Med. Imaging 35(2), 630\u2013644 (2016)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"3484_CR43","doi-asserted-by":"crossref","unstructured":"V\u00e1zquez, D., Bernal, J., S\u00e1nchez, F.J., Fern\u00e1ndez-Esparrach, G., L\u00f3pez, A.M., Romero, A., Drozdzal, M., Courville, A.C.: A benchmark for endoluminal scene segmentation of colonoscopy images. CoRR abs\/1612.00799 (2016)","DOI":"10.1155\/2017\/4037190"},{"key":"3484_CR44","doi-asserted-by":"crossref","unstructured":"Jha, D., Ali, S., Emanuelsen, K., Hicks, S.A., Thambawita, V., Garcia-Ceja, E., Riegler, M.A., Lange, T., Schmidt, P.T., Johansen, H.D., Johansen, D., Halvorsen, P.: Kvasir-instrument: Diagnostic and therapeutic tool segmentation dataset in gastrointestinal endoscopy. In: MultiMedia Modeling, pp. 218\u2013229 (2021)","DOI":"10.1007\/978-3-030-67835-7_19"},{"key":"3484_CR45","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Rahman\u00a0Siddiquee, M.M., Tajbakhsh, N., Liang, J.: Unet++: A nested u-net architecture for medical image segmentation. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, pp. 3\u201311 (2018)","DOI":"10.1007\/978-3-030-00889-5_1"},{"key":"3484_CR46","unstructured":"Oktay, O., Schlemper, J., Folgoc, L.L., Lee, M.J., Heinrich, M.P., Misawa, K., Mori, K., McDonagh, S.G., Hammerla, N.Y., Kainz, B., Glocker, B., Rueckert, D.: Attention u-net: Learning where to look for the pancreas. arXiv:1804.03999 (2018)"},{"key":"3484_CR47","unstructured":"Peng, Y., Sonka, M., Chen, D.Z.: U-net v2: Rethinking the skip connections of u-net for medical image segmentation. arXiv:2311.17791 (2023)"},{"key":"3484_CR48","first-page":"14","volume-title":"Medical image computing and computer assisted intervention\u2014MICCAI 2021","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Liu, H., Hu, Q.: Transfuse: Fusing transformers and CNNs for medical image segmentation. In: Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C. (eds.) Medical image computing and computer assisted intervention\u2014MICCAI 2021, pp. 14\u201324. Springer, Cham (2021)"},{"key":"3484_CR49","doi-asserted-by":"crossref","unstructured":"Aghdam, E.K., Azad, R., Zarvani, M., Merhof, D.: Attention swin u-net: Cross-contextual attention mechanism for skin lesion segmentation. In: 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI), pp. 1\u20135 (2023). IEEE","DOI":"10.1109\/ISBI53787.2023.10230337"},{"key":"3484_CR50","doi-asserted-by":"publisher","first-page":"9150015","DOI":"10.26599\/AIR.2023.9150015","volume":"2","author":"B Dong","year":"2023","unstructured":"Dong, B., Wang, W., Fan, D.-P., Li, J., Fu, H., Shao, L.: Polyp-pvt: Polyp segmentation with pyramid vision transformers. CAAI Artif. Intell. Res. 2, 9150015 (2023)","journal-title":"CAAI Artif. Intell. Res."},{"key":"3484_CR51","first-page":"14","volume-title":"Medical image computing and computer assisted intervention\u2014MICCAI 2023","author":"S Hu","year":"2023","unstructured":"Hu, S., Liao, Z., Xia, Y.: Devil is in channels: Contrastive single domain generalization for medical image segmentation. In: Greenspan, H., Madabhushi, A., Mousavi, P., Salcudean, S., Duncan, J., Syeda-Mahmood, T., Taylor, R. (eds.) Medical image computing and computer assisted intervention\u2014MICCAI 2023, pp. 14\u201323. Springer, Cham (2023)"},{"key":"3484_CR52","unstructured":"Gao, Y., Zhou, M., Liu, D., Metaxas, D.N.: A multi-scale transformer for medical image segmentation: Architectures, model efficiency, and benchmarks. arXiv:2203.00131 (2022)"},{"key":"3484_CR53","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. In: 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019 (2019)"},{"key":"3484_CR54","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings (2017)"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03484-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-024-03484-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03484-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,26]],"date-time":"2024-11-26T20:25:50Z","timestamp":1732652750000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-024-03484-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,18]]},"references-count":54,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2024,11]]}},"alternative-id":["3484"],"URL":"https:\/\/doi.org\/10.1007\/s11760-024-03484-8","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,8,18]]},"assertion":[{"value":"26 June 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 July 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 July 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 August 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}