{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T07:03:58Z","timestamp":1766127838631,"version":"3.48.0"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,10,17]],"date-time":"2025-10-17T00:00:00Z","timestamp":1760659200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,17]],"date-time":"2025-10-17T00:00:00Z","timestamp":1760659200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s00530-025-02031-w","type":"journal-article","created":{"date-parts":[[2025,10,17]],"date-time":"2025-10-17T09:41:38Z","timestamp":1760694098000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Ecca-unet: edge-aware and channel-enhanced cross-attention network for medical image segmentation"],"prefix":"10.1007","volume":"31","author":[{"given":"Tao","family":"Liu","sequence":"first","affiliation":[]},{"given":"JianLong","family":"Hu","sequence":"additional","affiliation":[]},{"given":"MengYu","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,17]]},"reference":[{"key":"2031_CR1","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-assisted intervention-MICCAI 2015: 18th International Conference, Munich, Germany, October 5\u20139, 2015, Proceedings, Part III 18, pp. 234\u2013241. Springer (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"2031_CR2","unstructured":"Howard, A.G.: Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"issue":"6","key":"2031_CR3","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Commun. ACM 60(6), 84\u201390 (2017). https:\/\/doi.org\/10.1145\/3065386","journal-title":"Commun. ACM"},{"key":"2031_CR4","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"2031_CR5","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"2031_CR6","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der\u00a0Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"2031_CR7","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16 x 16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"2031_CR8","unstructured":"Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., Lu, L., Yuille, A.L., Zhou, Y.: Transunet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:2102.04306 (2021)"},{"key":"2031_CR9","doi-asserted-by":"crossref","unstructured":"Cao, H., Wang, Y., Chen, J., Jiang, D., Zhang, X., Tian, Q., Wang, M.: Swin-unet: unet-like pure transformer for medical image segmentation. In: European Conference on Computer Vision, pp. 205\u2013218. Springer (2022)","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"2031_CR10","doi-asserted-by":"crossref","unstructured":"Wenxuan, W., Chen, C., Meng, D., Hong, Y., Sen, Z., Jiangyun, L.: Transbts: Multimodal brain tumor segmentation using transformer. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 109\u2013119. Springer (2021)","DOI":"10.1007\/978-3-030-87193-2_11"},{"key":"2031_CR11","doi-asserted-by":"crossref","unstructured":"Dong, X., Bao, J., Chen, D., Zhang, W., Yu, N., Yuan, L., Chen, D., Guo, B.: Cswin transformer: a general vision transformer backbone with cross-shaped windows. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12124\u201312134 (2022)","DOI":"10.1109\/CVPR52688.2022.01181"},{"key":"2031_CR12","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-00889-5_1","volume-title":"Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support","author":"Z Zhou","year":"2018","unstructured":"Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: Unet++: a nested u-net architecture for medical image segmentation. In: Stoyanov, D., Taylor, Z., Carneiro, G., Syeda-Mahmood, T., Martel, A., Maier-Hein, L., Tavares, J.M.R.S., Bradley, A., Papa, J.P., Belagiannis, V., Nascimento, J.C., Lu, Z., Conjeti, S., Moradi, M., Greenspan, H., Madabhushi, A. (eds.) Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, pp. 3\u201311. Springer, Cham (2018)"},{"key":"2031_CR13","doi-asserted-by":"crossref","unstructured":"Huang, H., Lin, L., Tong, R., Hu, H., Zhang, Q., Iwamoto, Y., Han, X., Chen, Y.-W., Wu, J.: Unet 3+: A full-scale connected unet for medical image segmentation. In: ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1055\u20131059 (2020)","DOI":"10.1109\/ICASSP40776.2020.9053405"},{"key":"2031_CR14","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1016\/j.neunet.2019.08.025","volume":"121","author":"N Ibtehaz","year":"2020","unstructured":"Ibtehaz, N., Rahman, M.S.: Multiresunet: rethinking the u-net architecture for multimodal biomedical image segmentation. Neural Netw. 121, 74\u201387 (2020)","journal-title":"Neural Netw."},{"key":"2031_CR15","first-page":"14","volume-title":"Medical Image Computing and Computer Assisted Intervention-MICCAI 2021","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Liu, H., Hu, Q.: Transfuse: fusing transformers and cnns for medical image segmentation. In: Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C. (eds.) Medical Image Computing and Computer Assisted Intervention-MICCAI 2021, pp. 14\u201324. Springer, Cham (2021)"},{"key":"2031_CR16","first-page":"2441","volume":"36","author":"H Wang","year":"2022","unstructured":"Wang, H., Cao, P., Wang, J., Zaiane, O.R.: Uctransnet: rethinking the skip connections in u-net from a channel-wise perspective with transformer. Proc. AAAI Conf. Artif. Intell. 36, 2441\u20132449 (2022)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"2031_CR17","doi-asserted-by":"publisher","first-page":"8180","DOI":"10.1109\/JSTARS.2021.3103176","volume":"14","author":"H Gao","year":"2021","unstructured":"Gao, H., Zhang, Y., Chen, Z., Li, C.: A multiscale dual-branch feature fusion and attention network for hyperspectral images classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 14, 8180\u20138192 (2021)","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"issue":"7","key":"2031_CR18","doi-asserted-by":"publisher","first-page":"2969","DOI":"10.1007\/s00371-022-02503-4","volume":"39","author":"G Wang","year":"2023","unstructured":"Wang, G., Gan, X., Cao, Q., Zhai, Q.: Mfanet: multi-scale feature fusion network with attention mechanism. Vis. Comput. 39(7), 2969\u20132980 (2023)","journal-title":"Vis. Comput."},{"issue":"4","key":"2031_CR19","doi-asserted-by":"publisher","first-page":"979","DOI":"10.3390\/electronics12040979","volume":"12","author":"J Lu","year":"2023","unstructured":"Lu, J., Ren, H., Shi, M., Cui, C., Zhang, S., Emam, M., Li, L.: A novel hybridoma cell segmentation method based on multi-scale feature fusion and dual attention network. Electronics 12(4), 979 (2023)","journal-title":"Electronics"},{"key":"2031_CR20","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Fu, H., Dai, H., Shen, J., Pang, Y., Shao, L.: Et-net: A generic edge-attention guidance network for medical image segmentation. In: Medical Image Computing and Computer Assisted Intervention\u2013MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13\u201317, 2019, Proceedings, Part I 22, pp. 442\u2013450. Springer (2019)","DOI":"10.1007\/978-3-030-32239-7_49"},{"key":"2031_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2024.108265","volume":"172","author":"X Fan","year":"2024","unstructured":"Fan, X., Zhou, J., Jiang, X., Xin, M., Hou, L.: Csap-unet: convolution and self-attention paralleling network for medical image segmentation with edge enhancement. Comput. Biol. Med. 172, 108265 (2024)","journal-title":"Comput. Biol. Med."},{"key":"2031_CR22","doi-asserted-by":"publisher","DOI":"10.3390\/diagnostics13061063","author":"B Lv","year":"2023","unstructured":"Lv, B., Liu, F., Li, Y., Nie, J., Gou, F., Wu, J.: Artificial intelligence-aided diagnosis solution by enhancing the edge features of medical images. Diagnostics (2023). https:\/\/doi.org\/10.3390\/diagnostics13061063","journal-title":"Diagnostics"},{"key":"2031_CR23","first-page":"1","volume":"71","author":"A Lin","year":"2022","unstructured":"Lin, A., Chen, B., Xu, J., Zhang, Z., Lu, G., Zhang, D.: Ds-transunet: dual swin transformer u-net for medical image segmentation. IEEE Trans. Instrum. Meas. 71, 1\u201315 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"5","key":"2031_CR24","doi-asserted-by":"publisher","first-page":"513","DOI":"10.3390\/diagnostics15050513","volume":"15","author":"S Umirzakova","year":"2025","unstructured":"Umirzakova, S., Muksimova, S., Baltayev, J., Cho, Y.I.: Force map-enhanced segmentation of a lightweight model for the early detection of cervical cancer. Diagnostics 15(5), 513 (2025)","journal-title":"Diagnostics"},{"key":"2031_CR25","doi-asserted-by":"crossref","unstructured":"Wang, J., Chen, K., Xu, R., Liu, Z., Loy, C.C., Lin, D.: Carafe: Content-aware reassembly of features. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00310"},{"key":"2031_CR26","doi-asserted-by":"crossref","unstructured":"Xie, S., Tu, Z.: Holistically-nested edge detection. arXiv:1504.06375 (2015)","DOI":"10.1109\/ICCV.2015.164"},{"key":"2031_CR27","unstructured":"Landman, B., Xu, Z., Iglesias, J., Styner, M., Langerak, T., Klein, A.: Miccai multi-atlas labeling beyond the cranial vault-workshop and challenge. In: Proceedings of the MICCAI Multi-Atlas Labeling Beyond Cranial Vault-Workshop Challenge, vol. 5, p. 12 (2015)"},{"issue":"12","key":"2031_CR28","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481\u20132495 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2031_CR29","doi-asserted-by":"crossref","unstructured":"\u00c7i\u00e7ek, \u00d6., Abdulkadir, A., Lienkamp, S.S., Brox, T., Ronneberger, O.: 3d u-net: learning dense volumetric segmentation from sparse annotation. In: Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2016: 19th International Conference, Athens, Greece, October 17\u201321, 2016, Proceedings, Part II 19, pp. 424\u2013432. Springer (2016)","DOI":"10.1007\/978-3-319-46723-8_49"},{"issue":"11","key":"2031_CR30","doi-asserted-by":"publisher","first-page":"2514","DOI":"10.1109\/TMI.2018.2837502","volume":"37","author":"O Bernard","year":"2018","unstructured":"Bernard, O., Lalande, A., Zotti, C., Cervenansky, F., Yang, X., Heng, P.-A., Cetin, I., Lekadir, K., Camara, O., Ballester, M.A.G., et al.: Deep learning techniques for automatic mri cardiac multi-structures segmentation and diagnosis: is the problem solved? IEEE Trans. Med. Imaging 37(11), 2514\u20132525 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"2031_CR31","doi-asserted-by":"publisher","unstructured":"Yurtkulu, S.C., \u015eahin, Y.H., Unal, G.: Semantic segmentation with extended deeplabv3 architecture. In: 2019 27th Signal Processing and Communications Applications Conference (SIU), pp. 1\u20134 (2019). https:\/\/doi.org\/10.1109\/SIU.2019.8806244","DOI":"10.1109\/SIU.2019.8806244"},{"key":"2031_CR32","unstructured":"Oktay, O., Schlemper, J., Folgoc, L.L., Lee, M., Heinrich, M., Misawa, K., Mori, K., McDonagh, S., Hammerla, N.Y., Kainz, B., et al.: Attention u-net: learning where to look for the pancreas. arXiv preprint arXiv:1804.03999 (2018)"},{"key":"2031_CR33","doi-asserted-by":"crossref","unstructured":"Wang, H., Xie, S., Lin, L., Iwamoto, Y., Han, X.-H., Chen, Y.-W., Tong, R.: Mixed transformer u-net for medical image segmentation. In: ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2390\u20132394 (2022)","DOI":"10.1109\/ICASSP43922.2022.9746172"},{"key":"2031_CR34","doi-asserted-by":"crossref","unstructured":"Xu, G., Zhang, X., He, X., Wu, X.: Levit-unet: make faster encoders with;transformer for;medical image segmentation. In: Pattern Recognition and Computer Vision: 6th Chinese Conference. PRCV 2023, Xiamen, China, October 13\u201315, 2023, Proceedings, Part VIII, pp. 42\u201353. Springer, Berlin (2023)","DOI":"10.1007\/978-981-99-8543-2_4"},{"key":"2031_CR35","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102634","volume":"113","author":"X Liu","year":"2025","unstructured":"Liu, X., Gao, P., Yu, T., Wang, F., Yuan, R.-Y.: Cswin-unet: transformer unet with cross-shaped windows for medical image segmentation. Inf. Fusion 113, 102634 (2025)","journal-title":"Inf. Fusion"},{"issue":"5","key":"2031_CR36","doi-asserted-by":"publisher","first-page":"1484","DOI":"10.1109\/TMI.2022.3230943","volume":"42","author":"X Huang","year":"2022","unstructured":"Huang, X., Deng, Z., Li, D., Yuan, X., Fu, Y.: Missformer: an effective transformer for 2d medical image segmentation. IEEE Trans. Med. Imaging 42(5), 1484\u20131494 (2022)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"2031_CR37","doi-asserted-by":"crossref","unstructured":"Hatamizadeh, A., Tang, Y., Nath, V., Yang, D., Myronenko, A., Landman, B., Roth, H.R., Xu, D.: Unetr: transformers for 3d medical image segmentation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 574\u2013584 (2022)","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"2031_CR38","doi-asserted-by":"crossref","unstructured":"Touvron, H., Cord, M., Sablayrolles, A., Synnaeve, G., J\u00e9gou, H.: Going deeper with image transformers. arXiv preprint arXiv:2103.17239 (2021)","DOI":"10.1109\/ICCV48922.2021.00010"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02031-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-02031-w","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-02031-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T06:59:24Z","timestamp":1766127564000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-02031-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,17]]},"references-count":38,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["2031"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-02031-w","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2025,10,17]]},"assertion":[{"value":"21 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 October 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"450"}}