{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T15:43:51Z","timestamp":1775317431580,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":37,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556304","type":"print"},{"value":"9789819556311","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5631-1_18","type":"book-chapter","created":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T07:07:56Z","timestamp":1769497676000},"page":"249-262","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multimodal Co-aware Scale-Spatial Network for\u00a0Medical Image Segmentation"],"prefix":"10.1007","author":[{"given":"Enming","family":"Huang","sequence":"first","affiliation":[]},{"given":"Teng Fei","family":"Gong","sequence":"additional","affiliation":[]},{"given":"Yaxiong","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Shengwu","family":"Xiong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,28]]},"reference":[{"issue":"5","key":"18_CR1","doi-asserted-by":"publisher","first-page":"1484","DOI":"10.1109\/TMI.2022.3230943","volume":"42","author":"X Huang","year":"2022","unstructured":"Huang, X., Deng, Z., Li, D., Yuan, X., Fu, Y.: Missformer: an effective transformer for 2d medical image segmentation. IEEE Trans. Med. Imaging 42(5), 1484\u20131494 (2022)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"12","key":"18_CR2","doi-asserted-by":"publisher","first-page":"3987","DOI":"10.1109\/TMI.2023.3320151","volume":"42","author":"S Jain","year":"2023","unstructured":"Jain, S., et al.: Coinnet: a convolution-involution network with a novel statistical attention for automatic polyp segmentation. IEEE Trans. Med. Imaging 42(12), 3987\u20134000 (2023)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"18_CR3","doi-asserted-by":"crossref","unstructured":"Li, Z., et al.: TFCNs: a CNN-transformer hybrid network for medical image segmentation. In: International Conference on Artificial Neural Networks, pp. 781\u2013792 (2022)","DOI":"10.1007\/978-3-031-15937-4_65"},{"key":"18_CR4","doi-asserted-by":"crossref","unstructured":"Chen, Y., et al.: Striving for simplicity: simple yet effective prior-aware pseudo-labeling for semi-supervised ultrasound image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 604\u2013614 (2024)","DOI":"10.1007\/978-3-031-72114-4_58"},{"key":"18_CR5","doi-asserted-by":"crossref","unstructured":"Li, L., Lian, S., Luo, Z., Wang, B., Li, S.: VCLIPSeg: voxel-wise CLIP-enhanced model for semi-supervised medical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 692\u2013701 (2024)","DOI":"10.1007\/978-3-031-72114-4_66"},{"key":"18_CR6","doi-asserted-by":"crossref","unstructured":"Zhang, S., Zhang, J., Xie, Y., Xia, Y.: TPRO: text-prompting-based weakly supervised histopathology tissue segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 109\u2013118 (2023)","DOI":"10.1007\/978-3-031-43907-0_11"},{"key":"18_CR7","doi-asserted-by":"crossref","unstructured":"Shah, N.A., VS, V., Patel, V.M.: LQMformer: language-aware query mask transformer for referring image segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12903\u201312913 (2024)","DOI":"10.1109\/CVPR52733.2024.01226"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Feng, Y., Chen, W., Gu, X., Xu, X., Zhang, M.: Multi-modal semi-supervised evidential recycle framework for Alzheimer\u2019s disease classification. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 130\u2013140 (2023)","DOI":"10.1007\/978-3-031-43907-0_13"},{"issue":"2","key":"18_CR9","doi-asserted-by":"publisher","first-page":"699","DOI":"10.1109\/TMI.2020.3035253","volume":"40","author":"R Gu","year":"2020","unstructured":"Gu, R., et al.: CA-net: comprehensive attention convolutional neural networks for explainable medical image segmentation. IEEE Trans. Med. Imaging 40(2), 699\u2013711 (2020)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"18_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102710","volume":"85","author":"Z Li","year":"2023","unstructured":"Li, Z., et al.: CAN: context-assisted full Attention Network for brain tissue segmentation. Med. Image Anal. 85, 102710 (2023)","journal-title":"Med. Image Anal."},{"key":"18_CR11","doi-asserted-by":"crossref","unstructured":"Zhou, L., Jiang, Y., Li, W., Hu, J., Zheng, S.: Shape-scale co-awareness network for 3D brain tumor segmentation. IEEE Trans. Med. Imaging (2024)","DOI":"10.1109\/TMI.2024.3368531"},{"key":"18_CR12","doi-asserted-by":"publisher","DOI":"10.1016\/j.compmedimag.2024.102456","volume":"118","author":"Y Huang","year":"2024","unstructured":"Huang, Y., Holcombe, S.A., Wang, S.C., Tang, J.: AFSegNet: few-shot 3D ankle-foot bone segmentation via hierarchical feature distillation and multi-scale attention and fusion. Comput. Med. Imaging Graph. 118, 102456 (2024)","journal-title":"Comput. Med. Imaging Graph."},{"key":"18_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.compmedimag.2024.102370","volume":"114","author":"M Xu","year":"2024","unstructured":"Xu, M., Ma, Q., Zhang, H., Kong, D., Zeng, T.: MEF-UNet: an end-to-end ultrasound image segmentation algorithm based on multi-scale feature extraction and fusion. Comput. Med. Imaging Graph. 114, 102370 (2024)","journal-title":"Comput. Med. Imaging Graph."},{"issue":"9","key":"18_CR14","doi-asserted-by":"publisher","first-page":"2763","DOI":"10.1109\/TMI.2023.3264513","volume":"42","author":"A He","year":"2023","unstructured":"He, A., Wang, K., Li, T., Du, C., Xia, S., Fu, H.: H2Former: an efficient hierarchical hybrid transformer for medical image segmentation. IEEE Trans. Med. Imaging 42(9), 2763\u20132775 (2023)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"Liu, W., et al.: Phtrans: parallelly aggregating global and local representations for medical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 235\u2013244 (2022)","DOI":"10.1007\/978-3-031-16443-9_23"},{"key":"18_CR16","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"4","key":"18_CR17","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/TPAMI.2017.2699184","volume":"40","author":"LC Chen","year":"2017","unstructured":"Chen, L.C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A.L.: DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834\u2013848 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"18_CR18","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1186\/s42492-022-00105-4","volume":"5","author":"W Yuan","year":"2022","unstructured":"Yuan, W., Peng, Y., Guo, Y., Ren, Y., Xue, Q.: DCAU-net: dense convolutional attention U-net for segmentation of intracranial aneurysm images. Vis. Comput. Ind. Biomed. Art 5(1), 9 (2022)","journal-title":"Vis. Comput. Ind. Biomed. Art"},{"issue":"6","key":"18_CR19","doi-asserted-by":"publisher","first-page":"1856","DOI":"10.1109\/TMI.2019.2959609","volume":"39","author":"Z Zhou","year":"2019","unstructured":"Zhou, Z., Siddiquee, M.M.R., Tajbakhsh, N., Liang, J.: Unet++: redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 39(6), 1856\u20131867 (2019)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"2","key":"18_CR20","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203\u2013211 (2021)","journal-title":"Nat. Methods"},{"key":"18_CR21","unstructured":"Oktay, O., et al.: Attention U-net: learning where to look for the pancreas. arXiv preprint arXiv:1804.03999 (2018)"},{"key":"18_CR22","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"18_CR23","doi-asserted-by":"crossref","unstructured":"Cao, H., et al.: Swin-Unet: Unet-like pure transformer for medical image segmentation. In: European Conference on Computer Vision, pp. 205\u2013218 (2022)","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"18_CR24","unstructured":"Zhou, H.Y., Guo, J., Zhang, Y., Yu, L., Wang, L., Yu, Y.: nnformer: Interleaved transformer for volumetric segmentation. arXiv preprint arXiv:2109.03201 (2021)"},{"issue":"1","key":"18_CR25","doi-asserted-by":"publisher","first-page":"96","DOI":"10.1109\/TMI.2023.3291719","volume":"43","author":"Z Li","year":"2023","unstructured":"Li, Z., et al.: LViT: language meets vision transformer in medical image segmentation. IEEE Trans. Med. Imaging 43(1), 96\u2013107 (2023)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"18_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2023.102939","volume":"90","author":"X Yu","year":"2023","unstructured":"Yu, X., et al.: Unest: local spatial representation learning with hierarchical transformer for efficient medical segmentation. Med. Image Anal. 90, 102939 (2023)","journal-title":"Med. Image Anal."},{"key":"18_CR27","doi-asserted-by":"crossref","unstructured":"Li, P., Gao, C., Lian, C., Meng, D.: Spatial prior-guided Bi-directional cross-attention transformers for tooth instance segmentation. IEEE Trans. Med. Imaging (2024)","DOI":"10.1109\/TMI.2024.3406015"},{"key":"18_CR28","doi-asserted-by":"crossref","unstructured":"Kong, Y., Zhang, X., Wang, W., Zhou, Y., Li, Y., Yuan, Y.: Multi-scale spatial-temporal attention networks for functional connectome classification. IEEE Trans. Med. Imaging (2024)","DOI":"10.1109\/TMI.2024.3448214"},{"key":"18_CR29","doi-asserted-by":"crossref","unstructured":"Liu, M., Wu, S., Chen, R., Lin, Z., Wang, Y., Meijering, E.: Brain image segmentation for ultrascale neuron reconstruction via an adaptive dual-task learning network. IEEE Trans. Med. Imaging (2024)","DOI":"10.1109\/TMI.2024.3367384"},{"key":"18_CR30","doi-asserted-by":"crossref","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics, pp. 4171\u20134186 (2019)","DOI":"10.18653\/v1\/N19-1423"},{"key":"18_CR31","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980\u20132988 (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"18_CR32","doi-asserted-by":"crossref","unstructured":"Degerli, A., Kiranyaz, S., Chowdhury, M.E., Gabbouj, M.: OSegnet: operational segmentation network for COVID-19 detection using chest x-ray images. In: 2022 IEEE International Conference on Image Processing, pp. 2306\u20132310 (2022)","DOI":"10.1109\/ICIP46576.2022.9897412"},{"key":"18_CR33","doi-asserted-by":"crossref","unstructured":"Morozov, S.P., et al.: MosMedData: chest CT scans with COVID-19 related findings dataset. arXiv preprint arXiv:2005.06465 (2020)","DOI":"10.1101\/2020.05.20.20100362"},{"key":"18_CR34","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s41747-020-00173-2","volume":"4","author":"J Hofmanninger","year":"2020","unstructured":"Hofmanninger, J., Prayer, F., Pan, J., R\u00f6hrich, S., Prosch, H., Langs, G.: Automatic lung segmentation in routine imaging is primarily a data diversity problem, not a methodology problem. Eur. Radiol. Exp. 4, 1\u201313 (2020)","journal-title":"Eur. Radiol. Exp."},{"key":"18_CR35","doi-asserted-by":"crossref","unstructured":"Wang, H., Cao, P., Wang, J., Zaiane, O.R. : UCTransnet: rethinking the skip connections in U-net from a channel-wise perspective with transformer. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 2441\u20132449 (2022)","DOI":"10.1609\/aaai.v36i3.20144"},{"key":"18_CR36","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763 (2021)"},{"key":"18_CR37","doi-asserted-by":"crossref","unstructured":"Yang, Z., Wang, J., Tang, Y., Chen, K., Zhao, H., Torr, P.H.: LAVT: language-aware vision transformer for referring image segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18155\u201318165 (2022)","DOI":"10.1109\/CVPR52688.2022.01762"}],"updated-by":[{"DOI":"10.1007\/978-981-95-5631-1_37","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2026,4,5]],"date-time":"2026-04-05T00:00:00Z","timestamp":1775347200000}}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5631-1_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T14:44:01Z","timestamp":1775313841000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5631-1_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556304","9789819556311"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5631-1_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"28 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"5 April 2026","order":2,"name":"change_date","label":"Change Date","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"Correction","order":3,"name":"change_type","label":"Change Type","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"A correction has been published.","order":4,"name":"change_details","label":"Change Details","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}