{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T03:25:04Z","timestamp":1769829904591,"version":"3.49.0"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031439865","type":"print"},{"value":"9783031439872","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-43987-2_7","type":"book-chapter","created":{"date-parts":[[2023,9,30]],"date-time":"2023-09-30T23:07:48Z","timestamp":1696115268000},"page":"64-74","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":18,"title":["SegNetr: Rethinking the Local-Global Interactions and Skip Connections in U-Shaped Networks"],"prefix":"10.1007","author":[{"given":"Junlong","family":"Cheng","sequence":"first","affiliation":[]},{"given":"Chengrui","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Fengjie","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Min","family":"Zhu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,1]]},"reference":[{"key":"7_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"7_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"358","DOI":"10.1007\/978-3-030-87193-2_34","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"Q Ma","year":"2021","unstructured":"Ma, Q., Zu, C., Wu, X., Zhou, J., Wang, Y.: Coarse-to-fine segmentation of organs at risk in nasopharyngeal carcinoma radiotherapy. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12901, pp. 358\u2013368. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_34"},{"key":"7_CR3","first-page":"109512","volume":"253","author":"Z Han","year":"2022","unstructured":"Han, Z., Jian, M., Wang, G.G.: ConvUNeXt: an efficient convolution neural network for medical image segmentation. KBS 253, 109512 (2022)","journal-title":"KBS"},{"key":"7_CR4","unstructured":"Oktay, O., Schlemper, J., Folgoc, L.L., et al.: Attention U-Net: learning where to look for the pancreas. arXiv preprint arXiv:1804.03999 (2018)"},{"key":"7_CR5","doi-asserted-by":"publisher","first-page":"102313","DOI":"10.1016\/j.media.2021.102313","volume":"76","author":"J Cheng","year":"2022","unstructured":"Cheng, J., Tian, S., Yu, L., et al.: ResGANet: residual group attention network for medical image classification and segmentation. Med. Image Anal. 76, 102313 (2022)","journal-title":"Med. Image Anal."},{"key":"7_CR6","doi-asserted-by":"publisher","first-page":"102447","DOI":"10.1016\/j.media.2022.102447","volume":"79","author":"K Wang","year":"2022","unstructured":"Wang, K., Zhan, B., Zu, C., et al.: Semi-supervised medical image segmentation via a tripled-uncertainty guided mean teacher model with contrastive learning. Med. Image Anal. 79, 102447 (2022)","journal-title":"Med. Image Anal."},{"issue":"10","key":"7_CR7","first-page":"2281","volume":"38","author":"Z Gu","year":"2019","unstructured":"Gu, Z., Cheng, J., Fu, H., et al.: Ce-net: context encoder network for 2D medical image segmentation. IEEE TMI 38(10), 2281\u20132292 (2019)","journal-title":"IEEE TMI"},{"key":"7_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00521-022-07859-1","volume":"35","author":"Y Wu","year":"2022","unstructured":"Wu, Y., et al.: D-former: a U-shaped dilated transformer for 3D medical image segmentation. Neural Comput. Appl. 35, 1\u201314 (2022). https:\/\/doi.org\/10.1007\/s00521-022-07859-1","journal-title":"Neural Comput. Appl."},{"key":"7_CR9","first-page":"102145","volume":"62","author":"J Cheng","year":"2020","unstructured":"Cheng, J., Tian, S., Yu, L., et al.: A deep learning algorithm using contrast-enhanced computed tomography (CT) images for segmentation and rapid automatic detection of aortic dissection. BSPC 62, 102145 (2020)","journal-title":"BSPC"},{"key":"7_CR10","unstructured":"Dosovitskiy, A., et al.: An image is worth 16$$\\times $$16 words: transformers for image recognition at scale. In: ICLR, pp. 3\u20137 (2021)"},{"key":"7_CR11","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NIPS, vol. 30 (2017)"},{"key":"7_CR12","unstructured":"Chen, J., et al.: TransUNet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:2102.04306 (2021)"},{"key":"7_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"205","DOI":"10.1007\/978-3-031-25066-8_9","volume-title":"Computer Vision - ECCV 2022 Workshops","author":"H Cao","year":"2023","unstructured":"Cao, H., et al.: Swin-Unet: Unet-like pure transformer for medical image segmentation. In: Karlinsky, L., Michaeli, T., Nishino, K. (eds.) ECCV 2022. Lecture Notes in Computer Science, vol. 13803, pp. 205\u2013218. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-25066-8_9"},{"key":"7_CR14","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: IEEE ICCV, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"7_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1007\/978-3-031-16443-9_3","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022","author":"JMJ Valanarasu","year":"2022","unstructured":"Valanarasu, J.M.J., Patel, V.M.: UNeXt: MLP-based rapid medical image segmentation network. In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) MICCAI 2022. Lecture Notes in Computer Science, vol. 13435, pp. 23\u201333. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-16443-9_3"},{"key":"7_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1007\/978-3-030-87193-2_2","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Liu, H., Hu, Q.: TransFuse: fusing transformers and CNNs for medical image segmentation. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12901, pp. 14\u201324. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_2"},{"key":"7_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1007\/978-3-030-87193-2_4","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"JMJ Valanarasu","year":"2021","unstructured":"Valanarasu, J.M.J., Oza, P., Hacihaliloglu, I., Patel, V.M.: Medical transformer: gated axial-attention for medical image segmentation. In: de Bruijne, M., Zheng, Y., Essert, C. (eds.) MICCAI 2021. LNCS, vol. 12901, pp. 36\u201346. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_4"},{"key":"7_CR18","doi-asserted-by":"publisher","first-page":"102327","DOI":"10.1016\/j.media.2021.102327","volume":"76","author":"H Wu","year":"2022","unstructured":"Wu, H., Chen, S., Chen, G., et al.: FAT-Net: feature adaptive transformers for automated skin lesion segmentation. Med. Image Anal. 76, 102327 (2022)","journal-title":"Med. Image Anal."},{"key":"7_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-00889-5_1","volume-title":"Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support","author":"Z Zhou","year":"2018","unstructured":"Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: UNet++: a nested U-Net architecture for medical image segmentation. In: Stoyanov, D., et al. (eds.) DLMIA\/ML-CDS -2018. LNCS, vol. 11045, pp. 3\u201311. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00889-5_1"},{"issue":"12","key":"7_CR20","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: SegNet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE TPAMI 39(12), 2481\u20132495 (2017)","journal-title":"IEEE TPAMI"},{"key":"7_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1007\/978-3-030-59710-8_8","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2020","author":"T Xiang","year":"2020","unstructured":"Xiang, T., Zhang, C., Liu, D., Song, Y., Huang, H., Cai, W.: BiO-Net: learning recurrent bi-directional connections for encoder-decoder architecture. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12261, pp. 74\u201384. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59710-8_8"},{"key":"7_CR22","doi-asserted-by":"crossref","unstructured":"Wang, H., et, al.: UCTransNet: rethinking the skip connections in U-Net from a channel-wise perspective with transformer. In: AAAI, vol. 36(3), pp. 2441\u20132449 (2022)","DOI":"10.1609\/aaai.v36i3.20144"},{"key":"7_CR23","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"459","DOI":"10.1007\/978-3-031-20053-3_27","volume-title":"Computer Vision - ECCV 2022","author":"Z Tu","year":"2022","unstructured":"Tu, Z., et al.: MaxViT: multi-axis vision transformer. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. Lecture Notes in Computer Science, vol. 13684, pp. 459\u2013479. Springer, Cham (2022)"},{"key":"7_CR24","unstructured":"Tan, M., Le, Q.: EfficientNet: rethinking model scaling for convolutional neural networks. In: ICML, PP. 6105\u20136114 (2019)"},{"key":"7_CR25","doi-asserted-by":"crossref","unstructured":"Quang, N.H.: Automatic skin lesion analysis towards melanoma detection. In: IES, pp. 106\u2013111. IEEE (2017)","DOI":"10.1109\/IESYS.2017.8233570"},{"key":"7_CR26","doi-asserted-by":"crossref","unstructured":"Mendon\u00e7a, T., et al.: PH 2-A dermoscopic image database for research and benchmarking. In: EMBC, pp. 5437\u20135440. IEEE (2013)","DOI":"10.1109\/EMBC.2013.6610779"},{"key":"7_CR27","doi-asserted-by":"crossref","unstructured":"Pedraza, L., et al.: An open access thyroid ultrasound image database. In: SPIE, vol. 9287, pp. 188\u2013193 (2015)","DOI":"10.1117\/12.2073532"},{"issue":"11","key":"7_CR28","first-page":"2514","volume":"37","author":"O Bernard","year":"2018","unstructured":"Bernard, O., Lalande, A., Zotti, C., et al.: Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: is the problem solved? IEEE TMI 37(11), 2514\u20132525 (2018)","journal-title":"IEEE TMI"},{"issue":"2","key":"7_CR29","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee, F., Jaeger, P.F., Kohl, S.A.A., et al.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203\u2013211 (2021)","journal-title":"Nat. Methods"},{"key":"7_CR30","doi-asserted-by":"publisher","first-page":"108636","DOI":"10.1016\/j.patcog.2022.108636","volume":"127","author":"K Wang","year":"2022","unstructured":"Wang, K., Zhang, X., Zhang, X., et al.: EANet: iterative edge attention network for medical image segmentation. Pattern Recogn. 127, 108636 (2022)","journal-title":"Pattern Recogn."}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-43987-2_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T19:18:03Z","timestamp":1730229483000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-43987-2_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031439865","9783031439872"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-43987-2_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"1 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vancouver, BC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Canada","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conferences.miccai.org\/2023\/en\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2250","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"730","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"32% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}