{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T06:21:39Z","timestamp":1743056499089,"version":"3.40.3"},"publisher-location":"Cham","reference-count":33,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031500770"},{"type":"electronic","value":"9783031500787"}],"license":[{"start":{"date-parts":[[2023,12,24]],"date-time":"2023-12-24T00:00:00Z","timestamp":1703376000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,24]],"date-time":"2023-12-24T00:00:00Z","timestamp":1703376000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-50078-7_18","type":"book-chapter","created":{"date-parts":[[2023,12,23]],"date-time":"2023-12-23T11:01:59Z","timestamp":1703329319000},"page":"222-236","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["ScaleNet: Rethinking Feature Interaction from\u00a0a\u00a0Scale-Wise Perspective for\u00a0Medical Image Segmentation"],"prefix":"10.1007","author":[{"given":"Yu","family":"Feng","sequence":"first","affiliation":[]},{"given":"Tai","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Zhengke","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Suwei","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Wen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,24]]},"reference":[{"key":"18_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"18_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-00889-5_1","volume-title":"Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support","author":"Z Zhou","year":"2018","unstructured":"Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: UNet++: a nested U-Net architecture for medical image segmentation. In: Stoyanov, D., et al. (eds.) DLMIA\/ML-CDS -2018. LNCS, vol. 11045, pp. 3\u201311. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00889-5_1"},{"issue":"5","key":"18_CR3","first-page":"1316","volume":"39","author":"S Hyunseok","year":"2019","unstructured":"Hyunseok, S., et al.: Modified U-Net (mU-Net) with incorporation of object-dependent high level features for improved liver and liver-tumor segmentation in CT images. IEEE Trans. Med. Imaging 39(5), 1316\u20131325 (2019)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"18_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"311","DOI":"10.1007\/978-3-030-11726-9_28","volume-title":"Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries","author":"A Myronenko","year":"2019","unstructured":"Myronenko, A.: 3D MRI brain tumor segmentation using autoencoder regularization. In: Crimi, A., Bakas, S., Kuijf, H., Keyvan, F., Reyes, M., van Walsum, T. (eds.) BrainLes 2018. LNCS, vol. 11384, pp. 311\u2013320. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11726-9_28"},{"key":"18_CR5","unstructured":"Oktay, O., Schlemper, J., Folgoc, L.L.: Attention U-Net: learning where to look for the pancreas. ArXiv preprint arXiv:1804.03999 (2018)"},{"key":"18_CR6","unstructured":"Luo, W., Li, Y., Urtasun, R., Zemel, R.: Understanding the effective receptive field in deep convolutional neural networks. In: Advances in Neural Information Processing Systems, vol. 29 (2016)"},{"key":"18_CR7","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. ArXiv preprint arXiv:2010.11929 (2020)"},{"key":"18_CR8","unstructured":"Touvron, H., Cord, M., Douze, M., et al.: Training data-efficient image transformers distillation through attention. In: International Conference on Machine Learning, PMLR, pp. 10347\u201310357 (2021)"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Huang, H., et al.: ScaleFormer: revisiting the transformer-based backbones from a scale-wise perspective for medical image segmentation. ArXiv preprint arXiv:2207.14552 (2022)","DOI":"10.24963\/ijcai.2022\/135"},{"key":"18_CR10","unstructured":"Chen, J., et al.: TransUNet: transformers make strong encoders for medical image segmentation. ArXiv preprint arXiv:2102.04306 (2021)"},{"key":"18_CR11","doi-asserted-by":"crossref","unstructured":"Xu, G., Wu, X., Zhang, X., He, X.: LeViT-UNet: make faster encoders with transformer for medical image segmentation. ArXiv preprint arXiv:2107.08623 (2021)","DOI":"10.2139\/ssrn.4116174"},{"key":"18_CR12","doi-asserted-by":"crossref","unstructured":"Shi, L., et al.: STM-UNet: an efficient U-shaped architecture based on Swin transformer and multi-scale MLP for medical image segmentation. ArXiv preprint arXiv:2304.12615 (2023)","DOI":"10.1109\/GLOBECOM54140.2023.10437634"},{"key":"18_CR13","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"18_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"205","DOI":"10.1007\/978-3-031-25066-8_9","volume-title":"Computer Vision-ECCV 2022 Workshops","author":"H Cao","year":"2022","unstructured":"Cao, H., et al.: Swin-UNet: UNet-like pure transformer for medical image segmentation. In: Karlinsky, L., Michaeli, T., Nishino, K. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 205\u2013218. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-25066-8_9"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"Rahman, M.M., Marculescu, R.: Medical image segmentation via cascaded attention decoding. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 6222\u20136231 (2023)","DOI":"10.1109\/WACV56688.2023.00616"},{"key":"18_CR16","doi-asserted-by":"crossref","unstructured":"Huang, X., Gong, H., Zhang, J.: HST-MRF: heterogeneous Swin transformer with multi-receptive field for medical image segmentation. ArXiv preprint arXiv:2304.04614 (2023)","DOI":"10.1109\/JBHI.2024.3397047"},{"key":"18_CR17","doi-asserted-by":"crossref","unstructured":"Huang, H., et al.: UNet3+: a full-scale connected UNet for medical image segmentation. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1055\u20131059 (2020)","DOI":"10.1109\/ICASSP40776.2020.9053405"},{"key":"18_CR18","unstructured":"Huang, X., Deng, Z., Li, D., Yuan, X: MISSFormer: an effective medical image segmentation transformer. ArXiv preprint arXiv:2109.07162 (2021)"},{"key":"18_CR19","unstructured":"You, C., Zhao, R., Liu, F.: Class-aware adversarial transformers for medical image segmentation. In: Advances in Neural Information Processing Systems, vol. 35, pp. 29582\u201329596 (2022)"},{"key":"18_CR20","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"18_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"656","DOI":"10.1007\/978-3-030-59710-8_64","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2020","author":"S Fu","year":"2020","unstructured":"Fu, S., et al.: Domain adaptive relational reasoning for 3D multi-organ segmentation. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12261, pp. 656\u2013666. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59710-8_64"},{"key":"18_CR22","doi-asserted-by":"publisher","first-page":"197","DOI":"10.1016\/j.media.2019.01.012","volume":"53","author":"J Schlemper","year":"2019","unstructured":"Schlemper, J., Oktay, O., Schaap, M.: Attention gated networks: learning to leverage salient regions in medical images. Med. Image Anal. 53, 197\u2013207 (2019)","journal-title":"Med. Image Anal."},{"key":"18_CR23","doi-asserted-by":"crossref","unstructured":"Wang, H., Xie, S., Lin, L.: Mixed transformer U-Net for medical image segmentation. In: Proceedings of the ICASSP, pp. 2390\u20132394 (2022)","DOI":"10.1109\/ICASSP43922.2022.9746172"},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"Heidari, M., et al.: HiFormer: hierarchical multi-scale representations using transformers for medical image segmentation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 6202\u20136212 (2023)","DOI":"10.1109\/WACV56688.2023.00614"},{"key":"18_CR25","unstructured":"Ben, L., et al.: Segmentation outside the cranial vault challenge. In: MICCAI: Multi Atlas Labeling Beyond Cranial Vault-Workshop Challenge (2015)"},{"issue":"11","key":"18_CR26","doi-asserted-by":"publisher","first-page":"2514","DOI":"10.1109\/TMI.2018.2837502","volume":"37","author":"O Bernard","year":"2018","unstructured":"Bernard, O., Lalande, A., Zotti, C.: Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: is the problem solved. IEEE Trans. Med. Imaging 37(11), 2514\u20132525 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"18_CR27","unstructured":"Codella, N., et al.: Skin lesion analysis toward melanoma detection, a challenge hosted by the international skin imaging collaboration (ISIC). ArXiv preprint arXiv:1902.03368 (2019)"},{"key":"18_CR28","doi-asserted-by":"publisher","first-page":"108205","DOI":"10.1109\/ACCESS.2022.3211501","volume":"10","author":"R Azad","year":"2022","unstructured":"Azad, R., Heidari, M., Merhof, D.: TransNorm: transformer provides a strong spatial normalization mechanism for a deep segmentation model. IEEE Access 10, 108205\u2013108215 (2022)","journal-title":"IEEE Access"},{"key":"18_CR29","doi-asserted-by":"crossref","unstructured":"Azad, R., Bozorgpour, A., Asadi-Aghbolaghi, M., Merhof, D., Escalera, S.: Deep frequency re-calibration U-Net for medical image segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3274\u20133283 (2021)","DOI":"10.1109\/ICCVW54120.2021.00366"},{"key":"18_CR30","doi-asserted-by":"crossref","unstructured":"Azad, R., Heidari, M., Wu, Y.: Contextual attention network: transformer meets U-Net. ArXiv preprint arXiv:2203.01932 (2022)","DOI":"10.1007\/978-3-031-21014-3_39"},{"key":"18_CR31","unstructured":"Azad, R., Jia, Y., Aghdam, E.K., Cohen-Adad, J., Merhof, D.: Enhancing medical image segmentation with TransCeption: a multi-scale feature fusion approach. ArXiv preprint arXiv:2301.10847 (2023)"},{"key":"18_CR32","doi-asserted-by":"publisher","first-page":"2160","DOI":"10.1109\/TIP.2023.3263111","volume":"32","author":"W Zongwei","year":"2023","unstructured":"Zongwei, W., Guillaume, A., Fabrice, M.: HiDAnet: RGB-D salient object detection via hierarchical depth awareness. IEEE Trans. Image Process. 32, 2160\u20132173 (2023)","journal-title":"IEEE Trans. Image Process."},{"key":"18_CR33","doi-asserted-by":"crossref","unstructured":"Zhou, T., Fu, H., Chen, G.: Specificity-preserving RGB-D saliency detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4681\u20134691 (2021)","DOI":"10.1109\/ICCV48922.2021.00464"}],"container-title":["Lecture Notes in Computer Science","Advances in Computer Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-50078-7_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,6]],"date-time":"2024-11-06T19:39:47Z","timestamp":1730921987000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-50078-7_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,24]]},"ISBN":["9783031500770","9783031500787"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-50078-7_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,12,24]]},"assertion":[{"value":"24 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CGI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Computer Graphics International Conference","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cgi2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"385","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"149","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"39% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}