{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T06:46:48Z","timestamp":1774594008030,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":26,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819620630","type":"print"},{"value":"9789819620647","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,12,28]],"date-time":"2024-12-28T00:00:00Z","timestamp":1735344000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,28]],"date-time":"2024-12-28T00:00:00Z","timestamp":1735344000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-2064-7_18","type":"book-chapter","created":{"date-parts":[[2024,12,27]],"date-time":"2024-12-27T19:25:03Z","timestamp":1735327503000},"page":"243-256","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Pubic Symphysis-Fetal Head Segmentation Network Using BiFormer Attention Mechanism and\u00a0Multipath Dilated Convolution"],"prefix":"10.1007","author":[{"given":"Pengzhou","family":"Cai","sequence":"first","affiliation":[]},{"given":"Lu","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Yanxin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Xiaojuan","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Libin","family":"Lan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,28]]},"reference":[{"issue":"6","key":"18_CR1","first-page":"1027","volume":"98","author":"M Fitzpatrick","year":"2001","unstructured":"Fitzpatrick, M., McQuillan, K., O\u2019Herlihy, C.: Influence of persistent occiput posterior position on delivery outcome. Obstetr. Gynecol. 98(6), 1027\u20131031 (2001)","journal-title":"Obstetr. Gynecol."},{"issue":"1","key":"18_CR2","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1111\/j.1523-536X.2009.00380.x","volume":"37","author":"P Simkin","year":"2010","unstructured":"Simkin, P.: The fetal occiput posterior position: state of the science and a new perspective. Birth 37(1), 61\u201371 (2010)","journal-title":"Birth"},{"issue":"3","key":"18_CR3","doi-asserted-by":"publisher","first-page":"258","DOI":"10.1046\/j.1469-0705.2002.00656.x","volume":"19","author":"DM Sherer","year":"2002","unstructured":"Sherer, D.M., Bradley, K.S., Langer, O.: Intrapartum fetal head position i: comparison between transvaginal digital examination and transabdominal ultrasound assessment during the active stage of labor. Ultrasound Obstet. Gynecol. 19(3), 258\u2013263 (2002)","journal-title":"Ultrasound Obstet. Gynecol."},{"key":"18_CR4","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"18_CR5","doi-asserted-by":"publisher","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"18_CR6","unstructured":"Oktay, O., et\u00a0al.: Attention u-net: learning where to look for the pancreas. arXiv preprint arXiv:1804.03999 (2018)"},{"issue":"1","key":"18_CR7","first-page":"5192338","volume":"2022","author":"L Yaosheng","year":"2022","unstructured":"Yaosheng, L., et al.: Multitask deep neural network for the fully automatic measurement of the angle of progression. Comput. Math. Methods Med. 2022(1), 5192338 (2022)","journal-title":"Comput. Math. Methods Med."},{"key":"18_CR8","doi-asserted-by":"publisher","DOI":"10.3389\/fphys.2022.940150","volume":"13","author":"J Bai","year":"2022","unstructured":"Bai, J., et al.: A framework for computing angle of progression from transperineal ultrasound images for evaluating fetal head descent using a novel double branch network. Front. Physiol. 13, 940150 (2022)","journal-title":"Front. Physiol."},{"key":"18_CR9","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems. vol. 30 (2017)"},{"key":"18_CR10","unstructured":"Chen, J., et al.: Transunet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:2102.04306 (2021)"},{"key":"18_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1007\/978-3-030-87193-2_2","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Liu, H., Hu, Q.: TransFuse: fusing transformers and CNNs for medical image segmentation. In: de Bruijne, M., et al. (eds.) MICCAI 2021, Part I. LNCS, vol. 12901, pp. 14\u201324. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_2"},{"key":"18_CR12","doi-asserted-by":"crossref","unstructured":"Heidari, M., et al.: Hiformer: hierarchical multi-scale representations using transformers for medical image segmentation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 6202\u20136212 (2023)","DOI":"10.1109\/WACV56688.2023.00614"},{"key":"18_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"377","DOI":"10.1007\/978-3-031-21014-3_39","volume-title":"Machine Learning in Medical Imaging - MLMI 2022","author":"R Azad","year":"2022","unstructured":"Azad, R., Heidari, M., Yuli, W., Merhof, D.: Contextual attention network: transformer meets U-net. In: Lian, C., Cao, X., Rekik, I., Xu, X., Cui, Z. (eds.) MLMI 2022. LNCS, vol. 13583, pp. 377\u2013386. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-21014-3_39"},{"key":"18_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"205","DOI":"10.1007\/978-3-031-25066-8_9","volume-title":"Computer Vision - ECCV 2022 Workshops","author":"H Cao","year":"2022","unstructured":"Cao, H., et al.: Swin-Unet: Unet-like pure transformer for medical image segmentation. In: Karlinsky, L., Michaeli, T., Nishino, K. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 205\u2013218. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-25066-8_9"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"18_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1007\/978-3-030-87193-2_4","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"JMJ Valanarasu","year":"2021","unstructured":"Valanarasu, J.M.J., Oza, P., Hacihaliloglu, I., Patel, V.M.: Medical transformer: gated axial-attention for medical image segmentation. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12901, pp. 36\u201346. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_4"},{"key":"18_CR17","doi-asserted-by":"crossref","unstructured":"Zhu, L., Wang, X., Ke, Z., Zhang, W., Lau, R.W.H.: Biformer: vision transformer with bi-level routing attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10323\u201310333 (2023)","DOI":"10.1109\/CVPR52729.2023.00995"},{"key":"18_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.109728","volume":"142","author":"G Chen","year":"2023","unstructured":"Chen, G., Li, L., Zhang, J., Dai, Yu.: Rethinking the unpretentious u-net for medical ultrasound image segmentation. Pattern Recogn. 142, 109728 (2023)","journal-title":"Pattern Recogn."},{"issue":"2","key":"18_CR19","doi-asserted-by":"publisher","first-page":"476","DOI":"10.1109\/TMI.2021.3116087","volume":"41","author":"Z Ning","year":"2021","unstructured":"Ning, Z., Zhong, S., Feng, Q., Chen, W., Zhang, Yu.: SMU-net: saliency-guided morphology-aware u-net for breast lesion segmentation in ultrasound image. IEEE Trans. Med. Imaging 41(2), 476\u2013490 (2021)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"18_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.123096","volume":"245","author":"Z Chen","year":"2024","unstructured":"Chen, Z., Zhanhong, O., Yaosheng, L., Bai, J.: Direction-guided and multi-scale feature screening for fetal head-pubic symphysis segmentation and angle of progression calculation. Expert Syst. Appl. 245, 123096 (2024)","journal-title":"Expert Syst. Appl."},{"key":"18_CR21","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"18_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.dib.2022.107904","volume":"41","author":"L Yaosheng","year":"2022","unstructured":"Yaosheng, L., et al.: The JNU-IFM dataset for segmenting pubic symphysis-fetal head. Data Brief 41, 107904 (2022)","journal-title":"Data Brief"},{"key":"18_CR23","doi-asserted-by":"crossref","unstructured":"Bai, J., et\u00a0al.: PSFHS challenge report: pubic symphysis and fetal head segmentation from intrapartum ultrasound images. Med. Image Anal. 103353 (2024)","DOI":"10.1016\/j.media.2024.103353"},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"van\u00a0den Heuvel, T.L.A., de\u00a0Bruijn, D., de\u00a0Korte, C.L., van Ginneken, B.: Automated measurement of fetal head circumference using 2d ultrasound images. PLoS One 13(8), e0200412 (2018)","DOI":"10.1371\/journal.pone.0200412"},{"key":"18_CR25","doi-asserted-by":"crossref","unstructured":"Chen, B., Liu, Y., Zhang, Z., Lu, G., Kong, A.W.K.: TransAttUnet: multi-level attention-guided u-net with transformer for medical image segmentation. IEEE Trans. Emerg. Top. Comput. Intell. (2023)","DOI":"10.1109\/TETCI.2023.3309626"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 618\u2013626 (2017)","DOI":"10.1109\/ICCV.2017.74"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-2064-7_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,27]],"date-time":"2024-12-27T20:04:10Z","timestamp":1735329850000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-2064-7_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,28]]},"ISBN":["9789819620630","9789819620647"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-2064-7_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,28]]},"assertion":[{"value":"28 December 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Nara","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 January 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 January 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/mmm2025.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}