{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T03:35:16Z","timestamp":1743132916407,"version":"3.40.3"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031450860"},{"type":"electronic","value":"9783031450877"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-45087-7_16","type":"book-chapter","created":{"date-parts":[[2023,10,7]],"date-time":"2023-10-07T19:01:37Z","timestamp":1696705297000},"page":"150-159","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Automated Segmentation of\u00a0Nasopharyngeal Carcinoma Based on\u00a0Dual-Sequence Magnetic Resonance Imaging Using Self-supervised Learning"],"prefix":"10.1007","author":[{"given":"Zongyou","family":"Cai","sequence":"first","affiliation":[]},{"given":"Yufeng","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Zhangnan","family":"Zhong","sequence":"additional","affiliation":[]},{"given":"Haiwei","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Ziyue","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Bin","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Deng","sequence":"additional","affiliation":[]},{"given":"Qiting","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Kaixin","family":"Lei","sequence":"additional","affiliation":[]},{"given":"Jiegeng","family":"Lyu","sequence":"additional","affiliation":[]},{"given":"Hanwei","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Bingsheng","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,8]]},"reference":[{"key":"16_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2019.101539","volume":"58","author":"L Chen","year":"2019","unstructured":"Chen, L., Bentley, P., Mori, K., Misawa, K., Fujiwara, M., Rueckert, D.: Self-supervised learning for medical image analysis using image context restoration. Med. Image Anal. 58, 101539 (2019)","journal-title":"Med. Image Anal."},{"key":"16_CR2","unstructured":"Feyjie, A.R., Azad, R., Pedersoli, M., Kauffman, C., Ayed, I.B., Dolz, J.: Semi-supervised few-shot learning for medical image segmentation. arXiv preprint arXiv:2003.08462 (2020)"},{"key":"16_CR3","doi-asserted-by":"crossref","unstructured":"Huang, B., et al.: Fully automated delineation of gross tumor volume for head and neck cancer on PET-CT using deep learning: a dual-center study. Contrast Media Molecular Imaging 2018 (2018)","DOI":"10.1155\/2018\/8923028"},{"key":"16_CR4","doi-asserted-by":"crossref","unstructured":"Huo, Y., Xu, Z., Bao, S., Assad, A., Abramson, R.G., Landman, B.A.: Adversarial synthesis learning enables segmentation without target modality ground truth. In: 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018), pp. 1217\u20131220. IEEE (2018)","DOI":"10.1109\/ISBI.2018.8363790"},{"issue":"2","key":"16_CR5","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203\u2013211 (2021)","journal-title":"Nat. Methods"},{"key":"16_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.oraloncology.2020.104862","volume":"110","author":"L Ke","year":"2020","unstructured":"Ke, L., et al.: Development of a self-constrained 3D DenseNet model in automatic detection and segmentation of nasopharyngeal carcinoma using magnetic resonance images. Oral Oncol. 110, 104862 (2020)","journal-title":"Oral Oncol."},{"key":"16_CR7","doi-asserted-by":"crossref","unstructured":"Kim, S., An, S., Chikontwe, P., Park, S.H.: Bidirectional RNN-based few shot learning for 3D medical image segmentation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 1808\u20131816 (2021)","DOI":"10.1609\/aaai.v35i3.16275"},{"key":"16_CR8","doi-asserted-by":"crossref","unstructured":"Lahiani, A., Gildenblat, J., Klaman, I., Navab, N., Klaiman, E.: Generalizing multistain immunohistochemistry tissue segmentation using one-shot color deconvolution deep neural networks. arXiv preprint arXiv:1805.06958 (2018)","DOI":"10.1049\/iet-ipr.2018.6513"},{"key":"16_CR9","doi-asserted-by":"crossref","unstructured":"Li, Q., et al.: Tumor segmentation in contrast-enhanced magnetic resonance imaging for nasopharyngeal carcinoma: deep learning with convolutional neural network. BioMed Res. Int. 2018 (2018)","DOI":"10.1155\/2018\/9128527"},{"issue":"3","key":"16_CR10","doi-asserted-by":"publisher","first-page":"677","DOI":"10.1148\/radiol.2019182012","volume":"291","author":"L Lin","year":"2019","unstructured":"Lin, L., et al.: Deep learning for automated contouring of primary tumor volumes by MRI for nasopharyngeal carcinoma. Radiology 291(3), 677\u2013686 (2019)","journal-title":"Radiology"},{"issue":"3","key":"16_CR11","first-page":"2511","volume":"16","author":"Z Ma","year":"2018","unstructured":"Ma, Z., Wu, X., Song, Q., Luo, Y., Wang, Y., Zhou, J.: Automated nasopharyngeal carcinoma segmentation in magnetic resonance images by combination of convolutional neural networks and graph cut. Exp. Ther. Med. 16(3), 2511\u20132521 (2018)","journal-title":"Exp. Ther. Med."},{"issue":"2","key":"16_CR12","doi-asserted-by":"publisher","DOI":"10.1088\/1361-6560\/aaf5da","volume":"64","author":"Z Ma","year":"2019","unstructured":"Ma, Z., et al.: Nasopharyngeal carcinoma segmentation based on enhanced convolutional neural networks using multi-modal metric learning. Phys. Med. Biol. 64(2), 025005 (2019)","journal-title":"Phys. Med. Biol."},{"key":"16_CR13","unstructured":"Mattes, D., Haynor, D.R., Vesselle, H., Lewellyn, T.K., Eubank, W.: Nonrigid multimodality image registration. In: Medical Imaging 2001: Image Processing, vol. 4322, pp. 1609\u20131620. SPIE (2001)"},{"key":"16_CR14","doi-asserted-by":"publisher","first-page":"315","DOI":"10.3389\/fonc.2017.00315","volume":"7","author":"K Men","year":"2017","unstructured":"Men, K., et al.: Deep deconvolutional neural network for target segmentation of nasopharyngeal cancer in planning computed tomography images. Front. Oncol. 7, 315 (2017)","journal-title":"Front. Oncol."},{"key":"16_CR15","unstructured":"Mondal, A.K., Dolz, J., Desrosiers, C.: Few-shot 3D multi-modal medical image segmentation using generative adversarial learning. arXiv preprint arXiv:1810.12241 (2018)"},{"key":"16_CR16","doi-asserted-by":"crossref","unstructured":"Peng, H., et al.: The current status of clinical trials focusing on nasopharyngeal carcinoma: a comprehensive analysis of ClinicalTrials. gov database. PLoS One 13(5), e0196730 (2018)","DOI":"10.1371\/journal.pone.0196730"},{"key":"16_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"16_CR18","doi-asserted-by":"publisher","first-page":"925","DOI":"10.1007\/s11548-018-1772-0","volume":"13","author":"T Ross","year":"2018","unstructured":"Ross, T., et al.: Exploiting the potential of unlabeled endoscopic video data with self-supervised learning. Int. J. Comput. Assist. Radiol. Surg. 13, 925\u2013933 (2018)","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"16_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2019.101587","volume":"59","author":"AG Roy","year":"2020","unstructured":"Roy, A.G., Siddiqui, S., P\u00f6lsterl, S., Navab, N., Wachinger, C.: Squeeze and excite guided few-shot segmentation of volumetric images. Med. Image Anal. 59, 101587 (2020)","journal-title":"Med. Image Anal."},{"key":"16_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"661","DOI":"10.1007\/978-3-030-78191-0_51","volume-title":"Information Processing in Medical Imaging","author":"A Taleb","year":"2021","unstructured":"Taleb, A., Lippert, C., Klein, T., Nabi, M.: Multimodal self-supervised learning for medical image analysis. In: Feragen, A., Sommer, S., Schnabel, J., Nielsen, M. (eds.) IPMI 2021. LNCS, vol. 12729, pp. 661\u2013673. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-78191-0_51"},{"key":"16_CR21","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1016\/j.neucom.2020.12.085","volume":"435","author":"P Tang","year":"2021","unstructured":"Tang, P., et al.: DA-DSUnet: dual attention-based dense SU-et for automatic head-and-neck tumor segmentation in MRI images. Neurocomputing 435, 103\u2013113 (2021)","journal-title":"Neurocomputing"},{"key":"16_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102381","volume":"78","author":"G Tao","year":"2022","unstructured":"Tao, G., et al.: SeqSeg: a sequential method to achieve nasopharyngeal carcinoma segmentation free from background dominance. Med. Image Anal. 78, 102381 (2022)","journal-title":"Med. Image Anal."},{"key":"16_CR23","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"238","DOI":"10.1007\/978-3-030-59719-1_24","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2020","author":"X Tao","year":"2020","unstructured":"Tao, X., Li, Y., Zhou, W., Ma, K., Zheng, Y.: Revisiting Rubik\u2019s Cube: self-supervised learning with volume-wise transformation for 3D medical image segmentation. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12264, pp. 238\u2013248. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59719-1_24"},{"key":"16_CR24","doi-asserted-by":"crossref","unstructured":"Wang, H., et al.: A collaborative dictionary learning model for nasopharyngeal carcinoma segmentation on multimodalities MR sequences. Comput. Math. Methods Med. 2020 (2020)","DOI":"10.1155\/2020\/7562140"},{"key":"16_CR25","doi-asserted-by":"publisher","first-page":"1323","DOI":"10.1007\/s11063-017-9759-3","volume":"48","author":"Y Wang","year":"2018","unstructured":"Wang, Y., et al.: Automatic tumor segmentation with deep convolutional neural networks for radiotherapy applications. Neural Process. Lett. 48, 1323\u20131334 (2018)","journal-title":"Neural Process. Lett."},{"key":"16_CR26","doi-asserted-by":"publisher","first-page":"166","DOI":"10.3389\/fonc.2020.00166","volume":"10","author":"Y Ye","year":"2020","unstructured":"Ye, Y., et al.: Fully-automated segmentation of nasopharyngeal carcinoma on dual-sequence MRI using convolutional neural networks. Front. Oncol. 10, 166 (2020)","journal-title":"Front. Oncol."},{"key":"16_CR27","doi-asserted-by":"crossref","unstructured":"Zhao, A., Balakrishnan, G., Durand, F., Guttag, J.V., Dalca, A.V.: Data augmentation using learned transformations for one-shot medical image segmentation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 8543\u20138553 (2019)","DOI":"10.1109\/CVPR.2019.00874"},{"key":"16_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2020.101840","volume":"67","author":"Z Zhou","year":"2021","unstructured":"Zhou, Z., Sodha, V., Pang, J., Gotway, M.B., Liang, J.: Models genesis. Med. Image Anal. 67, 101840 (2021)","journal-title":"Med. Image Anal."}],"container-title":["Lecture Notes in Computer Science","Computational Mathematics Modeling in Cancer Analysis"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-45087-7_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,10]],"date-time":"2023-12-10T14:03:07Z","timestamp":1702216987000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-45087-7_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031450860","9783031450877"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-45087-7_16","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"8 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"No relevant conflicts to disclose.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}},{"value":"CMMCA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Computational Mathematics Modeling in Cancer Analysis","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vancouver, BC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Canada","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cmmca2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/cmmca.github.io\/cmmca2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"25","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"17","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"68% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}