{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T12:49:55Z","timestamp":1775911795476,"version":"3.50.1"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783031090011","type":"print"},{"value":"9783031090028","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-09002-8_1","type":"book-chapter","created":{"date-parts":[[2022,7,14]],"date-time":"2022-07-14T12:05:34Z","timestamp":1657800334000},"page":"3-14","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":98,"title":["BiTr-Unet: A CNN-Transformer Combined Network for\u00a0MRI Brain Tumor Segmentation"],"prefix":"10.1007","author":[{"given":"Qiran","family":"Jia","sequence":"first","affiliation":[]},{"given":"Hai","family":"Shu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,7,15]]},"reference":[{"key":"1_CR1","unstructured":"Baid, U., et al.: The RSNA-ASNR-MICCAI BraTs 2021 benchmark on brain tumor segmentation and radiogenomic classification. arXiv preprint arXiv:2107.02314 (2021)"},{"key":"1_CR2","doi-asserted-by":"publisher","unstructured":"Bakas, S., et al.: Segmentation labels and radiomic features for the pre-operative scans of the TCGA-GBM collection, July 2017. https:\/\/doi.org\/10.7937\/K9\/TCIA.2017.KLXWJJ1Q","DOI":"10.7937\/K9\/TCIA.2017.KLXWJJ1Q"},{"key":"1_CR3","unstructured":"Bakas, S., et al.: Segmentation labels and radiomic features for the pre-operative scans of the TCGA-LGG collection. Cancer Imaging Arch. 286 (2017)"},{"issue":"1","key":"1_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/sdata.2017.117","volume":"4","author":"S Bakas","year":"2017","unstructured":"Bakas, S., et al.: Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci. Data 4(1), 1\u201313 (2017)","journal-title":"Sci. Data"},{"key":"1_CR5","doi-asserted-by":"crossref","unstructured":"Chen, C.F., Fan, Q., Panda, R.: Crossvit: cross-attention multi-scale vision transformer for image classification. arXiv preprint arXiv:2103.14899 (2021)","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"1_CR6","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"1_CR7","doi-asserted-by":"crossref","unstructured":"Hatamizadeh, A., Yang, D., Roth, H., Xu, D.: UnetR: transformers for 3D medical image segmentation. arXiv preprint arXiv:2103.10504 (2021)","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"1_CR8","doi-asserted-by":"crossref","unstructured":"Henry, T., et al.: Brain tumor segmentation with self-ensembled, deeply-supervised 3D u-net neural networks: a BraTs 2020 challenge solution. arXiv preprint arXiv:2011.01045 (2020)","DOI":"10.1007\/978-3-030-72084-1_30"},{"issue":"4","key":"1_CR9","doi-asserted-by":"publisher","first-page":"582","DOI":"10.1007\/s10278-019-00227-x","volume":"32","author":"MH Hesamian","year":"2019","unstructured":"Hesamian, M.H., Jia, W., He, X., Kennedy, P.: Deep learning techniques for medical image segmentation: achievements and challenges. J. Digit. Imaging 32(4), 582\u2013596 (2019). https:\/\/doi.org\/10.1007\/s10278-019-00227-x","journal-title":"J. Digit. Imaging"},{"key":"1_CR10","doi-asserted-by":"crossref","unstructured":"Isensee, F., et al.: nnu-Net: self-adapting framework for u-net-based medical image segmentation. arXiv preprint arXiv:1809.10486 (2018)","DOI":"10.1007\/978-3-658-25326-4_7"},{"key":"1_CR11","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q. (eds.) Advances in Neural Information Processing Systems, vol. 25. Curran Associates, Inc. (2012). https:\/\/proceedings.neurips.cc\/paper\/2012\/file\/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf"},{"key":"1_CR12","doi-asserted-by":"publisher","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521, 436\u201344 (2015). https:\/\/doi.org\/10.1038\/nature14539","DOI":"10.1038\/nature14539"},{"key":"1_CR13","doi-asserted-by":"crossref","unstructured":"Lin, H., et al.: Cat: cross attention in vision transformer. arXiv preprint arXiv:2106.05786 (2021)","DOI":"10.1109\/ICME52920.2022.9859720"},{"key":"1_CR14","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. arXiv preprint arXiv:2103.14030 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"1_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"435","DOI":"10.1007\/978-3-030-72084-1_39","volume-title":"Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries","author":"C Lyu","year":"2021","unstructured":"Lyu, C., Shu, H.: A two-stage cascade model with variational autoencoders and attention gates for MRI brain tumor segmentation. In: Crimi, A., Bakas, S. (eds.) BrainLes 2020. LNCS, vol. 12658, pp. 435\u2013447. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-72084-1_39"},{"issue":"10","key":"1_CR16","doi-asserted-by":"publisher","first-page":"1993","DOI":"10.1109\/TMI.2014.2377694","volume":"34","author":"BH Menze","year":"2014","unstructured":"Menze, B.H., et al.: The multimodal brain tumor image segmentation benchmark (brats). IEEE Trans. Med. Imaging 34(10), 1993\u20132024 (2014)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"4","key":"1_CR17","doi-asserted-by":"publisher","first-page":"722","DOI":"10.1007\/s10489-014-0629-7","volume":"42","author":"K Noda","year":"2014","unstructured":"Noda, K., Yamaguchi, Y., Nakadai, K., Okuno, H.G., Ogata, T.: Audio-visual speech recognition using deep learning. Appl. Intell. 42(4), 722\u2013737 (2014). https:\/\/doi.org\/10.1007\/s10489-014-0629-7","journal-title":"Appl. Intell."},{"key":"1_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"1_CR19","doi-asserted-by":"publisher","unstructured":"Shu, H., et al.: A deep learning approach to re-create raw full-field digital mammograms for breast density and texture analysis. Radiol. Artif. Intell. 3(4), e200097 (2021). https:\/\/doi.org\/10.1148\/ryai.2021200097","DOI":"10.1148\/ryai.2021200097"},{"key":"1_CR20","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition (2015)"},{"key":"1_CR21","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: International Conference on Machine Learning, pp. 10347\u201310357. PMLR (2021)"},{"key":"1_CR22","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"1_CR23","doi-asserted-by":"crossref","unstructured":"Wang, W., Chen, C., Ding, M., Li, J., Yu, H., Zha, S.: TransBTS: multimodal brain tumor segmentation using transformer. arXiv preprint arXiv:2103.04430 (2021)","DOI":"10.1007\/978-3-030-87193-2_11"},{"key":"1_CR24","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"1_CR25","doi-asserted-by":"crossref","unstructured":"Wu, H., et al.: CVT: introducing convolutions to vision transformers. arXiv preprint arXiv:2103.15808 (2021)","DOI":"10.1109\/ICCV48922.2021.00009"},{"key":"1_CR26","doi-asserted-by":"crossref","unstructured":"Xie, Y., Zhang, J., Shen, C., Xia, Y.: COTR: efficiently bridging CNN and transformer for 3D medical image segmentation. arXiv preprint arXiv:2103.03024 (2021)","DOI":"10.1007\/978-3-030-87199-4_16"},{"key":"1_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.neuroimage.2020.117368","volume":"223","author":"L Zhong","year":"2020","unstructured":"Zhong, L., et al.: 2WM: tumor segmentation and tract statistics for assessing white matter integrity with applications to glioblastoma patients. Neuroimage 223, 117368 (2020)","journal-title":"Neuroimage"}],"container-title":["Lecture Notes in Computer Science","Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-09002-8_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,11]],"date-time":"2023-02-11T15:42:16Z","timestamp":1676130136000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-09002-8_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031090011","9783031090028"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-09002-8_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"15 July 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"BrainLes","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International MICCAI Brainlesion Workshop","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iwb2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.brainlesion-workshop.org\/?msclkid=7759e32ed14111ecba82c5ba435279db","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"151","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"91","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"60% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1.5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}