{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,24]],"date-time":"2026-01-24T18:50:53Z","timestamp":1769280653076,"version":"3.49.0"},"publisher-location":"Cham","reference-count":32,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031164453","type":"print"},{"value":"9783031164460","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-16446-0_49","type":"book-chapter","created":{"date-parts":[[2022,9,16]],"date-time":"2022-09-16T09:02:47Z","timestamp":1663318967000},"page":"516-526","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":27,"title":["3D CVT-GAN: A 3D Convolutional Vision Transformer-GAN for PET Reconstruction"],"prefix":"10.1007","author":[{"given":"Pinxian","family":"Zeng","sequence":"first","affiliation":[]},{"given":"Luping","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Chen","family":"Zu","sequence":"additional","affiliation":[]},{"given":"Xinyi","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Zhengyang","family":"Jiao","sequence":"additional","affiliation":[]},{"given":"Xi","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Jiliu","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Dinggang","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,17]]},"reference":[{"key":"49_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"783","DOI":"10.1007\/978-3-030-59728-3_76","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2020","author":"Q Feng","year":"2020","unstructured":"Feng, Q., Liu, H.: Rethinking PET image reconstruction: ultra-low-dose, sinogram and deep learning. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12267, pp. 783\u2013792. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59728-3_76"},{"issue":"3","key":"49_CR2","doi-asserted-by":"publisher","first-page":"569","DOI":"10.1109\/TBME.2016.2564440","volume":"64","author":"Y Wang","year":"2016","unstructured":"Wang, Y., Ma, G., An, L., et al.: Semi-supervised tripled dictionary learning for standard-dose PET image prediction using low-dose PET and multimodal MRI. IEEE Trans. Biomed. Eng. 64(3), 569\u2013579 (2016)","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"6","key":"49_CR3","doi-asserted-by":"publisher","first-page":"1478","DOI":"10.1109\/TMI.2018.2832613","volume":"37","author":"K Kim","year":"2018","unstructured":"Kim, K., Wu, D., Gong, K., et al.: Penalized PET reconstruction using deep learning prior and local linear fitting. IEEE Trans. Med. Imaging 37(6), 1478\u20131487 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"6","key":"49_CR4","doi-asserted-by":"publisher","first-page":"1328","DOI":"10.1109\/TMI.2018.2884053","volume":"38","author":"Y Wang","year":"2018","unstructured":"Wang, Y., Zhou, L., Yu, B., et al.: 3D auto-context-based locality adaptive multi-modality GANs for PET synthesis. IEEE Trans. Med. Imaging 38(6), 1328\u20131339 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"49_CR5","doi-asserted-by":"publisher","first-page":"406","DOI":"10.1016\/j.neucom.2017.06.048","volume":"267","author":"L Xiang","year":"2017","unstructured":"Xiang, L., Qiao, Y., Nie, D., et al.: Deep auto-context convolutional neural networks for standard-dose PET image estimation from low-dose PET\/MRI. Neurocomputing 267, 406\u2013416 (2017)","journal-title":"Neurocomputing"},{"issue":"10","key":"49_CR6","doi-asserted-by":"publisher","first-page":"4928","DOI":"10.1002\/mp.14402","volume":"47","author":"K Spuhler","year":"2020","unstructured":"Spuhler, K., Serrano-Sosa, M., Cattell, R., et al.: Full-count PET recovery from low-count image using a dilated convolutional neural network. Med. Phys. 47(10), 4928\u20134938 (2020)","journal-title":"Med. Phys."},{"key":"49_CR7","doi-asserted-by":"publisher","first-page":"550","DOI":"10.1016\/j.neuroimage.2018.03.045","volume":"174","author":"Y Wang","year":"2018","unstructured":"Wang, Y., Yu, B., Wang, L., et al.: 3D conditional generative adversarial networks for high-quality PET image estimation at low dose. Neuroimage 174, 550\u2013562 (2018)","journal-title":"Neuroimage"},{"issue":"3","key":"49_CR8","doi-asserted-by":"publisher","first-page":"675","DOI":"10.1109\/TMI.2018.2869871","volume":"38","author":"K Gong","year":"2018","unstructured":"Gong, K., Guan, J., Kim, K., et al.: Iterative PET image reconstruction using convolutional neural network representation. IEEE Trans. Med. Imaging 38(3), 675\u2013685 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"49_CR9","doi-asserted-by":"publisher","first-page":"102339","DOI":"10.1016\/j.media.2021.102339","volume":"77","author":"B Zhan","year":"2022","unstructured":"Zhan, B., Xiao, J., Cao, C., et al.: Multi-constraint generative adversarial network for dose prediction in radiotherapy. Med. Image Anal. 77, 102339 (2022)","journal-title":"Med. Image Anal."},{"key":"49_CR10","doi-asserted-by":"publisher","first-page":"253","DOI":"10.1016\/j.media.2019.03.013","volume":"54","author":"I H\u00e4ggstr\u00f6m","year":"2019","unstructured":"H\u00e4ggstr\u00f6m, I., Schmidtlein, C.R., et al.: DeepPET: a deep encoder-decoder network for directly solving the PET image reconstruction inverse problem. Med. Image Anal. 54, 253\u2013262 (2019)","journal-title":"Med. Image Anal."},{"key":"49_CR11","doi-asserted-by":"publisher","first-page":"108021","DOI":"10.1016\/j.knosys.2021.108021","volume":"239","author":"L Hu","year":"2022","unstructured":"Hu, L., Li, J., Peng, X., et al.: Semi-supervised NPC segmentation with uncertainty and attention guided consistency. Knowl.-Based Syst. 239, 108021 (2022)","journal-title":"Knowl.-Based Syst."},{"issue":"1","key":"49_CR12","doi-asserted-by":"publisher","first-page":"54","DOI":"10.1109\/TRPMS.2020.3004408","volume":"5","author":"A Mehranian","year":"2020","unstructured":"Mehranian, A., Reader, A.J.: Model-based deep learning PET image reconstruction using forward-backward splitting expectation-maximization. IEEE Trans. Radiat. Plasma Med. Sci. 5(1), 54\u201364 (2020)","journal-title":"IEEE Trans. Radiat. Plasma Med. Sci."},{"key":"49_CR13","doi-asserted-by":"publisher","first-page":"108215","DOI":"10.1016\/j.knosys.2022.108215","volume":"241","author":"P Tang","year":"2022","unstructured":"Tang, P., Yang, P., et al.: Unified medical image segmentation by learning from uncertainty in an end-to-end manner. Knowl. Based Syst. 241, 108215 (2022)","journal-title":"Knowl. Based Syst."},{"key":"49_CR14","doi-asserted-by":"publisher","first-page":"101770","DOI":"10.1016\/j.media.2020.101770","volume":"65","author":"L Zhou","year":"2020","unstructured":"Zhou, L., Schaefferkoetter, J.D., et al.: Supervised learning with cyclegan for low-dose FDG PET image denoising. Med. Image Anal. 65, 101770 (2020)","journal-title":"Med. Image Anal."},{"key":"49_CR15","doi-asserted-by":"publisher","first-page":"102335","DOI":"10.1016\/j.media.2021.102335","volume":"77","author":"Y Luo","year":"2022","unstructured":"Luo, Y., Zhou, L., Zhan, B., et al.: Adaptive rectification based adversarial network with spectrum constraint for high-quality PET image synthesis. Med. Image Anal. 77, 102335 (2022)","journal-title":"Med. Image Anal."},{"key":"49_CR16","doi-asserted-by":"publisher","first-page":"102447","DOI":"10.1016\/j.media.2022.102447","volume":"79","author":"K Wang","year":"2022","unstructured":"Wang, K., Zhan, B., Zu, C., et al.: Semi-supervised medical image segmentation via a tripled-uncertainty guided mean teacher model with contrastive learning. Med. Image Anal. 79, 102447 (2022)","journal-title":"Med. Image Anal."},{"issue":"3","key":"49_CR17","doi-asserted-by":"publisher","first-page":"1123","DOI":"10.1109\/TCYB.2018.2797905","volume":"49","author":"D Nie","year":"2018","unstructured":"Nie, D., Wang, L., Adeli, E., et al.: 3D fully convolutional networks for multimodal isointense infant brain image segmentation. IEEE Trans. Cybern. 49(3), 1123\u20131136 (2018)","journal-title":"IEEE Trans. Cybern."},{"key":"49_CR18","doi-asserted-by":"publisher","first-page":"108566","DOI":"10.1016\/j.patcog.2022.108566","volume":"126","author":"Y Shi","year":"2022","unstructured":"Shi, Y., Zu, C., Hong, M., et al.: ASMFS: Adaptive-similarity-based multi-modality feature selection for classification of Alzheimer\u2019s disease. Pattern Recogn. 126, 108566 (2022)","journal-title":"Pattern Recogn."},{"key":"49_CR19","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al.: An image is worth 16 x 16 words: Transformers for image recognition at scale. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. IEEE, Venice (2020)"},{"key":"49_CR20","unstructured":"Hugo T., Matthieu C., et al.: Training data-efficient image transformers & distillation through attention. In: Proceedings of the 38th International Conference on Machine Learning, pp. 10347\u201310357. PMLR, Vienna (2021)"},{"key":"49_CR21","doi-asserted-by":"crossref","unstructured":"Wang, W., Xie, E., Li, X., et al.: Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 568\u2013578. IEEE, Montreal (2021)","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"49_CR22","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1007\/978-3-030-87231-1_6","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"Z Zhang","year":"2021","unstructured":"Zhang, Z., Yu, L., Liang, X., Zhao, W., Xing, L.: TransCT: dual-path transformer for low dose computed tomography. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12906, pp. 55\u201364. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87231-1_6"},{"key":"49_CR23","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"276","DOI":"10.1007\/978-3-030-87231-1_27","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"Y Luo","year":"2021","unstructured":"Luo, Y., et al.: 3D transformer-GAN for high-quality PET reconstruction. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12906, pp. 276\u2013285. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87231-1_27"},{"key":"49_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1007\/978-3-030-87193-2_11","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"W Wang","year":"2021","unstructured":"Wang, W., Chen, C., Ding, M., Yu, H., Zha, S., Li, J.: TransBTS: multimodal brain tumor segmentation using transformer. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12901, pp. 109\u2013119. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_11"},{"key":"49_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1007\/978-3-030-87193-2_2","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Liu, H., Hu, Q.: Transfuse: fusing transformers and CNNs for medical image segmentation. In: de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C. (eds.) MICCAI 2021. LNCS, vol. 12901, pp. 14\u201324. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_2"},{"key":"49_CR26","unstructured":"Chen, J., Lu, Y., Yu, Q., et al.: TransuNet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:2102.04306 (2021)"},{"key":"49_CR27","doi-asserted-by":"crossref","unstructured":"Gao, Y., Zhou, M., Metaxas, D.N.: UTNet: a hybrid transformer architecture for medical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 61\u201371. Springer, Cham (2021)","DOI":"10.1007\/978-3-030-87199-4_6"},{"key":"49_CR28","unstructured":"Luthra, A., Sulakhe, H., Mittal, T., et al.: Eformer: edge enhancement based transformer for medical image denoising. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021 (2021)"},{"key":"49_CR29","doi-asserted-by":"crossref","unstructured":"Wu, H., Xiao, B., Codella, N., et al.: CVT: introducing convolutions to vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 22\u201331 (2021)","DOI":"10.1109\/ICCV48922.2021.00009"},{"key":"49_CR30","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"186","DOI":"10.1007\/978-3-030-18305-9_15","volume-title":"Advances in Artificial Intelligence","author":"R Ye","year":"2019","unstructured":"Ye, R., Liu, F., Zhang, L.: 3D depthwise convolution: reducing model parameters in 3D vision tasks. In: Meurs, M., Rudzicz, F. (eds.) Canadian AI 2019. LNCS (LNAI), vol. 11489, pp. 186\u2013199. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-18305-9_15"},{"key":"49_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"424","DOI":"10.1007\/978-3-319-46723-8_49","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2016","author":"\u00d6 \u00c7i\u00e7ek","year":"2016","unstructured":"\u00c7i\u00e7ek, \u00d6., Abdulkadir, A., Lienkamp, S.S., Brox, T., Ronneberger, O.: 3D U-Net: learning dense volumetric segmentation from sparse annotation. In: Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W. (eds.) MICCAI 2016. LNCS, vol. 9901, pp. 424\u2013432. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46723-8_49"},{"key":"49_CR32","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, p. 30 (2017)"}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-16446-0_49","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,19]],"date-time":"2024-07-19T07:09:12Z","timestamp":1721372952000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-16446-0_49"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031164453","9783031164460"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-16446-0_49","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"17 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conferences.miccai.org\/2022\/en\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft Conference","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1831","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"574","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"31% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}