{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T20:02:15Z","timestamp":1776888135525,"version":"3.51.2"},"publisher-location":"Cham","reference-count":43,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031189098","type":"print"},{"value":"9783031189104","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-18910-4_16","type":"book-chapter","created":{"date-parts":[[2022,10,26]],"date-time":"2022-10-26T23:03:53Z","timestamp":1666825433000},"page":"191-203","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Cascade Multiscale Swin-Conv Network for Fast MRI Reconstruction"],"prefix":"10.1007","author":[{"given":"Shengcheng","family":"Ye","sequence":"first","affiliation":[]},{"given":"Xinyu","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Dongping","family":"Xiong","sequence":"additional","affiliation":[]},{"given":"Lijun","family":"Ouyang","sequence":"additional","affiliation":[]},{"given":"Xiaozhi","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,27]]},"reference":[{"key":"16_CR1","doi-asserted-by":"publisher","first-page":"560","DOI":"10.1109\/TSP.2002.807005","volume":"51","author":"JA Fessler","year":"2013","unstructured":"Fessler, J.A., Sutton, B.P.: Nonuniform fast Fourier transforms using min-max interpolation. IEEE Trans. Signal Process 51, 560\u2013574 (2013)","journal-title":"IEEE Trans. Signal Process"},{"issue":"1","key":"16_CR2","doi-asserted-by":"publisher","first-page":"61","DOI":"10.3390\/diagnostics11010061","volume":"11","author":"J Lv","year":"2021","unstructured":"Lv, J., Wang, C., Yang, G.: PIC-GAN: a parallel imaging coupled generative adversarial network for accelerated multi-channel MRI reconstruction. Diagnostics 11(1), 61 (2021)","journal-title":"Diagnostics"},{"key":"16_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2019.104969","volume":"188","author":"Z Chen","year":"2020","unstructured":"Chen, Z., Huang, C., Lin, S.: A new sparse representation framework for compressed sensing MRI. Knowl. Based Syst. 188, 104969 (2020)","journal-title":"Knowl. Based Syst."},{"issue":"4","key":"16_CR4","doi-asserted-by":"publisher","first-page":"893","DOI":"10.1109\/TMI.2010.2085084","volume":"30","author":"JP Haldar","year":"2010","unstructured":"Haldar, J.P., Hernando, D., Liang, Z.P.: Compressed-sensing MRI with random encoding. IEEE Trans. Med. Imaging 30(4), 893\u2013903 (2010)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"9","key":"16_CR5","doi-asserted-by":"publisher","first-page":"1985","DOI":"10.1109\/TBME.2018.2821699","volume":"65","author":"D Lee","year":"2018","unstructured":"Lee, D., Yoo, J., Tak, S., Ye, J.C.: Deep residual learning for accelerated MRI using magnitude and phase networks. IEEE Trans. Biomed. Eng. 65(9), 1985\u20131995 (2018)","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"2","key":"16_CR6","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee, F., Jaeger, P.F., Kohl, S.A., et al.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203\u2013211 (2021)","journal-title":"Nat. Methods"},{"key":"16_CR7","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","volume":"73","author":"H Li","year":"2021","unstructured":"Li, H., Wu, X.J., Kittler, J.: RFN-Nest: an end-to-end residual fusion network for infrared and visible images. Inf. Fus. 73, 72\u201386 (2021)","journal-title":"Inf. Fus."},{"key":"16_CR8","doi-asserted-by":"crossref","unstructured":"Tzinis, E., Wang, Z., Smaragdis, P.: Sudo RM-RF: efficient networks for universal audio source separation. In: 2020 IEEE 30th International Workshop on Machine Learning for Signal Processing (MLSP), pp. 1\u20136 (2020)","DOI":"10.1109\/MLSP49062.2020.9231900"},{"issue":"2","key":"16_CR9","first-page":"1097","volume":"25","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25(2), 1097\u20131105 (2012)","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"1","key":"16_CR10","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1109\/MSP.2019.2950557","volume":"37","author":"D Liang","year":"2020","unstructured":"Liang, D., Cheng, J., Ke, Z., et al.: Deep magnetic resonance image reconstruction: inverse problems meet neural networks. IEEE Signal Process. Mag. 37(1), 141\u2013151 (2020)","journal-title":"IEEE Signal Process. Mag."},{"key":"16_CR11","doi-asserted-by":"crossref","unstructured":"Wang, S., Su, Z., Ying, L., et al.: Accelerating magnetic resonance imaging via deep learning. In: 2016 IEEE 13th International Symposium on Biomedical Imaging, pp. 514\u2013517 (2016)","DOI":"10.1109\/ISBI.2016.7493320"},{"key":"16_CR12","unstructured":"Yang, Y., Sun, J., Li, H., et al.: Deep ADMM-Net for compressive sensing MRI. In: Proceedings of the 30th International Conference on Neural Information Processing Systems, pp. 10\u201318 (2016)"},{"issue":"2","key":"16_CR13","doi-asserted-by":"publisher","first-page":"491","DOI":"10.1109\/TMI.2017.2760978","volume":"37","author":"J Schlemper","year":"2017","unstructured":"Schlemper, J., Caballero, J., Hajnal, J.V., et al.: A deep cascade of convolutional neural networks for dynamic MR image reconstruction. IEEE Trans. Med. Imaging 37(2), 491\u2013503 (2017)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"13","key":"16_CR14","doi-asserted-by":"publisher","DOI":"10.1088\/1361-6560\/aac71a","volume":"63","author":"CM Hyun","year":"2018","unstructured":"Hyun, C.M., Kim, H.P., Lee, S.M., et al.: Deep learning for undersampled MRI re-construction. Phys. Med. Biol. 63(13), 135007 (2018)","journal-title":"Phys. Med. Biol."},{"key":"16_CR15","doi-asserted-by":"publisher","first-page":"317","DOI":"10.1016\/j.ins.2019.03.080","volume":"490","author":"Y Wu","year":"2019","unstructured":"Wu, Y., Ma, Y., Liu, J., et al.: Self-attention convolutional neural network for improved MR image reconstruction. Inf. Sci. 490, 317\u2013328 (2019)","journal-title":"Inf. Sci."},{"key":"16_CR16","unstructured":"Liang, D., Cheng, J., Ke, Z., et al.: Deep MRI reconstruction: Unrolled optimization algorithms meet neural networks. arXiv preprint arXiv:1907.11711 (2019)"},{"issue":"9","key":"16_CR17","doi-asserted-by":"publisher","first-page":"2772","DOI":"10.1109\/TMI.2020.2975344","volume":"39","author":"T Zhou","year":"2020","unstructured":"Zhou, T., Fu, H., Chen, G., et al.: Hi-Net: hybrid-fusion network for multi-modal mr image synthesis. IEEE Trans. Med. Imaging 39(9), 2772\u20132781 (2020)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"16_CR18","doi-asserted-by":"publisher","first-page":"93","DOI":"10.1016\/j.mri.2019.03.012","volume":"66","author":"Y Wu","year":"2020","unstructured":"Wu, Y., Ma, Y., Capaldi, D.P., et al.: Incorporating prior knowledge via volumetric deep residual network to optimize the reconstruction of sparsely sampled MRI. Magn. Reson. Imaging 66, 93\u2013103 (2020)","journal-title":"Magn. Reson. Imaging"},{"key":"16_CR19","unstructured":"Wang, P., Chen, E.Z., Chen, T., et al.: Pyramid convolutional RNN for MRI reconstruction. arXiv preprint arXiv:1912.00543 (2019)"},{"issue":"6","key":"16_CR20","doi-asserted-by":"publisher","first-page":"1310","DOI":"10.1109\/TMI.2017.2785879","volume":"37","author":"G Yang","year":"2018","unstructured":"Yang, G., Yu, S., Dong, H., et al.: DAGAN: deep de-aliasing generative adversarial networks for fast compressed sensing MRI reconstruction. IEEE Trans. Med. Imaging 37(6), 1310\u20131321 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"6","key":"16_CR21","doi-asserted-by":"publisher","first-page":"1488","DOI":"10.1109\/TMI.2018.2820120","volume":"37","author":"TM Quan","year":"2018","unstructured":"Quan, T.M., Nguyen-Duc, T., Jeong, W.K.: Compressed sensing MRI reconstruction using a generative adversarial network with a cyclic loss. IEEE Trans. Med. Imaging 37(6), 1488\u20131497 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"1","key":"16_CR22","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1109\/TMI.2018.2858752","volume":"38","author":"M Mardani","year":"2019","unstructured":"Mardani, M., Gong, E., Cheng, J.Y., et al.: Deep generative adversarial neural networks for compressive sensing MRI. IEEE Trans. Med. Imaging 38(1), 167\u2013179 (2019)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"16_CR23","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"16_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2019.115701","volume":"81","author":"W Zeng","year":"2020","unstructured":"Zeng, W., Peng, J., Wang, S., et al.: A comparative study of CNN-based super-resolution methods in MRI reconstruction and its beyond. Signal Process. Image Commun. 81, 115701 (2020)","journal-title":"Signal Process. Image Commun."},{"key":"16_CR25","doi-asserted-by":"publisher","first-page":"338","DOI":"10.1016\/j.neucom.2019.02.070","volume":"398","author":"X Li","year":"2020","unstructured":"Li, X., Cao, T., Tong, X., et al.: Deep residual network for highly accelerated fMRI reconstruction using variable density spiral trajectory. Neurocomputing 398, 338\u2013346 (2020)","journal-title":"Neurocomputing"},{"key":"16_CR26","doi-asserted-by":"publisher","first-page":"232","DOI":"10.1016\/j.jmr.2019.07.020","volume":"305","author":"L Bao","year":"2019","unstructured":"Bao, L., Ye, F., Cai, C., et al.: Undersampled MR image reconstruction using an enhanced recursive residual network. J. Magn. Reson. 305, 232\u2013246 (2019)","journal-title":"J. Magn. Reson."},{"key":"16_CR27","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Laurens, V., et al.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2016)","DOI":"10.1109\/CVPR.2017.243"},{"key":"16_CR28","unstructured":"Parmar, N., Vaswani, A., Uszkoreit, J., et al.: Image transformer. In: International Conference on Machine Learning, pp. 4055\u20134064 (2018)"},{"key":"16_CR29","unstructured":"Matsoukas, C., Haslum, J.F., S\u00f6derberg, M., et al.: Is it time to replace CNNs with transformers for medical images? arXiv preprint arXiv:2108.09038 (2021)"},{"key":"16_CR30","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"16_CR31","unstructured":"Ho, J., Kalchbrenner, N., Weissenborn, D., et al.: Axial attention in multidimensional transformers. arXiv preprint arXiv:1912.12180 (2019)"},{"key":"16_CR32","doi-asserted-by":"crossref","unstructured":"Strudel, R., Garcia, R., Laptev, I., et al.: Segmenter: transformer for semantic segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7262\u20137272 (2021)","DOI":"10.1109\/ICCV48922.2021.00717"},{"key":"16_CR33","doi-asserted-by":"crossref","unstructured":"Shamshad, F., Khan, S., Zamir, S.W., et al.: Transformers in medical imaging: a survey. arXiv preprint arXiv:2201.09873 (2022)","DOI":"10.1016\/j.media.2023.102802"},{"key":"16_CR34","doi-asserted-by":"crossref","unstructured":"Arnab, A., Dehghani, M., Heigold, G., et al.: ViViT: a video vision transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6836\u20136846 (2021)","DOI":"10.1109\/ICCV48922.2021.00676"},{"key":"16_CR35","doi-asserted-by":"crossref","unstructured":"Hatamizadeh, A., Yang, D., Roth, H., et al.: UNETR: transformers for 3D medical image segmentation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 574\u2013584 (2021)","DOI":"10.1109\/WACV51458.2022.00181"},{"key":"16_CR36","doi-asserted-by":"crossref","unstructured":"Chen, C.F.R., Fan, Q., Panda, R.: CrossViT: cross-attention multi-scale vision transformer for image classification. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 357\u2013366 (2021)","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"16_CR37","unstructured":"Feng, C.M., Yan, Y., Chen, G., et al.: Accelerated multi-modal mr imaging with transformers. arXiv preprint arXiv:2106.14248 (2021)"},{"key":"16_CR38","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"16_CR39","unstructured":"Cao, H., Wang, Y., Chen, J., et al.: Swin-Unet: Unet-like pure transformer for medical image segmentation. arXiv preprint arXiv:2105.05537 (2021)"},{"key":"16_CR40","doi-asserted-by":"crossref","unstructured":"Liang, J., Cao, J., Sun, G., et al.: SwinIR: image restoration using swin transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1833\u20131844 (2021)","DOI":"10.1109\/ICCVW54120.2021.00210"},{"key":"16_CR41","doi-asserted-by":"publisher","first-page":"281","DOI":"10.1016\/j.neucom.2022.04.051","volume":"493","author":"J Huang","year":"2022","unstructured":"Huang, J., Fang, Y., Wu, Y., et al.: Swin transformer for fast MRI. Neurocomputing 493, 281\u2013304 (2022)","journal-title":"Neurocomputing"},{"key":"16_CR42","doi-asserted-by":"crossref","unstructured":"Huang, J., Fang, Y., Wu, Y., et al.: Fast MRI reconstruction: how powerful transformers are? arXiv preprint arXiv:2201.09400 (2022)","DOI":"10.1109\/EMBC48229.2022.9871475"},{"key":"16_CR43","doi-asserted-by":"crossref","unstructured":"MASI Lab. https:\/\/my.vanderbilt.edu\/masi\/workshops\/. Accessed 14 Apr 2022","DOI":"10.3390\/toxins14090588"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-18910-4_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,29]],"date-time":"2023-11-29T16:48:47Z","timestamp":1701276527000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-18910-4_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031189098","9783031189104"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-18910-4_16","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"27 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shenzhen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/en.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"microsoft","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"564","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"233","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"41% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.03","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.35","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}