{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T09:40:50Z","timestamp":1742982050607,"version":"3.40.3"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031250651"},{"type":"electronic","value":"9783031250668"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25066-8_13","type":"book-chapter","created":{"date-parts":[[2023,2,17]],"date-time":"2023-02-17T08:18:05Z","timestamp":1676621885000},"page":"268-280","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Complementary Phase Encoding for\u00a0Pair-Wise Neural Deblurring of\u00a0Accelerated Brain MRI"],"prefix":"10.1007","author":[{"given":"Gali","family":"Hod","sequence":"first","affiliation":[]},{"given":"Michael","family":"Green","sequence":"additional","affiliation":[]},{"given":"Mark","family":"Waserman","sequence":"additional","affiliation":[]},{"given":"Eli","family":"Konen","sequence":"additional","affiliation":[]},{"given":"Shai","family":"Shrot","sequence":"additional","affiliation":[]},{"given":"Ilya","family":"Nelkenbaum","sequence":"additional","affiliation":[]},{"given":"Nahum","family":"Kiryati","sequence":"additional","affiliation":[]},{"given":"Arnaldo","family":"Mayer","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,18]]},"reference":[{"key":"13_CR1","unstructured":"Accelmr 2020 prediction challenge (2020). https:\/\/accelmrorg.wordpress.com\/"},{"key":"13_CR2","doi-asserted-by":"publisher","unstructured":"Bevilacqua, M., Roumy, A., Guillemot, C., line Alberi Morel, M.: Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In: Proceedings of the British Machine Vision Conference, pp. 135.1\u2013135.10. BMVA Press (2012). https:\/\/doi.org\/10.5244\/C.26.135","DOI":"10.5244\/C.26.135"},{"issue":"4","key":"13_CR3","doi-asserted-by":"publisher","first-page":"751","DOI":"10.1088\/0031-9155\/25\/4\/017","volume":"25","author":"WA Edelstein","year":"1980","unstructured":"Edelstein, W.A., Hutchison, J.M.S., Johnson, G., Redpath, T.: Spin warp NMR imaging and applications to human whole-body imaging. Phys. Med. Biol. 25(4), 751\u2013756 (1980). https:\/\/doi.org\/10.1088\/0031-9155\/25\/4\/017","journal-title":"Phys. Med. Biol."},{"key":"13_CR4","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition (2015). https:\/\/doi.org\/10.48550\/ARXIV.1512.03385, https:\/\/arxiv.org\/abs\/1512.03385","DOI":"10.48550\/ARXIV.1512.03385"},{"issue":"21","key":"13_CR5","doi-asserted-by":"publisher","first-page":"297","DOI":"10.1088\/0031-9155\/60\/21\/R297","volume":"60","author":"KG Hollingsworth","year":"2015","unstructured":"Hollingsworth, K.G.: Reducing acquisition time in clinical MRI by data undersampling and compressed sensing reconstruction. Phys. Med. Biol. 60(21), 297\u2013322 (2015)","journal-title":"Phys. Med. Biol."},{"key":"13_CR6","unstructured":"Iwamoto, Y., Takeda, K., Li, Y., Shiino, A., Chen, Y.W.: Unsupervised MRI super-resolution using deep external learning and guided residual dense network with multimodal image priors (2020)"},{"key":"13_CR7","doi-asserted-by":"publisher","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution (2016). https:\/\/doi.org\/10.48550\/ARXIV.1603.08155, https:\/\/arxiv.org\/abs\/1603.08155","DOI":"10.48550\/ARXIV.1603.08155"},{"key":"13_CR8","doi-asserted-by":"publisher","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization (2014). https:\/\/doi.org\/10.48550\/ARXIV.1412.6980, https:\/\/arxiv.org\/abs\/1412.6980","DOI":"10.48550\/ARXIV.1412.6980"},{"key":"13_CR9","doi-asserted-by":"publisher","unstructured":"Kiryati, N., Landau, Y.: Dataset growth in medical image analysis research. J. Imaging 7, 155 (2021). https:\/\/doi.org\/10.3390\/jimaging7080155","DOI":"10.3390\/jimaging7080155"},{"key":"13_CR10","doi-asserted-by":"crossref","unstructured":"Klein*, S., Staring*, M., Murphy, K., Viergever, M.A., Pluim, J.P.: elastix: a toolbox for intensity-based medical image registration. IEEE Trans. Medical Imag. 29(1), 196\u2013205 (2010)","DOI":"10.1109\/TMI.2009.2035616"},{"key":"13_CR11","doi-asserted-by":"publisher","unstructured":"Knoll, F., et al.: Advancing machine learning for MR image reconstruction with an open competition: Overview of the 2019 fastMRI challenge. Magn. Reson. Med. 84(6), 3054\u20133070 (2020). https:\/\/doi.org\/10.1002\/mrm.28338","DOI":"10.1002\/mrm.28338"},{"key":"13_CR12","doi-asserted-by":"publisher","unstructured":"Kupyn, O., Budzan, V., Mykhailych, M., Mishkin, D., Matas, J.: Deblurgan: blind motion deblurring using conditional adversarial networks. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8183\u20138192 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00854","DOI":"10.1109\/CVPR.2018.00854"},{"key":"13_CR13","doi-asserted-by":"crossref","unstructured":"Kupyn, O., Martyniuk, T., Wu, J., Wang, Z.: Deblurgan-v2: deblurring (orders-of-magnitude) faster and better. In: The IEEE International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00897"},{"key":"13_CR14","doi-asserted-by":"publisher","unstructured":"Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network (2016). https:\/\/doi.org\/10.48550\/ARXIV.1609.04802, https:\/\/arxiv.org\/abs\/1609.04802","DOI":"10.48550\/ARXIV.1609.04802"},{"key":"13_CR15","doi-asserted-by":"publisher","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts (2016). https:\/\/doi.org\/10.48550\/ARXIV.1608.03983, https:\/\/arxiv.org\/abs\/1608.03983","DOI":"10.48550\/ARXIV.1608.03983"},{"issue":"6","key":"13_CR16","doi-asserted-by":"publisher","first-page":"1182","DOI":"10.1002\/mrm.21391","volume":"58","author":"M Lustig","year":"2007","unstructured":"Lustig, M., Donoho, D., Pauly, J.M.: Sparse MRI: the application of compressed sensing for rapid MR imaging. Magn. Reson. Med. 58(6), 1182\u20131195 (2007). https:\/\/doi.org\/10.1002\/mrm.21391","journal-title":"Magn. Reson. Med."},{"issue":"2","key":"13_CR17","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1109\/MSP.2007.914728","volume":"25","author":"M Lustig","year":"2008","unstructured":"Lustig, M., Donoho, D.L., Santos, J.M., Pauly, J.M.: Compressed sensing MRI. IEEE Signal Process. Maga. 25(2), 72\u201382 (2008). https:\/\/doi.org\/10.1109\/MSP.2007.914728","journal-title":"IEEE Signal Process. Maga."},{"issue":"2","key":"13_CR18","doi-asserted-by":"publisher","first-page":"315","DOI":"10.1007\/s11548-021-02535-6","volume":"17","author":"M Mayberg","year":"2021","unstructured":"Mayberg, M., et al.: Anisotropic neural deblurring for MRI acceleration. Int. J. Comput. Assist. Radiol. Surg. 17(2), 315\u2013327 (2021). https:\/\/doi.org\/10.1007\/s11548-021-02535-6","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"13_CR19","doi-asserted-by":"publisher","unstructured":"Martin, D., Fowlkes, C., Tal, D., Malik, J.: A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In: Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001, vol. 2, pp. 416\u2013423 (2001). https:\/\/doi.org\/10.1109\/ICCV.2001.937655","DOI":"10.1109\/ICCV.2001.937655"},{"key":"13_CR20","doi-asserted-by":"publisher","unstructured":"Mei, Y., et al.: Pyramid attention networks for image restoration (2020). https:\/\/doi.org\/10.48550\/ARXIV.2004.13824, https:\/\/arxiv.org\/abs\/2004.13824","DOI":"10.48550\/ARXIV.2004.13824"},{"key":"13_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.compmedimag.2019.101647","volume":"77","author":"CH Pham","year":"2019","unstructured":"Pham, C.H., Tor-D\u00edez, C., Meunier, H., Bednarek, N., Fablet, R., Passat, N., Rousseau, F.: Multiscale brain MRI super-resolution using deep 3D convolutional networks. Comput. Med. Imaging Graph. 77, 101647 (2019). https:\/\/doi.org\/10.1016\/j.compmedimag.2019.101647","journal-title":"Comput. Med. Imaging Graph."},{"issue":"50","key":"13_CR22","first-page":"1","volume":"7","author":"DP Shamonin","year":"2014","unstructured":"Shamonin, D.P., Bron, E.E., Lelieveldt, B.P., Smits, M., Klein, S., Staring, M.: Fast parallel image registration on cpu and gpu for diagnostic classification of alzheimer\u2019s disease. Front. Neuroinf. 7(50), 1\u201315 (2014)","journal-title":"Front. Neuroinf."},{"key":"13_CR23","doi-asserted-by":"publisher","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition (2014). https:\/\/doi.org\/10.48550\/ARXIV.1409.1556, https:\/\/arxiv.org\/abs\/1409.1556","DOI":"10.48550\/ARXIV.1409.1556"},{"key":"13_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1007\/978-3-030-87231-1_41","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"E Tsiligianni","year":"2021","unstructured":"Tsiligianni, E., Zerva, M., Marivani, I., Deligiannis, N., Kondi, L.: Interpretable deep learning for\u00a0multimodal super-resolution of\u00a0medical images. In: de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C. (eds.) MICCAI 2021. LNCS, vol. 12906, pp. 421\u2013429. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87231-1_41"},{"issue":"4","key":"13_CR25","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A., Sheikh, H., Simoncelli, E.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004). https:\/\/doi.org\/10.1109\/TIP.2003.819861","journal-title":"IEEE Trans. Image Process."},{"issue":"7","key":"13_CR26","doi-asserted-by":"publisher","first-page":"2105","DOI":"10.1109\/TBME.2018.2883958","volume":"66","author":"L Xiang","year":"2019","unstructured":"Xiang, L., et al.: Deep-learning-based multi-modal fusion for fast mr reconstruction. IEEE Trans. Biomed. Eng. 66(7), 2105\u20132114 (2019). https:\/\/doi.org\/10.1109\/TBME.2018.2883958","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"6","key":"13_CR27","doi-asserted-by":"publisher","first-page":"1310","DOI":"10.1109\/TMI.2017.2785879","volume":"37","author":"G Yang","year":"2018","unstructured":"Yang, G., et al.: Dagan: deep de-aliasing generative adversarial networks for fast compressed sensing mri reconstruction. IEEE Trans. Med. Imaging 37(6), 1310\u20131321 (2018). https:\/\/doi.org\/10.1109\/TMI.2017.2785879","journal-title":"IEEE Trans. Med. Imaging"},{"key":"13_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"711","DOI":"10.1007\/978-3-642-27413-8_47","volume-title":"Curves and Surfaces","author":"R Zeyde","year":"2012","unstructured":"Zeyde, R., Elad, M., Protter, M.: On single image scale-up using sparse-representations. In: Boissonnat, J.-D., Chenin, P., Cohen, A., Gout, C., Lyche, T., Mazure, M.-L., Schumaker, L. (eds.) Curves and Surfaces 2010. LNCS, vol. 6920, pp. 711\u2013730. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-27413-8_47"},{"key":"13_CR29","doi-asserted-by":"publisher","unstructured":"Zhou, B., Zhou, S.K.: Dudornet: learning a dual-domain recurrent network for fast mri reconstruction with deep t1 prior (2020). https:\/\/doi.org\/10.48550\/ARXIV.2001.03799, https:\/\/arxiv.org\/abs\/2001.03799","DOI":"10.48550\/ARXIV.2001.03799"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25066-8_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T13:18:37Z","timestamp":1709817517000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25066-8_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031250651","9783031250668"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25066-8_13","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"18 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}