{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,13]],"date-time":"2025-05-13T16:29:33Z","timestamp":1747153773922,"version":"3.40.5"},"reference-count":36,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2022,9,19]],"date-time":"2022-09-19T00:00:00Z","timestamp":1663545600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,9,19]],"date-time":"2022-09-19T00:00:00Z","timestamp":1663545600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2023,3]]},"DOI":"10.1007\/s11042-022-13815-x","type":"journal-article","created":{"date-parts":[[2022,9,19]],"date-time":"2022-09-19T06:02:47Z","timestamp":1663567367000},"page":"12117-12133","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Accurate stereo image super-resolution using spatial-attention-enhance residual network"],"prefix":"10.1007","volume":"82","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4390-7828","authenticated-orcid":false,"given":"Wenyuan","family":"Ying","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2231-1569","authenticated-orcid":false,"given":"Tianyang","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Chen","family":"Shentu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,19]]},"reference":[{"key":"13815_CR1","doi-asserted-by":"crossref","unstructured":"Ahn N, Kang B, Sohn K (2018) Fast, accurate, and lightweight super-resolution with cascading residual network. In: Proceedings of the European conference on computer vision (ECCV), pp 252\u2013268","DOI":"10.1109\/CVPRW.2018.00123"},{"key":"13815_CR2","doi-asserted-by":"publisher","first-page":"1119","DOI":"10.1007\/s11042-019-08218-4","volume":"79","author":"S Barzegar","year":"2020","unstructured":"Barzegar S, Sharifi A, Manthouri M (2020) Super-resolution using lightweight detailnet network. Multimed Tools Appl 79:1119\u20131136","journal-title":"Multimed Tools Appl"},{"issue":"9","key":"13815_CR3","doi-asserted-by":"publisher","first-page":"1721","DOI":"10.1109\/TPAMI.2010.90","volume":"32","author":"AV Bhavsar","year":"2010","unstructured":"Bhavsar AV, Rajagopalan AN (2010) Resolution enhancement in multi-image stereo. IEEE Trans Pattern AnalMach Intell 32(9):1721\u20131728","journal-title":"IEEE Trans Pattern AnalMach Intell"},{"issue":"4","key":"13815_CR4","doi-asserted-by":"publisher","first-page":"596","DOI":"10.1109\/LSP.2018.2815003","volume":"25","author":"K Chang","year":"2018","unstructured":"Chang K, Ding PLK, Li B (2018) Single image super resolution using joint regularization. IEEE Signal Proc Lett 25(4):596\u2013600","journal-title":"IEEE Signal Proc Lett"},{"key":"13815_CR5","doi-asserted-by":"crossref","unstructured":"Chen C, Qing C, Xu X, Dickinson P (2021)\u00a0Cross parallax attention network for stereo image super-resolution. IEEE Trans Multimed 24:202\u2013216","DOI":"10.1109\/TMM.2021.3050092"},{"issue":"7","key":"13815_CR6","doi-asserted-by":"publisher","first-page":"946","DOI":"10.1109\/LSP.2018.2820057","volume":"25","author":"J Chu","year":"2018","unstructured":"Chu J, Zhang J, Lu W, Huang X (2018) A novel multiconnected convolutional network for super-resolution. IEEE Signal Proc Lett 25(7):946\u2013950","journal-title":"IEEE Signal Proc Lett"},{"key":"13815_CR7","doi-asserted-by":"crossref","unstructured":"Dai T, Cai J, Zhang Y, Xia S.-T, Zhang L\u00a0(2019) Second-order attention network for single image super-resolution. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 11065\u201311074","DOI":"10.1109\/CVPR.2019.01132"},{"key":"13815_CR8","doi-asserted-by":"crossref","unstructured":"Dong C, Loy CC, He K, Tang X (2014)\u00a0Learning a deep convolutional network for image super-resolution. In: European conference on computer vision. Springer,\u00a0pp 184\u2013199","DOI":"10.1007\/978-3-319-10593-2_13"},{"key":"13815_CR9","doi-asserted-by":"crossref","unstructured":"Geiger A, Lenz P, Urtasun R (2012) Are we ready for autonomous driving? the kitti vision benchmark suite. In: 2012 IEEE conference on computer vision and pattern recognition. IEEE,\u00a0pp 3354\u20133361","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"13815_CR10","doi-asserted-by":"crossref","unstructured":"Haris M, Shakhnarovich G, Ukita N (2018)\u00a0Deep back-projection networks for super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1664\u20131673","DOI":"10.1109\/CVPR.2018.00179"},{"key":"13815_CR11","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016)\u00a0Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"13815_CR12","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018)\u00a0Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"key":"13815_CR13","doi-asserted-by":"publisher","first-page":"29639","DOI":"10.1007\/s11042-020-09524-y","volume":"79","author":"J Huang","year":"2020","unstructured":"Huang J (2020) Image super-resolution reconstruction based on generative adversarial network model with double discriminators. Multimed Tools Appl 79:29639\u201329662","journal-title":"Multimed Tools Appl"},{"key":"13815_CR14","doi-asserted-by":"crossref","unstructured":"Jeon DS, Baek S-H, Choi I, Kim MH (2018) Enhancing the spatial resolution of stereo images using a parallax prior. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1721\u20131730","DOI":"10.1109\/CVPR.2018.00185"},{"key":"13815_CR15","doi-asserted-by":"crossref","unstructured":"Kim J, Lee JK, Lee KM (2016)\u00a0Accurate image super-resolution using very deep convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1646\u20131654","DOI":"10.1109\/CVPR.2016.182"},{"key":"13815_CR16","doi-asserted-by":"crossref","unstructured":"Kim J, Lee JK, Lee KM (2016)\u00a0Deeply-recursive convolutional network for image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1637\u20131645","DOI":"10.1109\/CVPR.2016.181"},{"key":"13815_CR17","unstructured":"Kingma D, Ba J (2015)\u00a0Adam: a method for stochastic optimization. In: International Conference on Learning Representations (Poster)"},{"key":"13815_CR18","doi-asserted-by":"crossref","unstructured":"Lai W-S, Huang J-B, Ahuja N, Yang M-H\u00a0(2017) Deep laplacian pyramid networks for fast and accurate super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5835\u20135843","DOI":"10.1109\/CVPR.2017.618"},{"key":"13815_CR19","doi-asserted-by":"crossref","unstructured":"Li K, Wu Z, Peng K-C, Ernst J, Fu Y (2018)\u00a0Tell me where to look: Guided attention inference network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 9215\u20139223","DOI":"10.1109\/CVPR.2018.00960"},{"key":"13815_CR20","doi-asserted-by":"crossref","unstructured":"Li F, Cong R, Bai H, He Y (2020)\u00a0Deep interleaved network for single image super-resolution with asymmetric co-attention. In: IJCAI, pp 537\u2013543","DOI":"10.24963\/ijcai.2020\/75"},{"key":"13815_CR21","doi-asserted-by":"crossref","unstructured":"Lim B, Son S, Kim H, Nah Sm, Lee KM (2017) Enhanced deep residual networks for single image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp 1132\u20131140","DOI":"10.1109\/CVPRW.2017.151"},{"key":"13815_CR22","doi-asserted-by":"crossref","unstructured":"Menze M, Geiger A (2015)\u00a0Object scene flow for autonomous vehicles. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3061\u20133070","DOI":"10.1109\/CVPR.2015.7298925"},{"key":"13815_CR23","doi-asserted-by":"crossref","unstructured":"Scharstein D (2014) H. Hirschm\u00a8uller, Y. Kitajima, G. Krathwohl, N. Nesic, X. Wang and P. Westling, \u201chigh-resolution stereo datasets with subpixel-accurate ground truth,\u201d in German Conf. For. Pattern Recogn 8753:31\u201342","DOI":"10.1007\/978-3-319-11752-2_3"},{"key":"13815_CR24","doi-asserted-by":"publisher","first-page":"28087","DOI":"10.1007\/s11042-021-10888-y","volume":"80","author":"P Shen","year":"2021","unstructured":"Shen P, Zhang L, Wang M, Yin G (2021) Deeper super-resolution generative adversarial network with gradient penalty for sonar image enhancement. Multimed Tools Appl 80:28087\u201328107","journal-title":"Multimed Tools Appl"},{"key":"13815_CR25","doi-asserted-by":"crossref","unstructured":"Shi W, Caballero J, Husz\u00e1r F, Totz J, Aitken AP, Bishop R, Rueckert D, Wang Z (2016) Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1637\u20131645","DOI":"10.1109\/CVPR.2016.207"},{"key":"13815_CR26","doi-asserted-by":"publisher","first-page":"9765","DOI":"10.1007\/s11042-020-10152-9","volume":"80","author":"Z Song","year":"2021","unstructured":"Song Z, Zhao X, Jiang H (2021) Gradual deep residual network for super-resolution. Multimed Tools Appl 80:9765\u20139778","journal-title":"Multimed Tools Appl"},{"key":"13815_CR27","doi-asserted-by":"crossref","unstructured":"Tai Y, Yang J, Liu X (2017)\u00a0Image super-resolution via deep recursive residual network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2790\u20132798","DOI":"10.1109\/CVPR.2017.298"},{"key":"13815_CR28","doi-asserted-by":"crossref","unstructured":"Wang F, Jiang M, Qian C, Yang S, C. Li, Zhang H, Wang X, Tang X (2017) Residual attention network for image classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6450\u20136458","DOI":"10.1109\/CVPR.2017.683"},{"key":"13815_CR29","doi-asserted-by":"crossref","unstructured":"Wang L, Wang Y, Liang Z, Lin Z, Yang J, An W, Guo Y (2019) Learning parallax attention for stereo image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 12242\u201312251","DOI":"10.1109\/CVPR.2019.01253"},{"key":"13815_CR30","doi-asserted-by":"crossref","unstructured":"Wang Y, Wang L, Yang J, An W, Guo Y (2019) Flickr1024: A large-scale dataset for stereo image super-resolution. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops","DOI":"10.1109\/ICCVW.2019.00478"},{"key":"13815_CR31","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee J-Y, Kweon IS (2018)\u00a0CBAM: Convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV), pp 3\u201319","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"13815_CR32","doi-asserted-by":"publisher","first-page":"4859","DOI":"10.1007\/s11042-021-11258-4","volume":"81","author":"H Yang","year":"2021","unstructured":"Yang H, Tong J, Dou Q, Xiao L, Jeon G, Yang X (2021) Wide receptive field networks for single image super-resolution. Multimed Tools Appl 81:4859\u20134876","journal-title":"Multimed Tools Appl"},{"key":"13815_CR33","doi-asserted-by":"publisher","first-page":"496","DOI":"10.1109\/LSP.2020.2973813","volume":"27","author":"X Ying","year":"2020","unstructured":"Ying X, Wang Y, Wang L, Sheng W, An W, Guo Y (2020) A stereo attention module for stereo image super-resolution. IEEE Signal Proc Lett 27:496\u2013500","journal-title":"IEEE Signal Proc Lett"},{"key":"13815_CR34","doi-asserted-by":"crossref","unstructured":"Zhang Y, Li K, Li K, Wang L, Zhong B and Fu Y (2018) Image super-resolution using very deep residual channel attention networks. In: Proceedings of the European conference on computervision (ECCV), pp 286\u2013301","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"13815_CR35","doi-asserted-by":"crossref","unstructured":"Zhang Y, Tian Y, Kong Y, Zhong B, Fu Y(2018)\u00a0Residual dense network for image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2472\u20132481","DOI":"10.1109\/CVPR.2018.00262"},{"key":"13815_CR36","doi-asserted-by":"publisher","unstructured":"Zhu X, Guo K, Fang H, Chen L, Ren S Hu B (n.d.) \u201cCross View Capture for Stereo Image Super-Resolution,\u201d in IEEE Transactions on Multimedia, https:\/\/doi.org\/10.1109\/TMM.2021.3092571","DOI":"10.1109\/TMM.2021.3092571"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-022-13815-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-022-13815-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-022-13815-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,3]],"date-time":"2023-03-03T09:54:03Z","timestamp":1677837243000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-022-13815-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,9,19]]},"references-count":36,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2023,3]]}},"alternative-id":["13815"],"URL":"https:\/\/doi.org\/10.1007\/s11042-022-13815-x","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"type":"print","value":"1380-7501"},{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2022,9,19]]},"assertion":[{"value":"17 September 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 December 2021","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 September 2022","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 September 2022","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}