{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,8]],"date-time":"2025-06-08T04:01:16Z","timestamp":1749355276856,"version":"3.41.0"},"publisher-location":"Singapore","reference-count":39,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819665983","type":"print"},{"value":"9789819665969","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-6596-9_20","type":"book-chapter","created":{"date-parts":[[2025,6,7]],"date-time":"2025-06-07T05:30:54Z","timestamp":1749274254000},"page":"283-296","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MASR: Efficient Multi-attention Network For Single Image Super-Resolution"],"prefix":"10.1007","author":[{"given":"Tao","family":"Gong","sequence":"first","affiliation":[]},{"given":"Zihao","family":"Jian","sequence":"additional","affiliation":[]},{"given":"Honghao","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Yongxuan","family":"Lai","sequence":"additional","affiliation":[]},{"given":"Liang","family":"Song","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,8]]},"reference":[{"key":"20_CR1","doi-asserted-by":"publisher","unstructured":"Bevilacqua, M., Roumy, A.: Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In: Proceedings of the British Machine Vision Conference 2012 (2012). https:\/\/doi.org\/10.5244\/c.26.135, http:\/\/dx.doi.org\/10.5244\/c.26.135","DOI":"10.5244\/c.26.135"},{"key":"20_CR2","doi-asserted-by":"crossref","unstructured":"Chen, X., Wang, X.: Activating more pixels in image super-resolution transformer (2022)","DOI":"10.1109\/CVPR52729.2023.02142"},{"key":"20_CR3","doi-asserted-by":"crossref","unstructured":"Chen, Z., Zhang, Y.: Dual aggregation transformer for image super-resolution. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 12312\u201312321 (2023)","DOI":"10.1109\/ICCV51070.2023.01131"},{"key":"20_CR4","doi-asserted-by":"publisher","unstructured":"Dai, T., Cai, J.: Second-order attention network for single image super-resolution. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019). https:\/\/doi.org\/10.1109\/cvpr.2019.01132, http:\/\/dx.doi.org\/10.1109\/cvpr.2019.01132","DOI":"10.1109\/cvpr.2019.01132"},{"key":"20_CR5","unstructured":"Dong, C., Loy, C.C., He, K., Tang, X.: Image super-resolution using deep convolutional networks (2014)"},{"key":"20_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"184","DOI":"10.1007\/978-3-319-10593-2_13","volume-title":"Computer Vision \u2013 ECCV 2014","author":"C Dong","year":"2014","unstructured":"Dong, C., Loy, C.C., He, K., Tang, X.: Learning a deep convolutional network for image super-resolution. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8692, pp. 184\u2013199. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10593-2_13"},{"key":"20_CR7","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv: Computer Vision and Pattern Recognition (2020)"},{"key":"20_CR8","doi-asserted-by":"crossref","unstructured":"Han, D., Pan, X.: Flatten transformer: vision transformer using focused linear attention. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 5961\u20135971 (2023)","DOI":"10.1109\/ICCV51070.2023.00548"},{"key":"20_CR9","doi-asserted-by":"crossref","unstructured":"Hassani, A., Walton, S., Li, J., Li, S., Shi, H.: Neighborhood attention transformer (2022)","DOI":"10.1109\/CVPR52729.2023.00599"},{"key":"20_CR10","unstructured":"Katharopoulos, A., Vyas, A., Pappas, N., Fleuret, F.: Transformers are RNNs: Fast autoregressive transformers with linear attention. arXiv: Learning (2020)"},{"key":"20_CR11","doi-asserted-by":"publisher","unstructured":"Kim, J., Lee, J.K., Lee, K.M.: Accurate image super-resolution using very deep convolutional networks. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016). https:\/\/doi.org\/10.1109\/cvpr.2016.182, http:\/\/dx.doi.org\/10.1109\/cvpr.2016.182","DOI":"10.1109\/cvpr.2016.182"},{"key":"20_CR12","doi-asserted-by":"publisher","unstructured":"Kim, J., Lee, J.K., Lee, K.M.: Deeply-recursive convolutional network for image super-resolution. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016). https:\/\/doi.org\/10.1109\/cvpr.2016.181","DOI":"10.1109\/cvpr.2016.181"},{"key":"20_CR13","unstructured":"Kingma, D., Ba, J.: Adam: A method for stochastic optimization. arXiv: Learning (2014)"},{"key":"20_CR14","doi-asserted-by":"crossref","unstructured":"Li, G., Xu, D., Cheng, X., Si, L., Zheng, C.: SimViT: exploring a simple vision transformer with sliding windows. In: 2022 IEEE International Conference on Multimedia and Expo (ICME), pp.\u00a01\u20136. IEEE (2022)","DOI":"10.1109\/ICME52920.2022.9859907"},{"key":"20_CR15","first-page":"14290","volume":"35","author":"G Li","year":"2022","unstructured":"Li, G., Zheng, H., Liu, D., Wang, C., Su, B., Zheng, C.: SemMAE: semantic-guided masking for learning masked autoencoders. Adv. Neural. Inf. Process. Syst. 35, 14290\u201314302 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"20_CR16","doi-asserted-by":"publisher","unstructured":"Liang, J., Cao, J.: SwinIR: image restoration using swin transformer. In: 2021 IEEE\/CVF International Conference on Computer Vision Workshops (ICCVW) (2021). https:\/\/doi.org\/10.1109\/iccvw54120.2021.00210","DOI":"10.1109\/iccvw54120.2021.00210"},{"key":"20_CR17","doi-asserted-by":"publisher","unstructured":"Lim, B., Son, S., Kim, H., Nah, S., Lee, K.M.: Enhanced deep residual networks for single image super-resolution. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (2017). https:\/\/doi.org\/10.1109\/cvprw.2017.151","DOI":"10.1109\/cvprw.2017.151"},{"key":"20_CR18","doi-asserted-by":"crossref","unstructured":"Liu, J., Zhang, W.: Residual feature aggregation network for image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020)","DOI":"10.1109\/CVPR.2019.00399"},{"key":"20_CR19","doi-asserted-by":"publisher","unstructured":"Liu, Z., Lin, Y.: Swin transformer: Hierarchical vision transformer using shifted windows. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV) (2021). https:\/\/doi.org\/10.1109\/iccv48922.2021.00986","DOI":"10.1109\/iccv48922.2021.00986"},{"key":"20_CR20","volume-title":"SOFT: softmax-free transformer with linear complexity","author":"J Lu","year":"2021","unstructured":"Lu, J., et al.: SOFT: softmax-free transformer with linear complexity. Cornell University - arXiv, Cornell University - arXiv (Oct (2021)"},{"key":"20_CR21","doi-asserted-by":"publisher","unstructured":"Martin, D., Fowlkes, C.: A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In: Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001 (2002). https:\/\/doi.org\/10.1109\/iccv.2001.937655","DOI":"10.1109\/iccv.2001.937655"},{"key":"20_CR22","doi-asserted-by":"publisher","unstructured":"Martin, D., Fowlkes, C.: A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In: Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001 (2002). https:\/\/doi.org\/10.1109\/iccv.2001.937655","DOI":"10.1109\/iccv.2001.937655"},{"key":"20_CR23","doi-asserted-by":"publisher","unstructured":"Matsui, Y., Ito: Sketch-based manga retrieval using manga109 dataset. Multimedia Tools Appl. 21811\u201321838 (2017). https:\/\/doi.org\/10.1007\/s11042-016-4020-z","DOI":"10.1007\/s11042-016-4020-z"},{"key":"20_CR24","doi-asserted-by":"publisher","unstructured":"Mei, Y., Fan, Y., Zhou, Y.: Image super-resolution with non-local sparse attention. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2021). https:\/\/doi.org\/10.1109\/cvpr46437.2021.00352","DOI":"10.1109\/cvpr46437.2021.00352"},{"key":"20_CR25","volume-title":"Single Image Super-Resolution via a Holistic Attention Network","author":"B Niu","year":"2020","unstructured":"Niu, B., Wen, W.: Single Image Super-Resolution via a Holistic Attention Network. Cornell University - arXiv, Cornell University - arXiv (2020)"},{"key":"20_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1007\/978-3-030-58610-2_12","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Niu","year":"2020","unstructured":"Niu, B., et al.: Single image super-resolution via a holistic attention network. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 191\u2013207. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_12"},{"key":"20_CR27","doi-asserted-by":"crossref","unstructured":"Shi, W., Caballero, J., Huszar, F.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016)","DOI":"10.1109\/CVPR.2016.207"},{"key":"20_CR28","doi-asserted-by":"publisher","unstructured":"Shi, W., et al.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016). https:\/\/doi.org\/10.1109\/cvpr.2016.207","DOI":"10.1109\/cvpr.2016.207"},{"key":"20_CR29","doi-asserted-by":"publisher","unstructured":"Tai, Y., Yang, J., Liu, X.: Image super-resolution via deep recursive residual network. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017). https:\/\/doi.org\/10.1109\/cvpr.2017.298","DOI":"10.1109\/cvpr.2017.298"},{"key":"20_CR30","doi-asserted-by":"publisher","unstructured":"Tai, Y., Yang, J., Liu, X., Xu, C.: MemNET: a persistent memory network for image restoration. In: 2017 IEEE International Conference on Computer Vision (ICCV) (2017). https:\/\/doi.org\/10.1109\/iccv.2017.486","DOI":"10.1109\/iccv.2017.486"},{"key":"20_CR31","doi-asserted-by":"publisher","unstructured":"Timofte, R., Agustsson, E.: NTIRE 2017 challenge on single image super-resolution: methods and results. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (2017). https:\/\/doi.org\/10.1109\/cvprw.2017.149","DOI":"10.1109\/cvprw.2017.149"},{"key":"20_CR32","volume-title":"Attention is All You Need","author":"A Vaswani","year":"2017","unstructured":"Vaswani, A., Shazeer, N.: Attention is All You Need. Neural Information Processing Systems, Neural Information Processing Systems (2017)"},{"key":"20_CR33","doi-asserted-by":"publisher","unstructured":"Wang, W., et al.: Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV) (2021). https:\/\/doi.org\/10.1109\/iccv48922.2021.00061","DOI":"10.1109\/iccv48922.2021.00061"},{"key":"20_CR34","doi-asserted-by":"publisher","unstructured":"Wang, Z., Bovik, A., Sheikh, H., Simoncelli, E.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 600\u2013612 (2004). https:\/\/doi.org\/10.1109\/tip.2003.819861","DOI":"10.1109\/tip.2003.819861"},{"key":"20_CR35","doi-asserted-by":"publisher","unstructured":"Xiong, Y., et al.: Nystr\u00f6mformer: a nystr\u00f6m-based algorithm for approximating self-attention. Proceedings of the AAAI Conference on Artificial Intelligence, pp. 14138\u201314148 (2022). https:\/\/doi.org\/10.1609\/aaai.v35i16.17664","DOI":"10.1609\/aaai.v35i16.17664"},{"key":"20_CR36","doi-asserted-by":"crossref","unstructured":"You, H., et al.: Castling-ViT: Compressing self-attention via switching towards linear-angular attention during vision transformer inference (2022)","DOI":"10.1109\/CVPR52729.2023.01387"},{"key":"20_CR37","unstructured":"Zhang, X., Zeng, H., Guo, S., Zhang, L.: Efficient long-range attention network for image super-resolution"},{"key":"20_CR38","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Li, K., Li, K., Wang, L., Zhong, B., Fu, Y.: Image super-resolution using very deep residual channel attention networks, pp. 286\u2013301 (2018)","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"20_CR39","first-page":"3499","volume":"33","author":"S Zhou","year":"2020","unstructured":"Zhou, S., Zhang, J., Zuo, W., Loy, C.C.: Cross-scale internal graph neural network for image super-resolution. Adv. Neural. Inf. Process. Syst. 33, 3499\u20133509 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."}],"container-title":["Lecture Notes in Computer Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-6596-9_20","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,7]],"date-time":"2025-06-07T05:31:02Z","timestamp":1749274262000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-6596-9_20"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819665983","9789819665969"],"references-count":39,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-6596-9_20","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"8 June 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Auckland","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/iconip2024.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}