{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T14:48:31Z","timestamp":1768747711329,"version":"3.49.0"},"reference-count":33,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2021,5,19]],"date-time":"2021-05-19T00:00:00Z","timestamp":1621382400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,5,19]],"date-time":"2021-05-19T00:00:00Z","timestamp":1621382400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61672292"],"award-info":[{"award-number":["61672292"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61532009"],"award-info":[{"award-number":["61532009"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2022,2]]},"DOI":"10.1007\/s00530-021-00796-4","type":"journal-article","created":{"date-parts":[[2021,5,19]],"date-time":"2021-05-19T11:02:43Z","timestamp":1621422163000},"page":"131-139","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":12,"title":["Visual saliency prediction using multi-scale attention gated network"],"prefix":"10.1007","volume":"28","author":[{"given":"Yubao","family":"Sun","sequence":"first","affiliation":[]},{"given":"Mengyang","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Shaojing","family":"Fan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,5,19]]},"reference":[{"key":"796_CR1","unstructured":"Borji, A., Itti, L.: Cat2000: A large scale fixation dataset for boosting saliency research. arXiv preprint. arXiv:1505.03581 (2015)"},{"issue":"3","key":"796_CR2","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1109\/TPAMI.2018.2815601","volume":"41","author":"Z Bylinskii","year":"2018","unstructured":"Bylinskii, Z., Judd, T., Oliva, A., Torralba, A., Durand, F.: What do different evaluation metrics tell us about saliency models? IEEE Trans. Pattern Anal. Mach. Intell. 41(3), 740\u2013757 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"796_CR3","doi-asserted-by":"crossref","unstructured":"Cornia, M., Baraldi, L., Serra, G., Cucchiara, R.: A deep multi-level network for saliency prediction. In: 2016 23rd International Conference on Pattern Recognition (ICPR), pp. 3488\u20133493. IEEE (2016)","DOI":"10.1109\/ICPR.2016.7900174"},{"issue":"10","key":"796_CR4","doi-asserted-by":"publisher","first-page":"5142","DOI":"10.1109\/TIP.2018.2851672","volume":"27","author":"M Cornia","year":"2018","unstructured":"Cornia, M., Baraldi, L., Serra, G., Cucchiara, R.: Predicting human eye fixations via an lstm-based saliency attentive model. IEEE Trans. Image Process. 27(10), 5142\u20135154 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"796_CR5","doi-asserted-by":"crossref","unstructured":"Fan, S., Shen, Z., Jiang, M., Koenig, B.L., Xu, J., Kankanhalli, M.S., Zhao, Q.: Emotional attention: a study of image sentiment and visual attention. In: Proceedings of the IEEE Conference on computer vision and pattern recognition, pp. 7521\u20137531 (2018)","DOI":"10.1109\/CVPR.2018.00785"},{"key":"796_CR6","doi-asserted-by":"crossref","unstructured":"Harel, J., Koch, C., Perona, P.: Graph-based visual saliency. In: Proceedings of the Conference on Neural Information processing Systems (NIPS), vol.\u00a019, pp. 545\u2013552 (2007)","DOI":"10.7551\/mitpress\/7503.003.0073"},{"key":"796_CR7","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"796_CR8","doi-asserted-by":"crossref","unstructured":"Huang, X., Shen, C., Boix, X., Zhao, Q.: Salicon: Reducing the semantic gap in saliency prediction by adapting deep neural networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 262\u2013270 (2015)","DOI":"10.1109\/ICCV.2015.38"},{"issue":"11","key":"796_CR9","doi-asserted-by":"publisher","first-page":"1254","DOI":"10.1109\/34.730558","volume":"20","author":"L Itti","year":"1998","unstructured":"Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 20(11), 1254\u20131259 (1998)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"796_CR10","doi-asserted-by":"crossref","unstructured":"Jetley, S., Murray, N., Vig, E.: End-to-end saliency mapping via probability distribution prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5753\u20135761 (2016)","DOI":"10.1109\/CVPR.2016.620"},{"key":"796_CR11","doi-asserted-by":"crossref","unstructured":"Jiang, M., Huang, S., Duan, J., Zhao, Q.: Salicon: Saliency in context. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1072\u20131080 (2015)","DOI":"10.1109\/CVPR.2015.7298710"},{"key":"796_CR12","doi-asserted-by":"crossref","unstructured":"Koch, C., Ullman, S.: Shifts in selective visual attention: towards the underlying neural circuitry. In: Matters of intelligence, pp. 115\u2013141. Springer (1987)","DOI":"10.1007\/978-94-009-3833-5_5"},{"issue":"9","key":"796_CR13","doi-asserted-by":"publisher","first-page":"4446","DOI":"10.1109\/TIP.2017.2710620","volume":"26","author":"SS Kruthiventi","year":"2017","unstructured":"Kruthiventi, S.S., Ayush, K., Babu, R.V.: Deepfix: a fully convolutional neural network for predicting human eye fixations. IEEE Trans. Image Process. 26(9), 4446\u20134456 (2017)","journal-title":"IEEE Trans. Image Process."},{"key":"796_CR14","unstructured":"K\u00fcmmerer, M., Theis, L., Bethge, M.: Deep gaze i: boosting saliency prediction with feature maps trained on imagenet. arXiv preprint. arXiv:1411.1045 (2014)"},{"issue":"52","key":"796_CR15","doi-asserted-by":"publisher","first-page":"16054","DOI":"10.1073\/pnas.1510393112","volume":"112","author":"M K\u00fcmmerer","year":"2015","unstructured":"K\u00fcmmerer, M., Wallis, T.S., Bethge, M.: Information-theoretic model comparison unifies saliency metrics. Proc. Natl. Acad. Sci. 112(52), 16054\u201316059 (2015)","journal-title":"Proc. Natl. Acad. Sci."},{"issue":"1","key":"796_CR16","doi-asserted-by":"publisher","first-page":"251","DOI":"10.3758\/s13428-012-0226-9","volume":"45","author":"O Le Meur","year":"2013","unstructured":"Le Meur, O., Baccino, T.: Methods for comparing scanpaths and saliency maps: strengths and weaknesses. Behav. Res. Methods 45(1), 251\u2013266 (2013)","journal-title":"Behav. Res. Methods"},{"issue":"19","key":"796_CR17","doi-asserted-by":"publisher","first-page":"2483","DOI":"10.1016\/j.visres.2007.06.015","volume":"47","author":"O Le Meur","year":"2007","unstructured":"Le Meur, O., Le Callet, P., Barba, D.: Predicting visual fixations on video based on low-level visual features. Vis. Res. 47(19), 2483\u20132498 (2007)","journal-title":"Vis. Res."},{"issue":"2","key":"796_CR18","doi-asserted-by":"publisher","first-page":"392","DOI":"10.1109\/TNNLS.2016.2628878","volume":"29","author":"N Liu","year":"2016","unstructured":"Liu, N., Han, J., Liu, T., Li, X.: Learning to predict eye fixations via multiresolution convolutional neural networks. IEEE Trans. Neural Netw. Learn. Syst. 29(2), 392\u2013404 (2016)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"796_CR19","unstructured":"Pan, J., Ferrer, C.C., McGuinness, K., O\u2019Connor, N.E., Torres, J., Sayrol, E., Giro-i Nieto, X.: Salgan: Visual saliency prediction with generative adversarial networks. arXiv preprint. arXiv:1701.01081 (2017)"},{"key":"796_CR20","doi-asserted-by":"crossref","unstructured":"Pan, J., Sayrol, E., Giro-i Nieto, X., McGuinness, K., O\u2019Connor, N.E.: Shallow and deep convolutional networks for saliency prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 598\u2013606 (2016)","DOI":"10.1109\/CVPR.2016.71"},{"issue":"18","key":"796_CR21","doi-asserted-by":"publisher","first-page":"2397","DOI":"10.1016\/j.visres.2005.03.019","volume":"45","author":"RJ Peters","year":"2005","unstructured":"Peters, R.J., Iyer, A., Itti, L., Koch, C.: Components of bottom-up gaze allocation in natural images. Vis. Res. 45(18), 2397\u20132416 (2005)","journal-title":"Vis. Res."},{"key":"796_CR22","doi-asserted-by":"crossref","unstructured":"Ramanathan, S., Katti, H., Sebe, N., Kankanhalli, M., Chua, T.S.: An eye fixation database for saliency detection in images. In: European Conference on Computer Vision, pp. 30\u201343. Springer (2010)","DOI":"10.1007\/978-3-642-15561-1_3"},{"key":"796_CR23","doi-asserted-by":"crossref","unstructured":"Sun, K., Xiao, B., Liu, D., Wang, J.: Deep high-resolution representation learning for human pose estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5693\u20135703 (2019)","DOI":"10.1109\/CVPR.2019.00584"},{"issue":"1","key":"796_CR24","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1007\/BF00130487","volume":"7","author":"MJ Swain","year":"1991","unstructured":"Swain, M.J., Ballard, D.H.: Color indexing. Int. J. Comput. Vis. 7(1), 11\u201332 (1991)","journal-title":"Int. J. Comput. Vis."},{"issue":"12","key":"796_CR25","doi-asserted-by":"publisher","first-page":"1736","DOI":"10.1109\/LSP.2016.2617340","volume":"23","author":"H Tang","year":"2016","unstructured":"Tang, H., Chen, C., Pei, X.: Visual saliency detection via sparse residual and outlier detection. IEEE Signal Process. Lett. 23(12), 1736\u20131740 (2016)","journal-title":"IEEE Signal Process. Lett."},{"issue":"5","key":"796_CR26","doi-asserted-by":"publisher","first-page":"643","DOI":"10.1016\/j.visres.2004.09.017","volume":"45","author":"BW Tatler","year":"2005","unstructured":"Tatler, B.W., Baddeley, R.J., Gilchrist, I.D.: Visual correlates of fixation selection: effects of scale and time. Vis. Res. 45(5), 643\u2013659 (2005)","journal-title":"Vis. Res."},{"issue":"1","key":"796_CR27","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1016\/0010-0285(80)90005-5","volume":"12","author":"AM Treisman","year":"1980","unstructured":"Treisman, A.M., Gelade, G.: A feature-integration theory of attention. Cogn. Psychol. 12(1), 97\u2013136 (1980)","journal-title":"Cogn. Psychol."},{"key":"796_CR28","doi-asserted-by":"crossref","unstructured":"Vig, E., Dorr, M., Cox, D.: Large-scale optimization of hierarchical features for saliency prediction in natural images. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2798\u20132805 (2014)","DOI":"10.1109\/CVPR.2014.358"},{"issue":"5","key":"796_CR29","doi-asserted-by":"publisher","first-page":"2368","DOI":"10.1109\/TIP.2017.2787612","volume":"27","author":"W Wang","year":"2017","unstructured":"Wang, W., Shen, J.: Deep visual attention prediction. IEEE Trans. Image Process. 27(5), 2368\u20132378 (2017)","journal-title":"IEEE Trans. Image Process."},{"issue":"9","key":"796_CR30","doi-asserted-by":"publisher","first-page":"e24038","DOI":"10.1371\/journal.pone.0024038","volume":"6","author":"N Wilming","year":"2011","unstructured":"Wilming, N., Betz, T., Kietzmann, T.C., K\u00f6nig, P.: Measures and limits of models of fixation selection. PLoS One 6(9), e24038 (2011)","journal-title":"PLoS One"},{"issue":"3","key":"796_CR31","doi-asserted-by":"publisher","first-page":"419","DOI":"10.1037\/0096-1523.15.3.419","volume":"15","author":"JM Wolfe","year":"1989","unstructured":"Wolfe, J.M., Cave, K.R., Franzel, S.L.: Guided search: an alternative to the feature integration model for visual search. J. Exp. Psychol. Hum. Percept. Perform. 15(3), 419 (1989)","journal-title":"J. Exp. Psychol. Hum. Percept. Perform."},{"key":"796_CR32","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., So\u00a0Kweon, I.: Cbam: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"796_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, J., Sclaroff, S.: Saliency detection: a boolean map approach. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 153\u2013160 (2013)","DOI":"10.1109\/ICCV.2013.26"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-021-00796-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-021-00796-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-021-00796-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,3]],"date-time":"2023-11-03T22:18:51Z","timestamp":1699049931000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-021-00796-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,5,19]]},"references-count":33,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2022,2]]}},"alternative-id":["796"],"URL":"https:\/\/doi.org\/10.1007\/s00530-021-00796-4","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,5,19]]},"assertion":[{"value":"25 October 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 April 2021","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 May 2021","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}