{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T16:02:23Z","timestamp":1772121743531,"version":"3.50.1"},"publisher-location":"Cham","reference-count":35,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783319464534","type":"print"},{"value":"9783319464541","type":"electronic"}],"license":[{"start":{"date-parts":[[2016,1,1]],"date-time":"2016-01-01T00:00:00Z","timestamp":1451606400000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2016,1,1]],"date-time":"2016-01-01T00:00:00Z","timestamp":1451606400000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2016]]},"DOI":"10.1007\/978-3-319-46454-1_49","type":"book-chapter","created":{"date-parts":[[2016,9,15]],"date-time":"2016-09-15T09:15:09Z","timestamp":1473930909000},"page":"809-824","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":86,"title":["Where Should Saliency Models Look Next?"],"prefix":"10.1007","author":[{"given":"Zoya","family":"Bylinskii","sequence":"first","affiliation":[]},{"given":"Adri\u00e0","family":"Recasens","sequence":"additional","affiliation":[]},{"given":"Ali","family":"Borji","sequence":"additional","affiliation":[]},{"given":"Aude","family":"Oliva","sequence":"additional","affiliation":[]},{"given":"Antonio","family":"Torralba","sequence":"additional","affiliation":[]},{"given":"Fr\u00e9do","family":"Durand","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2016,9,16]]},"reference":[{"key":"49_CR1","unstructured":"Bylinskii, Z., Judd, T., Borji, A., Itti, L., Durand, F., Oliva, A., Torralba, A.: MIT saliency benchmark. http:\/\/saliency.mit.edu\/"},{"key":"49_CR2","doi-asserted-by":"crossref","unstructured":"Kienzle, W., Wichmann, F.A., Franz, M.O., Sch\u00f6lkopf, B.: A nonparametric approach to bottom-up visual saliency. In: Advances in Neural Information Processing Systems, pp. 689\u2013696 (2006)","DOI":"10.7551\/mitpress\/7503.003.0091"},{"key":"49_CR3","doi-asserted-by":"crossref","unstructured":"Judd, T., Ehinger, K., Durand, F., Torralba, A.: Learning to predict where humans look. In: IEEE 12th International Conference on Computer Vision, pp. 2106\u20132113 (2009)","DOI":"10.1109\/ICCV.2009.5459462"},{"key":"49_CR4","doi-asserted-by":"crossref","unstructured":"Borji, A.: Boosting bottom-up and top-down visual features for saliency estimation. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 438\u2013445 (2012)","DOI":"10.1109\/CVPR.2012.6247706"},{"issue":"1","key":"49_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1167\/14.1.28","volume":"14","author":"J Xu","year":"2014","unstructured":"Xu, J., Jiang, M., Wang, S., Kankanhalli, M.S., Zhao, Q.: Predicting human gaze beyond pixels. J. Vis. 14(1), 1\u201320 (2014)","journal-title":"J. Vis."},{"issue":"3","key":"49_CR6","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1167\/11.3.9","volume":"11","author":"Q Zhao","year":"2011","unstructured":"Zhao, Q., Koch, C.: Learning a saliency map using fixated locations in natural scenes. J. Vis. 11(3), 9 (2011)","journal-title":"J. Vis."},{"issue":"1","key":"49_CR7","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1016\/0010-0285(80)90005-5","volume":"12","author":"AM Treisman","year":"1980","unstructured":"Treisman, A.M., Gelade, G.: A feature-integration theory of attention. Cogn. Psychol. 12(1), 97\u2013136 (1980)","journal-title":"Cogn. Psychol."},{"key":"49_CR8","first-page":"219","volume":"4","author":"C Koch","year":"1985","unstructured":"Koch, C., Ullman, S.: Shifts in selective visual attention: towards the underlying neural circuitry. Hum. Neurbiology 4, 219\u2013227 (1985)","journal-title":"Hum. Neurbiology"},{"key":"49_CR9","doi-asserted-by":"publisher","first-page":"1254","DOI":"10.1109\/34.730558","volume":"11","author":"L Itti","year":"1998","unstructured":"Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 11, 1254\u20131259 (1998)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"49_CR10","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1016\/S0042-6989(01)00250-4","volume":"42","author":"D Parkhurst","year":"2002","unstructured":"Parkhurst, D., Law, K., Niebur, E.: Modeling the role of salience in the allocation of overt visual attention. Vis. Res. 42(1), 107\u2013123 (2002)","journal-title":"Vis. Res."},{"issue":"9","key":"49_CR11","doi-asserted-by":"publisher","first-page":"950","DOI":"10.1167\/7.9.950","volume":"7","author":"N Bruce","year":"2007","unstructured":"Bruce, N., Tsotsos, J.: Attention based on information maximization. J. Vis. 7(9), 950 (2007)","journal-title":"J. Vis."},{"issue":"1","key":"49_CR12","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1109\/TPAMI.2012.89","volume":"35","author":"A Borji","year":"2013","unstructured":"Borji, A., Itti, L.: State-of-the-art in visual attention modeling. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 185\u2013207 (2013)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"49_CR13","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1109\/TIP.2012.2210727","volume":"22","author":"A Borji","year":"2013","unstructured":"Borji, A., Sihite, D.N., Itti, L.: Quantitative analysis of human-model agreement in visual saliency modeling: a comparative study. IEEE Trans. Image Process. 22(1), 55\u201369 (2013)","journal-title":"IEEE Trans. Image Process."},{"key":"49_CR14","doi-asserted-by":"crossref","unstructured":"Vig, E., Dorr, M., Cox, D.: Large-scale optimization of hierarchical features for saliency prediction in natural images. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2798\u20132805 (2014)","DOI":"10.1109\/CVPR.2014.358"},{"key":"49_CR15","unstructured":"K\u00fcmmerer, M., Theis, L., Bethge, M.: Deep Gaze I: Boosting saliency prediction with feature maps trained on ImageNet. arXiv preprint (2014). arXiv:1411.1045"},{"key":"49_CR16","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097\u20131105 (2012)"},{"key":"49_CR17","doi-asserted-by":"crossref","unstructured":"Liu, N., Han, J., Zhang, D., Wen, S., Liu, T.: Predicting eye fixations using convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 362\u2013370 (2015)","DOI":"10.1109\/CVPR.2015.7298633"},{"key":"49_CR18","doi-asserted-by":"crossref","unstructured":"Jiang, M., Huang, S., Duan, J., Zhao, Q.: Salicon: saliency in context. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2015","DOI":"10.1109\/CVPR.2015.7298710"},{"key":"49_CR19","unstructured":"Kruthiventi, S.S., Ayush, K., Babu, R.V.: Deepfix: A fully convolutional neural network for predicting human eye fixations. arXiv preprint (2015). arXiv:1510.02927"},{"key":"49_CR20","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint (2014). arXiv:1409.1556"},{"key":"49_CR21","unstructured":"Borji, A., Itti, L.: Cat2000: A large scale fixation dataset for boosting saliency research. arXiv preprint (2015). arXiv:1505.03581"},{"key":"49_CR22","doi-asserted-by":"crossref","unstructured":"Pan, J., Sayrol, E., Giro-i-Nieto, X., McGuinness, K., O\u2019Connor, N.E.: Shallow and deep convolutional networks for saliency prediction. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016","DOI":"10.1109\/CVPR.2016.71"},{"key":"49_CR23","doi-asserted-by":"crossref","unstructured":"Zhao, R., Ouyang, W., Li, H., Wang, X.: Saliency detection by multi-context deep learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1265\u20131274 (2015)","DOI":"10.1109\/CVPR.2015.7298731"},{"key":"49_CR24","doi-asserted-by":"crossref","unstructured":"Li, G., Yu, Y.: Visual saliency based on multiscale deep features. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5455\u20135463 (2015)","DOI":"10.1109\/CVPR.2015.7299184"},{"key":"49_CR25","doi-asserted-by":"crossref","unstructured":"Wang, L., Lu, H., Ruan, X., Yang, M.H.: Deep networks for saliency detection via local estimation and global search. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3183\u20133192 (2015)","DOI":"10.1109\/CVPR.2015.7298938"},{"issue":"8","key":"49_CR26","doi-asserted-by":"publisher","first-page":"3919","DOI":"10.1109\/TIP.2016.2579306","volume":"25","author":"X Li","year":"2016","unstructured":"Li, X., Zhao, L., Wei, L., Yang, M.H., Wu, F., Zhuang, Y., Ling, H., Wang, J.: Deepsaliency: multi-task deep neural network model for salient object detection. IEEE Trans. Image Process. 25(8), 3919\u20133930 (2016)","journal-title":"IEEE Trans. Image Process."},{"key":"49_CR27","unstructured":"Judd, T., Durand, F., Torralba, A.: A benchmark of computational models of saliency to predict human fixations. In: MIT Technical report (2012)"},{"key":"49_CR28","doi-asserted-by":"crossref","unstructured":"Yao, B., Jiang, X., Khosla, A., Lin, A.L., Guibas, L., Fei-Fei, L.: Human action recognition by learning bases of action attributes and parts. In: IEEE International Conference on Computer Vision (ICCV), pp. 1331\u20131338 (2011)","DOI":"10.1109\/ICCV.2011.6126386"},{"key":"49_CR29","doi-asserted-by":"crossref","unstructured":"Xiao, J., Hays, J., Ehinger, K.A., Oliva, A., Torralba, A.: Sun database: large-scale scene recognition from abbey to zoo. In: IEEE conference on Computer Vision and Pattern Recognition (CVPR), pp. 3485\u20133492 (2010)","DOI":"10.1109\/CVPR.2010.5539970"},{"key":"49_CR30","doi-asserted-by":"crossref","unstructured":"Zhang, J., Sclaroff, S.: Saliency detection: a boolean map approach. In: IEEE International Conference on Computer Vision (2013)","DOI":"10.1109\/ICCV.2013.26"},{"issue":"52","key":"49_CR31","doi-asserted-by":"publisher","first-page":"16054","DOI":"10.1073\/pnas.1510393112","volume":"112","author":"M K\u00fcmmerer","year":"2015","unstructured":"K\u00fcmmerer, M., Wallis, T.S., Bethge, M.: Information-theoretic model comparison unifies saliency metrics. Proc. Nat. Acad. Sci. 112(52), 16054\u201316059 (2015)","journal-title":"Proc. Nat. Acad. Sci."},{"key":"49_CR32","unstructured":"Bylinskii, Z., Judd, T., Oliva, A., Torralba, A., Durand, F.: What do different evaluation metrics tell us about saliency models? arXiv preprint (2016). arXiv:1604.03605"},{"issue":"10","key":"49_CR33","first-page":"1","volume":"12","author":"M Cerf","year":"2009","unstructured":"Cerf, M., Frady, E.P., Koch, C.: Faces and text attract gaze independent of the task: experimental data and computer model. J. Vis. 12(10), 1\u201315 (2009)","journal-title":"J. Vis."},{"key":"49_CR34","unstructured":"Recasens, A., Khosla, A., Vondrick, C., Torralba, A.: Where are they looking? In: Advances in Neural Information Processing Systems, pp. 199\u2013207 (2015)"},{"key":"49_CR35","doi-asserted-by":"crossref","unstructured":"Soo Park, H., Shi, J.: Social saliency prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4777\u20134785 (2015)","DOI":"10.1109\/CVPR.2015.7299110"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2016"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-319-46454-1_49","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,10]],"date-time":"2025-06-10T19:01:22Z","timestamp":1749582082000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-319-46454-1_49"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2016]]},"ISBN":["9783319464534","9783319464541"],"references-count":35,"URL":"https:\/\/doi.org\/10.1007\/978-3-319-46454-1_49","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2016]]},"assertion":[{"value":"16 September 2016","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Amsterdam","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"The Netherlands","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2016","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2016","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 October 2016","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2016","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.eccv2016.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"This content has been made available to all.","name":"free","label":"Free to read"}]}}