{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T18:51:22Z","timestamp":1742928682835,"version":"3.40.3"},"publisher-location":"Cham","reference-count":37,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030226428"},{"type":"electronic","value":"9783030226435"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-22643-5_26","type":"book-chapter","created":{"date-parts":[[2019,7,9]],"date-time":"2019-07-09T23:03:41Z","timestamp":1562713421000},"page":"332-347","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["The Assessment of Sencogi: A Visual Complexity Model Predicting Visual Fixations"],"prefix":"10.1007","author":[{"given":"Maria Laura","family":"Mele","sequence":"first","affiliation":[]},{"given":"Silvia","family":"Colabrese","sequence":"additional","affiliation":[]},{"given":"Luca","family":"Calabria","sequence":"additional","affiliation":[]},{"given":"Damon","family":"Millar","sequence":"additional","affiliation":[]},{"given":"Christiaan Erik","family":"Rijnders","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,6,27]]},"reference":[{"issue":"1","key":"26_CR1","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1016\/0010-0285(80)90005-5","volume":"12","author":"AM Treisman","year":"1980","unstructured":"Treisman, A.M., Gelade, G.: A feature-integration theory of attention. Cogn. Psychol. 12(1), 97\u2013136 (1980). https:\/\/doi.org\/10.1016\/0010-0285(80)90005-5","journal-title":"Cogn. Psychol."},{"key":"26_CR2","series-title":"Synthese Library (Studies in Epistemology, Logic, Methodology, and Philosophy of Science)","doi-asserted-by":"publisher","first-page":"115","DOI":"10.1007\/978-94-009-3833-5_5","volume-title":"Matters of Intelligence","author":"C Koch","year":"1987","unstructured":"Koch, C., Ullman, S.: Shifts in selective visual attention: towards the underlying neural circuitry. In: Vaina, L.M. (ed.) Matters of Intelligence. Synthese Library (Studies in Epistemology, Logic, Methodology, and Philosophy of Science), vol. 188, pp. 115\u2013141. Springer, Dordrecht (1987). https:\/\/doi.org\/10.1007\/978-94-009-3833-5_5"},{"key":"26_CR3","doi-asserted-by":"publisher","unstructured":"Kummerer, M., Wallis, T.S., Gatys, L.A., Bethge, M.: Understanding low-and high-level contributions to fixation prediction. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4789\u20134798 (2017). https:\/\/doi.org\/10.1109\/iccv.2017.513","DOI":"10.1109\/iccv.2017.513"},{"issue":"3","key":"26_CR4","doi-asserted-by":"publisher","first-page":"194","DOI":"10.1038\/35058500","volume":"2","author":"L Itti","year":"2001","unstructured":"Itti, L., Koch, C.: Computational modelling of visual attention. Nat. Rev. Neurosci. 2(3), 194 (2001). https:\/\/doi.org\/10.1038\/35058500","journal-title":"Nat. Rev. Neurosci."},{"key":"26_CR5","doi-asserted-by":"publisher","first-page":"271","DOI":"10.1016\/j.image.2018.10.005","volume":"70","author":"YY Zhang","year":"2019","unstructured":"Zhang, Y.Y., Zhang, S., Zhang, P., Zhang, X.: Saliency detection via background and foreground null space learning. Sig. Process. Image Commun. 70, 271\u2013281 (2019). https:\/\/doi.org\/10.1016\/j.image.2018.10.005","journal-title":"Sig. Process. Image Commun."},{"key":"26_CR6","doi-asserted-by":"publisher","first-page":"62","DOI":"10.1016\/j.visres.2012.06.001","volume":"65","author":"J Shen","year":"2012","unstructured":"Shen, J., Itti, L.: Top-down influences on visual attention during listening are modulated by observer sex. Vision. Res. 65, 62\u201376 (2012). https:\/\/doi.org\/10.1016\/j.visres.2012.06.001","journal-title":"Vision. Res."},{"key":"26_CR7","doi-asserted-by":"publisher","first-page":"62","DOI":"10.1016\/j.visres.2013.07.016","volume":"91","author":"A Borji","year":"2013","unstructured":"Borji, A., Sihite, D.N., Itti, L.: What stands out in a scene? A study of human explicit saliency judgment. Vision. Res. 91, 62\u201377 (2013). https:\/\/doi.org\/10.1016\/j.visres.2013.07.016","journal-title":"Vision. Res."},{"issue":"3","key":"26_CR8","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1167\/14.3.14","volume":"14","author":"K Koehler","year":"2014","unstructured":"Koehler, K., Guo, F., Zhang, S., Eckstein, M.P.: What do saliency models predict? J. Vis. 14(3), 14 (2014). https:\/\/doi.org\/10.1167\/14.3.14","journal-title":"J. Vis."},{"issue":"6","key":"26_CR9","doi-asserted-by":"publisher","first-page":"1266","DOI":"10.1109\/TNNLS.2015.2461603","volume":"27","author":"W Zhang","year":"2016","unstructured":"Zhang, W., Borji, A., Wang, Z., Le Callet, P., Liu, H.: The application of visual saliency models in objective image quality assessment: a statistical evaluation. IEEE Trans. Neural Netw. Learn. Syst. 27(6), 1266\u20131278 (2016). https:\/\/doi.org\/10.1109\/TNNLS.2015.2461603","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"26_CR10","doi-asserted-by":"publisher","unstructured":"Muddamsetty, S.M., Sidibe, D., Tremeau, A., Meriaudeau, F.: Spatio-temporal saliency detection in dynamic scenes using local binary patterns. In: 2014 22nd International Conference on Pattern Recognition (2014). https:\/\/doi.org\/10.1109\/icpr.2014.408","DOI":"10.1109\/icpr.2014.408"},{"key":"26_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-23862-3","volume-title":"Intelligence Science and Big Data Engineering. Big Data and Machine Learning Techniques","year":"2015","unstructured":"He, X. (ed.): IScIDE 2015. LNCS, vol. 9243. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-23862-3"},{"key":"26_CR12","doi-asserted-by":"publisher","first-page":"241","DOI":"10.1007\/s12559-010-9094-8","volume":"3","author":"T Yubing","year":"2011","unstructured":"Yubing, T., Cheikh, F.A., Guraya, F.F.E., Konik, H., Tr\u00e9meau, A.: A spatiotemporal saliency model for video surveillance. Cogn. Comput. 3, 241\u2013263 (2011). https:\/\/doi.org\/10.1007\/s12559-010-9094-8","journal-title":"Cogn. Comput."},{"key":"26_CR13","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1016\/j.jvcir.2018.11.012","volume":"58","author":"N Mu","year":"2019","unstructured":"Mu, N., Xu, X., Zhang, X.: A spatial-frequency-temporal domain based saliency model for low contrast video sequences. J. Vis. Commun. Image Represent. 58, 79\u201388 (2019). https:\/\/doi.org\/10.1016\/j.jvcir.2018.11.012","journal-title":"J. Vis. Commun. Image Represent."},{"key":"26_CR14","doi-asserted-by":"publisher","unstructured":"Rapantzikos, K., Avrithis, Y., Kollias, S.: Dense saliency-based spatiotemporal feature points for action recognition (2009). https:\/\/doi.org\/10.1109\/cvpr.2009.5206525","DOI":"10.1109\/cvpr.2009.5206525"},{"issue":"1","key":"26_CR15","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1007\/978-3-642-37431-9_41","volume":"32","author":"V Mahadevan","year":"2010","unstructured":"Mahadevan, V., Vasconcelos, N.: Spatiotemporal saliency in dynamic scenes. IEEE Trans. Pattern Anal. Mach. Intell. 32(1), 171\u2013177 (2010). https:\/\/doi.org\/10.1007\/978-3-642-37431-9_41","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"26_CR16","doi-asserted-by":"publisher","first-page":"710","DOI":"10.1109\/tsmcb.2005.861864","volume":"36","author":"A Oikonomopoulos","year":"2005","unstructured":"Oikonomopoulos, A., Patras, I., Pantic, M.: Spatiotemporal salient points for visual recognition of human actions. IEEE Trans. Syst. Man Cybern. Part B (Cybernetics) 36(3), 710\u2013719 (2005). https:\/\/doi.org\/10.1109\/tsmcb.2005.861864","journal-title":"IEEE Trans. Syst. Man Cybern. Part B (Cybernetics)"},{"key":"26_CR17","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1016\/j.engappai.2018.08.014","volume":"77","author":"C Dhiman","year":"2019","unstructured":"Dhiman, C., Vishwakarma, D.K.: A review of state-of-the-art techniques for abnormal human activity recognition. Eng. Appl. Artif. Intell. 77, 21\u201345 (2019). https:\/\/doi.org\/10.1016\/j.engappai.2018.08.014","journal-title":"Eng. Appl. Artif. Intell."},{"key":"26_CR18","doi-asserted-by":"publisher","unstructured":"Achanta, R., Hemami, S., Estrada, F., Susstrunk, S.: Frequency-tuned salient region detection. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2009, pp. 1597\u20131604. IEEE, June 2009. https:\/\/doi.org\/10.1109\/cvpr.2009.5206596","DOI":"10.1109\/cvpr.2009.5206596"},{"key":"26_CR19","doi-asserted-by":"publisher","unstructured":"Mele, M.L., Millar, D., Rijnders, C.E.: The web-based subjective quality assessment of an adaptive image compression plug-in. In: 1st International Conference on Human Computer Interaction Theory and Applications, HUCAPP, Porto, Portugal (2017). https:\/\/doi.org\/10.5220\/0006226401330137","DOI":"10.5220\/0006226401330137"},{"key":"26_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"353","DOI":"10.1007\/978-3-319-58071-5_27","volume-title":"Human-Computer Interaction. User Interface Design, Development and Multimodality","author":"ML Mele","year":"2017","unstructured":"Mele, M.L., Millar, D., Rijnders, C.E.: Using spatio-temporal saliency to predict subjective video quality: a new high-speed objective assessment metric. In: Kurosu, M. (ed.) HCI 2017. LNCS, vol. 10271, pp. 353\u2013368. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-58071-5_27"},{"key":"26_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"552","DOI":"10.1007\/978-3-319-91244-8_43","volume-title":"Human-Computer Interaction. Interaction in Context","author":"ML Mele","year":"2018","unstructured":"Mele, M.L., Millar, D., Rijnders, C.E.: Sencogi spatio-temporal saliency: a new metric for predicting subjective video quality on mobile devices. In: Kurosu, M. (ed.) HCI 2018. LNCS, vol. 10902, pp. 552\u2013564. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-91244-8_43"},{"issue":"2","key":"26_CR22","doi-asserted-by":"publisher","first-page":"422","DOI":"10.1177\/000306517001800210","volume":"18","author":"Max Schur","year":"1970","unstructured":"Schneirla, T.C.: An evolutionary and developmental theory of biphasic processes underlying approach and withdrawal (1959). https:\/\/doi.org\/10.1177\/000306517001800210","journal-title":"Journal of the American Psychoanalytic Association"},{"key":"26_CR23","doi-asserted-by":"publisher","unstructured":"Ekman, P., Friesen, W.V.: Manual for the Facial Action Coding System. Consulting Psychologists Press (1978). https:\/\/doi.org\/10.4135\/9781483381411","DOI":"10.4135\/9781483381411"},{"key":"26_CR24","doi-asserted-by":"publisher","unstructured":"Redi, J., Liu, H., Zunino, R., Heynderickx, I.: Interactions of visual attention and quality perception. In: Human Vision and Electronic Imaging XVI (2011). https:\/\/doi.org\/10.1117\/12.876712","DOI":"10.1117\/12.876712"},{"key":"26_CR25","unstructured":"Rijnders, C.E.: U.S. Patent Application No. 15\/899,331 (2018)"},{"issue":"3","key":"26_CR26","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1109\/TPAMI.2018.2815601","volume":"41","author":"Zoya Bylinskii","year":"2019","unstructured":"Bylinskii, Z., Judd, T., Oliva, A., Torralba, A., Durand, F.: What do different evaluation metrics tell us about saliency models? IEEE Trans. Pattern Anal. Mach. Intell. (2018). https:\/\/doi.org\/10.1109\/tpami.2018.2815601","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"26_CR27","doi-asserted-by":"publisher","first-page":"370","DOI":"10.1080\/00029238.1961.11080571","volume":"10","author":"HH Jasper","year":"1958","unstructured":"Jasper, H.H.: The ten-twenty electrode system of the International Federation. Electroencephalogr. Clin. Neurophysiol. 10, 370\u2013375 (1958). https:\/\/doi.org\/10.1080\/00029238.1961.11080571","journal-title":"Electroencephalogr. Clin. Neurophysiol."},{"issue":"2","key":"26_CR28","doi-asserted-by":"publisher","first-page":"148","DOI":"10.1177\/002076407902500222","volume":"25","author":"Peter Hall","year":"1979","unstructured":"John, E.R.: Neurometrics: clinical applications of quantitative electrophysiology, vol. 2. Wiley (1977). https:\/\/doi.org\/10.1177\/002076407902500222","journal-title":"International Journal of Social Psychiatry"},{"key":"26_CR29","unstructured":"Ekman, P.: Facial action coding system (FACS). A human face (2002)"},{"issue":"1","key":"26_CR30","doi-asserted-by":"publisher","first-page":"251","DOI":"10.3758\/s13428-012-0226-9","volume":"45","author":"O Le Meur","year":"2013","unstructured":"Le Meur, O., Baccino, T.: Methods for comparing scanpaths and saliency maps: strengths and weaknesses. Behav. Res. Methods 45(1), 251\u2013266 (2013). https:\/\/doi.org\/10.3758\/s13428-012-0226-9","journal-title":"Behav. Res. Methods"},{"key":"26_CR31","unstructured":"Judd, T., Durand, F., Torralba, A.: A benchmark of computational models of saliency to predict human fixations (2012)"},{"key":"26_CR32","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"809","DOI":"10.1007\/978-3-319-46454-1_49","volume-title":"Computer Vision \u2013 ECCV 2016","author":"Z Bylinskii","year":"2016","unstructured":"Bylinskii, Z., Recasens, A., Borji, A., Oliva, A., Torralba, A., Durand, F.: Where should saliency models look next? In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 809\u2013824. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46454-1_49"},{"key":"26_CR33","doi-asserted-by":"publisher","unstructured":"Judd, T., Ehinger, K., Durand, F., Torralba, A.: Learning to predict where humans look. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 2106\u20132113. IEEE, September 2009. https:\/\/doi.org\/10.1109\/iccv.2009.5459462","DOI":"10.1109\/iccv.2009.5459462"},{"key":"26_CR34","unstructured":"BT.500: Methodology for the subjective assessment of the quality of television pictures (n.d.). http:\/\/www.itu.int\/rec\/R-REC-BT.500-7-199510-S\/en. Accessed 9 Oct 2017"},{"key":"26_CR35","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4899-5379-7","volume-title":"Eye Movements and Vision","author":"Alfred L. Yarbus","year":"1967","unstructured":"Yarbus, A.L.: Eye Movements and Vision, New York (1967). https:\/\/doi.org\/10.1007\/978-1-4899-5379-7"},{"issue":"2","key":"26_CR36","doi-asserted-by":"publisher","first-page":"152","DOI":"10.1145\/123078.128728","volume":"9","author":"RJ Jacob","year":"1991","unstructured":"Jacob, R.J.: The use of eye movements in human-computer interaction techniques: what you look at is what you get. ACM Trans. Inf. Syst. (TOIS) 9(2), 152\u2013169 (1991). https:\/\/doi.org\/10.1145\/123078.128728","journal-title":"ACM Trans. Inf. Syst. (TOIS)"},{"issue":"8","key":"26_CR37","doi-asserted-by":"publisher","first-page":"e105106","DOI":"10.1371\/journal.pone.0105106","volume":"9","author":"ML Mele","year":"2014","unstructured":"Mele, M.L., Federici, S., Dennis, J.L.: Believing is seeing: fixation duration predicts implicit negative attitudes. PLoS ONE 9(8), e105106 (2014). https:\/\/doi.org\/10.1371\/journal.pone.0105106","journal-title":"PLoS ONE"}],"container-title":["Lecture Notes in Computer Science","Human-Computer Interaction. Recognition and Interaction Technologies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-22643-5_26","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,19]],"date-time":"2023-07-19T00:33:13Z","timestamp":1689726793000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-22643-5_26"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9783030226428","9783030226435"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-22643-5_26","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"27 June 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"HCII","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Human-Computer Interaction","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Orlando, FL","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 July 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31 July 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"hcii2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2019.hci.international\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"This content has been made available to all.","name":"free","label":"Free to read"}]}}