{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,21]],"date-time":"2025-10-21T15:31:03Z","timestamp":1761060663792,"version":"3.37.3"},"reference-count":76,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"8","license":[{"start":{"date-parts":[[2019,8,1]],"date-time":"2019-08-01T00:00:00Z","timestamp":1564617600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,8,1]],"date-time":"2019-08-01T00:00:00Z","timestamp":1564617600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,8,1]],"date-time":"2019-08-01T00:00:00Z","timestamp":1564617600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Reverse Engineering Visual Intelligence for cognitiVe Enhancement (REVIVE)","award":["1335H00098"],"award-info":[{"award-number":["1335H00098"]}]},{"DOI":"10.13039\/501100001348","name":"A*STAR","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100001348","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001352","name":"National University of Singapore","doi-asserted-by":"publisher","award":["R-263-000-C08-133"],"award-info":[{"award-number":["R-263-000-C08-133"]}],"id":[{"id":"10.13039\/501100001352","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Ministry of Education of Singapore AcRF Tier One","award":["R-263-000-C21-112"],"award-info":[{"award-number":["R-263-000-C21-112"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2019,8,1]]},"DOI":"10.1109\/tpami.2018.2871688","type":"journal-article","created":{"date-parts":[[2018,9,24]],"date-time":"2018-09-24T21:42:59Z","timestamp":1537825379000},"page":"1783-1796","source":"Crossref","is-referenced-by-count":16,"title":["Anticipating Where People will Look Using Adversarial Networks"],"prefix":"10.1109","volume":"41","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2694-7097","authenticated-orcid":false,"given":"Mengmi","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Keng Teck","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Joo Hwee","family":"Lim","sequence":"additional","affiliation":[]},{"given":"Qi","family":"Zhao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6843-0064","authenticated-orcid":false,"given":"Jiashi","family":"Feng","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref73","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298625"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2013.350"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2015.05.006"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1167\/9.7.4"},{"key":"ref76","doi-asserted-by":"publisher","DOI":"10.1109\/VSPETS.2005.1570899"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.510"},{"key":"ref39","first-page":"1486","article-title":"Deep generative image models using a Laplacian pyramid of adversarial networks","author":"denton","year":"2015","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-005-1838-7"},{"article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","year":"2015","author":"radford","key":"ref38"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299189"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2012.6247710"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2010.69"},{"key":"ref30","first-page":"362","article-title":"Predicting eye fixations using convolutional neural networks","author":"liu","year":"2015","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"key":"ref37","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc Int Conf Neural Inf Process"},{"article-title":"Recurrent mixture density network for spatiotemporal visual attention","year":"2016","author":"bazzani","key":"ref36"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2013.152"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2567391"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2014.2366154"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.147"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.118"},{"article-title":"What do different evaluation metrics tell us about saliency models?","year":"2016","author":"bylinskii","key":"ref63"},{"article-title":"Deep gaze I: Boosting saliency prediction with feature maps trained on ImageNet","year":"2014","author":"k\u00fcmmerer","key":"ref28"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1145\/1143844.1143874"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.358"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2011.11.007"},{"key":"ref66","first-page":"155","article-title":"Saliency based on information maximization","author":"bruce","year":"2005","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref29","first-page":"1097","article-title":"ImageNet classification with deep convolutional neural networks","author":"krizhevsky","year":"2012","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1016\/S0042-6989(99)00163-7"},{"key":"ref68","doi-asserted-by":"crossref","first-page":"194","DOI":"10.1109\/TPAMI.2011.146","article-title":"Image signature: Highlighting sparse salient regions","volume":"34","author":"hou","year":"2012","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206697"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-25367-6_25"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2015.2409731"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2012.6247706"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1023\/A:1009715923555"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1167\/14.1.28"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1167\/11.3.9"},{"key":"ref23","first-page":"1871","article-title":"LIBLINEAR: A library for large linear classification","volume":"9","author":"fan","year":"2008","journal-title":"J Mach Learn Res"},{"key":"ref26","article-title":"Saliency detection within a deep convolutional architecture","author":"lin","year":"0","journal-title":"Proc Workshops 28th AAAI Conf Artif Intell"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.38"},{"key":"ref50","first-page":"835","article-title":"An uncertain future: Forecasting from static images using variational autoencoders","author":"walker","year":"2016","journal-title":"Proc Eur Conf Comput Vis"},{"article-title":"Deep predictive coding networks for video prediction and unsupervised learning","year":"2016","author":"lotter","key":"ref51"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206557"},{"key":"ref58","first-page":"314","article-title":"Learning to recognize daily actions using gaze","author":"fathi","year":"2012","journal-title":"Proc Eur Conf Comput Vis"},{"article-title":"Adam: A method for stochastic optimization","year":"2014","author":"kingma","key":"ref57"},{"key":"ref56","first-page":"613","article-title":"Generating videos with scene dynamics","author":"vondrick","year":"2016","journal-title":"Adv Neural Inf Process Syst"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553469"},{"key":"ref54","first-page":"527","article-title":"Shuffle and learn: Unsupervised learning using temporal order verification","author":"misra","year":"2016","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.320"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2013.6639343"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.3389\/fpsyg.2015.01049"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/0010-0285(80)90005-5"},{"key":"ref40","first-page":"318","article-title":"Generative image modeling using style and structure adversarial networks","author":"wang","year":"2016","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1007\/978-94-009-3833-5_5"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/34.730558"},{"key":"ref14","first-page":"545","article-title":"Graph-based visual saliency","author":"harel","year":"2006","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1167\/8.7.32"},{"key":"ref16","doi-asserted-by":"crossref","first-page":"194","DOI":"10.1109\/TPAMI.2011.146","article-title":"Image signature: Highlighting sparse salient regions","volume":"34","author":"hou","year":"2012","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.26"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1167\/9.3.5"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2009.5459462"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.377"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.399"},{"article-title":"Gaze-enhanced user interface design","year":"2007","author":"kumar","key":"ref6"},{"key":"ref5","first-page":"8055","article-title":"A gaze-controlled interface to virtual reality applications for motor- and speech-impaired users","volume":"1","author":"ding","year":"2009","journal-title":"HCI Int"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1002\/(SICI)1099-1778(199901\/03)10:1<39::AID-VIS195>3.3.CO;2-U"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1080\/15252019.2011.10722185"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46484-8_16"},{"article-title":"Look-that-there: Exploiting gaze in virtual reality interactions","year":"2005","author":"zeleznik","key":"ref9"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.18"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.416"},{"key":"ref48","first-page":"91","article-title":"Visual dynamics: Probabilistic future frame synthesis via cross convolutional networks","author":"xue","year":"2016","journal-title":"Proc Int Conf Neural Inf Process"},{"article-title":"Video pixel networks","year":"2016","author":"kalchbrenner","key":"ref47"},{"article-title":"Conditional generative adversarial nets","year":"2014","author":"mirza","key":"ref42"},{"article-title":"Deep multi-scale video prediction beyond mean square error","year":"2015","author":"mathieu","key":"ref41"},{"article-title":"Video (language) modeling: A baseline for generative models of natural videos","year":"2014","author":"ranzato","key":"ref44"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2015.2430335"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/34\/8752185\/08471119.pdf?arnumber=8471119","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,13]],"date-time":"2022-07-13T20:48:54Z","timestamp":1657745334000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8471119\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,8,1]]},"references-count":76,"journal-issue":{"issue":"8"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2018.2871688","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"type":"print","value":"0162-8828"},{"type":"electronic","value":"2160-9292"},{"type":"electronic","value":"1939-3539"}],"subject":[],"published":{"date-parts":[[2019,8,1]]}}}