{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,30]],"date-time":"2025-12-30T15:37:04Z","timestamp":1767109024640,"version":"3.37.3"},"reference-count":57,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"12","license":[{"start":{"date-parts":[[2019,12,1]],"date-time":"2019-12-01T00:00:00Z","timestamp":1575158400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,12,1]],"date-time":"2019-12-01T00:00:00Z","timestamp":1575158400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,12,1]],"date-time":"2019-12-01T00:00:00Z","timestamp":1575158400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"NSFC","doi-asserted-by":"crossref","award":["61502304"],"award-info":[{"award-number":["61502304"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100008982","name":"National Science Foundation","doi-asserted-by":"publisher","award":["HCC-1319598","IIS-1422477"],"award-info":[{"award-number":["HCC-1319598","IIS-1422477"]}],"id":[{"id":"10.13039\/501100008982","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2019,12,1]]},"DOI":"10.1109\/tpami.2018.2866563","type":"journal-article","created":{"date-parts":[[2018,8,23]],"date-time":"2018-08-23T19:27:14Z","timestamp":1535052434000},"page":"2975-2989","source":"Crossref","is-referenced-by-count":36,"title":["Personalized Saliency and Its Prediction"],"prefix":"10.1109","volume":"41","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8926-7833","authenticated-orcid":false,"given":"Yanyu","family":"Xu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1626-2040","authenticated-orcid":false,"given":"Shenghua","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Junru","family":"Wu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4172-4940","authenticated-orcid":false,"given":"Nianyi","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jingyi","family":"Yu","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","first-page":"64","article-title":"Unsupervised learning for physical interaction through video prediction","author":"finn","year":"2016","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref38","first-page":"667","article-title":"Dynamic filter networks","author":"de brabandere","year":"2016","journal-title":"Proc Int Conf Neural Inf Process"},{"journal-title":"Neuroimaging Personality Social Cognition and Character","year":"2016","author":"absher","key":"ref33"},{"journal-title":"Neuroscience of Preference and Choice Cognitive and Neural Mechanisms","year":"2011","author":"dolan","key":"ref32"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.512"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1016\/j.paid.2010.06.016"},{"key":"ref37","first-page":"30","article-title":"Image question answering using convolutional neural network with dynamic parameter prediction","author":"noh","year":"2015","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"journal-title":"MRC International Report","article-title":"Gender Specific Gaze Differences?","year":"2011","key":"ref36"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1111\/infa.12144"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/s10919-007-0043-5"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1007\/s10919-007-0043-5"},{"key":"ref27","first-page":"362","article-title":"Predicting eye fixations using convolutional neural networks","author":"liu","year":"2015","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"key":"ref29","first-page":"55","article-title":"The role of sexually-oriented stimuli in advertising: Theory and literature review","author":"wilson","year":"1979","journal-title":"ACR North American Advances"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1145\/1149488.1149499"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2004.834657"},{"key":"ref20","first-page":"562","article-title":"Deeply-supervised nets","author":"lee","year":"2015","journal-title":"Proc Int Conf Artif Intell Statist"},{"key":"ref22","first-page":"94","article-title":"Facial landmark detection by deep multi-task learning","author":"zhang","year":"2014","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2016.2579306"},{"article-title":"MIT saliency benchmark","year":"2015","author":"bylinskii","key":"ref24"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1145\/2647868.2654889"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.26"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2016.2522440"},{"key":"ref50","first-page":"2106","article-title":"Learning to predict where humans look","author":"judd","year":"2010","journal-title":"Proc IEEE Int Conf Comput Vis"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2017\/543"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.398"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.512"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.513"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.354"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0193149"},{"key":"ref52","article-title":"TurkerGaze: Crowdsourcing saliency with webcam based eye tracking","author":"xu","year":"2015","journal-title":"Comput Sci"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.71"},{"key":"ref11","first-page":"362","article-title":"Predicting eye fixations using convolutional neural networks","author":"liu","year":"2015","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"key":"ref40","first-page":"1097","article-title":"ImageNet classification with deep convolutional neural networks","author":"krizhevsky","year":"2012","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2710620"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.623"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1167\/14.1.28"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298710"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-007-0090-8"},{"article-title":"A benchmark of computational models of saliency to predict human fixations","year":"2012","author":"judd","key":"ref18"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2016.7900174"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.205"},{"key":"ref3","doi-asserted-by":"crossref","first-page":"453","DOI":"10.1007\/978-3-319-40621-3_33","article-title":"Automatic information positioning scheme in AR-assisted maintenance based on visual saliency","author":"chang","year":"2016","journal-title":"Int Conf on Augmented Reality Virtual Reality and Computer Graphics"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2009.5459462"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2015.2487833"},{"article-title":"TurkerGaze: Crowdsourcing saliency with webcam based eye tracking","year":"2015","author":"xu","key":"ref8"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.43"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2012.6247706"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.38"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/34.730558"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1167\/13.4.11"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2013.03.009"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.358"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"article-title":"Very deep convolutional networks for large-scale image recognition","year":"2014","author":"simonyan","key":"ref44"},{"article-title":"Deep gaze I: Boosting saliency prediction with feature maps trained on ImageNet","year":"2014","author":"kmmerer","key":"ref43"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/34\/8890757\/08444709.pdf?arnumber=8444709","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,13]],"date-time":"2022-07-13T20:48:55Z","timestamp":1657745335000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8444709\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,12,1]]},"references-count":57,"journal-issue":{"issue":"12"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2018.2866563","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"type":"print","value":"0162-8828"},{"type":"electronic","value":"2160-9292"},{"type":"electronic","value":"1939-3539"}],"subject":[],"published":{"date-parts":[[2019,12,1]]}}}