{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T17:48:32Z","timestamp":1772905712758,"version":"3.50.1"},"reference-count":51,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"10","license":[{"start":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T00:00:00Z","timestamp":1569888000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T00:00:00Z","timestamp":1569888000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,10,1]],"date-time":"2019-10-01T00:00:00Z","timestamp":1569888000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61472389"],"award-info":[{"award-number":["61472389"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61650202"],"award-info":[{"award-number":["61650202"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61332016"],"award-info":[{"award-number":["61332016"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61620106009"],"award-info":[{"award-number":["61620106009"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U1636214"],"award-info":[{"award-number":["U1636214"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Key Research Program of Frontier Sciences","award":["CAS: QYZDJ-SSW-SYS013"],"award-info":[{"award-number":["CAS: QYZDJ-SSW-SYS013"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Circuits Syst. Video Technol."],"published-print":{"date-parts":[[2019,10]]},"DOI":"10.1109\/tcsvt.2018.2870954","type":"journal-article","created":{"date-parts":[[2018,9,27]],"date-time":"2018-09-27T18:58:26Z","timestamp":1538074706000},"page":"2960-2971","source":"Crossref","is-referenced-by-count":26,"title":["Learning Coupled Convolutional Networks Fusion for Video Saliency Prediction"],"prefix":"10.1109","volume":"29","author":[{"given":"Zhe","family":"Wu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4038-753X","authenticated-orcid":false,"given":"Li","family":"Su","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7542-296X","authenticated-orcid":false,"given":"Qingming","family":"Huang","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2015.2473844"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref33","first-page":"568","article-title":"Two-stream convolutional networks for action recognition in videos","author":"simonyan","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref32","first-page":"3174","article-title":"Fixation bank: Learning to reweight fixation candidates","author":"zhao","year":"2015","journal-title":"Proc IEEE Conf Comput Vis Pattern Recognit"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2014.2336549"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/1631272.1631370"},{"key":"ref37","author":"yu","year":"2015","journal-title":"Multi-scale context aggregation by dilated convolutions"},{"key":"ref36","author":"simonyan","year":"2014","journal-title":"Very Deep Convolutional Networks for Large-scale Image Recognition"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref34","author":"wang","year":"2015","journal-title":"Towards good practices for very deep two-stream convnets"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2016.7532629"},{"key":"ref27","author":"cornia","year":"2016","journal-title":"Predicting human eye fixations via an lstm-based saliency attentive model"},{"key":"ref29","author":"bazzani","year":"2016","journal-title":"Recurrent mixture density network for spatiotemporal visual attention"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/34.730558"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2012.89"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299189"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2013.152"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33709-3_60"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.71"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.38"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.620"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2016.7900174"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/EUSIPCO.2015.7362638"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298710"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2014.2308642"},{"key":"ref11","doi-asserted-by":"crossref","first-page":"185","DOI":"10.1109\/TIP.2009.2030969","article-title":"A novel multiresolution spatiotemporal saliency detection model and its applications in image and video compression","volume":"19","author":"guo","year":"2010","journal-title":"IEEE Trans Image Process"},{"key":"ref40","first-page":"214","article-title":"A duality based approach for realtime TV-\n${L}^{1}$\n optical flow","author":"zach","year":"2007","journal-title":"Proc Joint Pattern Recognit Symp (DAGM-OAGM)"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2013.2273613"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2567391"},{"key":"ref14","first-page":"545","article-title":"Graph-based visual saliency","author":"harel","year":"2006","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-02490-0_31"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2009.112"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2014.2332213"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2014.2308652"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2013.2277884"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-009-0215-3"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1117\/12.512618"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2013.6738712"},{"key":"ref5","first-page":"2944","article-title":"Sunday: Saliency using natural statistics for dynamic analysis of scenes","author":"zhang","year":"2009","journal-title":"Proc 31st Annu Conf Cogn Sci Soc"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/1180639.1180824"},{"key":"ref7","first-page":"547","article-title":"Bayesian surprise attracts human attention","author":"itti","year":"2005","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1167\/9.12.15"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2013.2270367"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2012.2210727"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2011.2165292"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-013-0678-0"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1016\/j.visres.2004.09.017"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1145\/2647868.2654889"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2012.120"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1007\/s12559-010-9074-z"},{"key":"ref43","author":"chen","year":"2016","journal-title":"Deeplab Semantic image segmentation with deep convolutional nets atrous convolution and fully connected crfs"}],"container-title":["IEEE Transactions on Circuits and Systems for Video Technology"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/76\/8855034\/08474977.pdf?arnumber=8474977","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,13]],"date-time":"2022-07-13T20:48:40Z","timestamp":1657745320000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8474977\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,10]]},"references-count":51,"journal-issue":{"issue":"10"},"URL":"https:\/\/doi.org\/10.1109\/tcsvt.2018.2870954","relation":{},"ISSN":["1051-8215","1558-2205"],"issn-type":[{"value":"1051-8215","type":"print"},{"value":"1558-2205","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,10]]}}}