{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:17:11Z","timestamp":1750220231473,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":28,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T00:00:00Z","timestamp":1638316800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National key research and development program of China","award":["2019QY(Y)0202"],"award-info":[{"award-number":["2019QY(Y)0202"]}]},{"name":"Research Programme on Applied Fundamentals and Frontier Technologies of Wuhan","award":["2020010601012182"],"award-info":[{"award-number":["2020010601012182"]}]},{"name":"Natural Science Foundation of China","award":["61972169"],"award-info":[{"award-number":["61972169"]}]},{"DOI":"10.13039\/501100005090","name":"Beijing Nova Program","doi-asserted-by":"publisher","award":["Z201100006820123"],"award-info":[{"award-number":["Z201100006820123"]}],"id":[{"id":"10.13039\/501100005090","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,12]]},"DOI":"10.1145\/3469877.3490597","type":"proceedings-article","created":{"date-parts":[[2022,1,10]],"date-time":"2022-01-10T18:27:21Z","timestamp":1641839241000},"page":"1-6","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Video Saliency Prediction via Deep Eye Movement Learning"],"prefix":"10.1145","author":[{"given":"Jiazhong","family":"Chen","sequence":"first","affiliation":[{"name":"Huazhong University of Science and Technology, CN"}]},{"given":"Jie","family":"Chen","sequence":"additional","affiliation":[{"name":"Jinan University, CN"}]},{"given":"Yuan","family":"Dong","sequence":"additional","affiliation":[{"name":"Huazhong University of Science and Technology, CN"}]},{"given":"Dakai","family":"Ren","sequence":"additional","affiliation":[{"name":"Beijing University of Posts and Telecommunications, CN"}]},{"given":"Shiqi","family":"Zhang","sequence":"additional","affiliation":[{"name":"Huazhong University of Science and Technology, CN"}]},{"given":"Zongyi","family":"Li","sequence":"additional","affiliation":[{"name":"Huazhong University of Science and Technology, CN"}]}],"member":"320","published-online":{"date-parts":[[2022,1,10]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2017.2777665"},{"key":"e_1_3_2_1_2_1","unstructured":"L. Bazzani H. Larochelle and L. Torresani. 2017. Recurrent mixture density network for spatiotemporal visual attention. In ICLR. 1\u201315.  L. Bazzani H. Larochelle and L. Torresani. 2017. Recurrent mixture density network for spatiotemporal visual attention. In ICLR. 1\u201315."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"S. Chaabouni J. Benois-Pineau and C.\u00a0B. Amar. 2016. Transfer learning with deep networks for saliency prediction in natural video. In ICIP. 1604\u20131608.  S. Chaabouni J. Benois-Pineau and C.\u00a0B. Amar. 2016. Transfer learning with deep networks for saliency prediction in natural video. In ICIP. 1604\u20131608.","DOI":"10.1109\/ICIP.2016.7532629"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"crossref","unstructured":"S. Gorji and J. Clark. 2018. Going from image to video saliency: Augmenting image salience with dynamic attentional push. In CVPR. 7501\u20137511.  S. Gorji and J. Clark. 2018. Going from image to video saliency: Augmenting image salience with dynamic attentional push. In CVPR. 7501\u20137511.","DOI":"10.1109\/CVPR.2018.00783"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1126\/science.181.4095.179"},{"key":"e_1_3_2_1_6_1","unstructured":"L. Jiang M. Xu and Z. Wang. 2017. Predicting video saliency with object-to-motion CNN and two-layer convolutional LSTM. arXiv preprint arXiv:1709.06316(2017).  L. Jiang M. Xu and Z. Wang. 2017. Predicting video saliency with object-to-motion CNN and two-layer convolutional LSTM. arXiv preprint arXiv:1709.06316(2017)."},{"key":"e_1_3_2_1_7_1","volume-title":"SALICON: Saliency in context. In CVPR. 1072\u20131080.","author":"Jiang M.","year":"2015","unstructured":"M. Jiang , S. Huang , J. Duan , and Q. Zhao . 2015 . SALICON: Saliency in context. In CVPR. 1072\u20131080. M. Jiang, S. Huang, J. Duan, and Q. Zhao. 2015. SALICON: Saliency in context. In CVPR. 1072\u20131080."},{"key":"e_1_3_2_1_8_1","unstructured":"W. Kay J. Carreira K. Simonyan B. Zhang C. Hillier S. Vijayanarasimhan F. Viola T. Green T. Back and P. Natsev. 2017. The kinetics human action video dataset. arXiv preprint arXiv:1705.06950(2017).  W. Kay J. Carreira K. Simonyan B. Zhang C. Hillier S. Vijayanarasimhan F. Viola T. Green T. Back and P. Natsev. 2017. The kinetics human action video dataset. arXiv preprint arXiv:1705.06950(2017)."},{"key":"e_1_3_2_1_9_1","volume-title":"Adam: A method for stochastic optimization. In ICLR. 1\u201315.","author":"Kingma D.","year":"2015","unstructured":"D. Kingma and J. Ba . 2015 . Adam: A method for stochastic optimization. In ICLR. 1\u201315. D. Kingma and J. Ba. 2015. Adam: A method for stochastic optimization. In ICLR. 1\u201315."},{"key":"e_1_3_2_1_10_1","first-page":"12410","article-title":"Relation of cortical areas MT and MST to pursuit eye movements. I. Localization and visual properties of neurons","volume":"60","author":"Komatsu H.","year":"1988","unstructured":"H. Komatsu . 1988 . Relation of cortical areas MT and MST to pursuit eye movements. I. Localization and visual properties of neurons . Journal of Neurophysiology 60 (1988), 12410 \u2013 12417 . H. Komatsu. 1988. Relation of cortical areas MT and MST to pursuit eye movements. I. Localization and visual properties of neurons. Journal of Neurophysiology 60 (1988), 12410\u201312417.","journal-title":"Journal of Neurophysiology"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF00961876"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2019.2936112"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"G. Leifman D. Rudoy T. Swedish E. Bayro-Corrochano and R. Raskar. 2017. Learning gaze transitions from depth to improve video saliency estimation. In ICCV. 1707\u20131716.  G. Leifman D. Rudoy T. Swedish E. Bayro-Corrochano and R. Raskar. 2017. Learning gaze transitions from depth to improve video saliency estimation. In ICCV. 1707\u20131716.","DOI":"10.1109\/ICCV.2017.188"},{"key":"e_1_3_2_1_14_1","unstructured":"P. Linardos E. Mohedano J.\u00a0J. Nieto N.\u00a0E. O\u2019Connor X. Giro-i-Nieto and K. McGuinness. 2019. Simple vs complex temporal recurrences for video saliency prediction. arXiv preprint arXiv:1907.01869(2019).  P. Linardos E. Mohedano J.\u00a0J. Nieto N.\u00a0E. O\u2019Connor X. Giro-i-Nieto and K. McGuinness. 2019. Simple vs complex temporal recurrences for video saliency prediction. arXiv preprint arXiv:1907.01869(2019)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.19-06-02224.1999"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2014.2366154"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"K. Min and J. Corso. 2019. TASED-Net: Temporally aggregating spatial encoder-decoder network for video saliency detection. In ICCV. 2394\u20132403.  K. Min and J. Corso. 2019. TASED-Net: Temporally aggregating spatial encoder-decoder network for video saliency detection. In ICCV. 2394\u20132403.","DOI":"10.1109\/ICCV.2019.00248"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"crossref","unstructured":"N. Riche M. Duvinage M. Mancas B. Gosselin and T. Dutoit. 2014. Saliency and human fixations: State-of-the-art and study of comparison metrics. In ICCV. 1153\u20131160.  N. Riche M. Duvinage M. Mancas B. Gosselin and T. Dutoit. 2014. Saliency and human fixations: State-of-the-art and study of comparison metrics. In ICCV. 1153\u20131160.","DOI":"10.1109\/ICCV.2013.147"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF00199548"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1113\/jphysiol.1965.sp007718"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF00363977"},{"key":"e_1_3_2_1_22_1","unstructured":"X. Shi Z. Chen H. Wang D.-Y. Yeung and W.-K. Wong. 2015. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In NIPS. 802\u2013810.  X. Shi Z. Chen H. Wang D.-Y. Yeung and W.-K. Wong. 2015. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In NIPS. 802\u2013810."},{"key":"e_1_3_2_1_23_1","unstructured":"K. Simonyan and A. Zisserman. 2015. Very deep convolutional networks for large-scale image recognition. In ICLR. 1\u201314.  K. Simonyan and A. Zisserman. 2015. Very deep convolutional networks for large-scale image recognition. In ICLR. 1\u201314."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2018.2832053"},{"key":"e_1_3_2_1_25_1","first-page":"213","article-title":"A Model of the Smooth Pursuit Eye Movement System","volume":"18","author":"Tomohiro S.","year":"2005","unstructured":"S. Tomohiro , T. Hiromitsu , S. Schaal , and M. Kawato . 2005 . A Model of the Smooth Pursuit Eye Movement System . Biological Cybernetics 18 (2005), 213 \u2013 224 . S. Tomohiro, T. Hiromitsu, S. Schaal, and M. Kawato. 2005. A Model of the Smooth Pursuit Eye Movement System. Biological Cybernetics 18 (2005), 213\u2013224.","journal-title":"Biological Cybernetics"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"W. Wang J. Shen F. Guo M. Cheng and A. Borji. 2018. Revisiting video saliency: A large-scale benchmark and a new model. In CVPR. 4894\u20134903.  W. Wang J. Shen F. Guo M. Cheng and A. Borji. 2018. Revisiting video saliency: A large-scale benchmark and a new model. In CVPR. 4894\u20134903.","DOI":"10.1109\/CVPR.2018.00514"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"X. Wu Z. Wu J. Zhang L. Ju and S. Wang. 2020. SalSAC: A video saliency prediction model with shuffled attentions and correlation-based ConvLSTM. In AAAI. 12410\u201312417.  X. Wu Z. Wu J. Zhang L. Ju and S. Wang. 2020. SalSAC: A video saliency prediction model with shuffled attentions and correlation-based ConvLSTM. In AAAI. 12410\u201312417.","DOI":"10.1609\/aaai.v34i07.6927"},{"volume-title":"A revised stochastic sampled data model for eye tracking movements","author":"Young R.","key":"e_1_3_2_1_28_1","unstructured":"L.\u00a0 R. Young , J.\u00a0 D. Forster , and N. van Houtte . 1968. A revised stochastic sampled data model for eye tracking movements . In Fourth Ann NASA\u2013University Conference on Manual Control. 12410\u201312417. L.\u00a0R. Young, J.\u00a0D. Forster, and N. van Houtte. 1968. A revised stochastic sampled data model for eye tracking movements. In Fourth Ann NASA\u2013University Conference on Manual Control. 12410\u201312417."}],"event":{"name":"MMAsia '21: ACM Multimedia Asia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Gold Coast Australia","acronym":"MMAsia '21"},"container-title":["ACM Multimedia Asia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3469877.3490597","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3469877.3490597","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:30:16Z","timestamp":1750188616000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3469877.3490597"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12]]},"references-count":28,"alternative-id":["10.1145\/3469877.3490597","10.1145\/3469877"],"URL":"https:\/\/doi.org\/10.1145\/3469877.3490597","relation":{},"subject":[],"published":{"date-parts":[[2021,12]]},"assertion":[{"value":"2022-01-10","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}