{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,8]],"date-time":"2025-10-08T16:33:09Z","timestamp":1759941189068,"version":"3.37.3"},"reference-count":69,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2023,9,20]],"date-time":"2023-09-20T00:00:00Z","timestamp":1695168000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,9,20]],"date-time":"2023-09-20T00:00:00Z","timestamp":1695168000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100013076","name":"National Major Science and Technology Projects of China","doi-asserted-by":"publisher","award":["2020YFA0713504"],"award-info":[{"award-number":["2020YFA0713504"]}],"id":[{"id":"10.13039\/501100013076","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61972333"],"award-info":[{"award-number":["61972333"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014472","name":"Scientific Research Foundation of Hunan Provincial Education Department","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100014472","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-16593-2","type":"journal-article","created":{"date-parts":[[2023,9,20]],"date-time":"2023-09-20T13:01:56Z","timestamp":1695214916000},"page":"32301-32320","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Fusion hierarchy motion feature for video saliency detection"],"prefix":"10.1007","volume":"83","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7511-9418","authenticated-orcid":false,"given":"Fen","family":"Xiao","sequence":"first","affiliation":[]},{"given":"Huiyu","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Wenlei","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Zhen","family":"Li","sequence":"additional","affiliation":[]},{"given":"Xieping","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,20]]},"reference":[{"issue":"11","key":"16593_CR1","doi-asserted-by":"publisher","first-page":"7413","DOI":"10.1007\/s11042-019-08535-8","volume":"79","author":"X Wang","year":"2020","unstructured":"Wang X, Qi C (2020) Detecting action-relevant regions for action recognition using a three-stage saliency detection technique. Multimed Tools Appl 79(11):7413\u20137433","journal-title":"Multimed Tools Appl"},{"issue":"12","key":"16593_CR2","doi-asserted-by":"publisher","first-page":"17457","DOI":"10.1007\/s11042-022-12442-w","volume":"81","author":"K Cizmeciler","year":"2022","unstructured":"Cizmeciler K, Erdem E, Erdem A (2022) Leveraging semantic saliency maps for query-specific video summarization. Multimed Tools Appl 81(12):17457\u201317482","journal-title":"Multimed Tools Appl"},{"issue":"6","key":"16593_CR3","doi-asserted-by":"publisher","first-page":"7429","DOI":"10.1007\/s11042-017-4655-4","volume":"77","author":"J Ullah","year":"2018","unstructured":"Ullah J, Khan A, Jaffar MA (2018) Motion cues and saliency based unconstrained video segmentation. Multimed Tools Appl 77(6):7429\u20137446","journal-title":"Multimed Tools Appl"},{"issue":"11","key":"16593_CR4","doi-asserted-by":"publisher","first-page":"2409","DOI":"10.1109\/TCSVT.2016.2589878","volume":"27","author":"S Li","year":"2016","unstructured":"Li S, Xu M, Wang Z, Sun X (2016) Optimal bit allocation for ctu level rate control in hevc. IEEE Trans Circ Syst Video Technol 27(11):2409\u20132424","journal-title":"IEEE Trans Circ Syst Video Technol"},{"issue":"9","key":"16593_CR5","doi-asserted-by":"publisher","first-page":"4529","DOI":"10.1109\/TIP.2018.2837106","volume":"27","author":"M Xu","year":"2018","unstructured":"Xu M, Liu Y, Hu R, He F (2018) Find who to look at: turning from action to saliency. IEEE Trans Image Proc 27(9):4529\u20134544","journal-title":"IEEE Trans Image Proc"},{"issue":"7","key":"16593_CR6","doi-asserted-by":"publisher","first-page":"3156","DOI":"10.1109\/TIP.2017.2670143","volume":"26","author":"C Chen","year":"2017","unstructured":"Chen C, Li S, Wang Y, Qin H, Hao A (2017) Video saliency detection via spatial-temporal fusion and low-rank coherency diffusion. IEEE Trans Image Proc 26(7):3156\u20133170","journal-title":"IEEE Trans Image Proc"},{"issue":"12","key":"16593_CR7","doi-asserted-by":"publisher","first-page":"3324","DOI":"10.1109\/TMM.2018.2839523","volume":"20","author":"C Chen","year":"2018","unstructured":"Chen C, Li S, Qin H, Pan Z, Yang G (2018) Bilevel feature learning for video saliency detection. IEEE Trans Multimed 20(12):3324\u20133336","journal-title":"IEEE Trans Multimed"},{"issue":"5","key":"16593_CR8","doi-asserted-by":"publisher","first-page":"1153","DOI":"10.1109\/TMM.2019.2940851","volume":"22","author":"Y Li","year":"2019","unstructured":"Li Y, Li S, Chen C, Hao A, Qin H (2019) Accurate and robust video saliency detection via self-paced diffusion. IEEE Trans Multimed 22(5):1153\u20131167","journal-title":"IEEE Trans Multimed"},{"key":"16593_CR9","doi-asserted-by":"publisher","first-page":"1090","DOI":"10.1109\/TIP.2019.2934350","volume":"29","author":"C Chen","year":"2019","unstructured":"Chen C, Wang G, Peng C, Zhang X, Qin H (2019) Improved robust video saliency detection based on long-term spatial-temporal information. IEEE Trans Image Proc 29:1090\u20131100","journal-title":"IEEE Trans Image Proc"},{"key":"16593_CR10","doi-asserted-by":"publisher","first-page":"256","DOI":"10.1016\/j.neucom.2019.10.024","volume":"377","author":"P Zhang","year":"2020","unstructured":"Zhang P, Liu J, Wang X, Pu T, Fei C, Guo Z (2020) Stereoscopic video saliency detection based on spatiotemporal correlation and depth confidence optimization. Neurocomputing 377:256\u2013268","journal-title":"Neurocomputing"},{"key":"16593_CR11","unstructured":"Wang G, Chen C, Fan D, Hao A, Qin H (2021) Weakly supervised visual-auditory saliency detection with multigranularity perception. arXiv preprint arXiv:2112.13697"},{"key":"16593_CR12","doi-asserted-by":"crossref","unstructured":"Li H, Chen G, Li G, Yu Y (2019) Motion guided attention for video salient object detection. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 7274\u20137283","DOI":"10.1109\/ICCV.2019.00737"},{"issue":"5","key":"16593_CR13","doi-asserted-by":"publisher","first-page":"2732","DOI":"10.1109\/TCSVT.2021.3095843","volume":"32","author":"C Chen","year":"2021","unstructured":"Chen C, Song J, Peng C, Wang G, Fang Y (2021) A novel video salient object detection method via semisupervised motion quality perception. IEEE Trans Circ Syst Video Technol 32(5):2732\u20132745","journal-title":"IEEE Trans Circ Syst Video Technol"},{"issue":"11","key":"16593_CR14","doi-asserted-by":"publisher","first-page":"7662","DOI":"10.1109\/TCSVT.2022.3185252","volume":"32","author":"C Chen","year":"2022","unstructured":"Chen C, Wang H, Fang Y, Peng C (2022) A novel long-term iterative mining scheme for video salient object detection. IEEE Trans Circ Syst Video Technol 32(11):7662\u20137676","journal-title":"IEEE Trans Circ Syst Video Technol"},{"issue":"12","key":"16593_CR15","doi-asserted-by":"publisher","first-page":"5706","DOI":"10.1109\/TIP.2015.2487833","volume":"24","author":"A Borji","year":"2015","unstructured":"Borji A, Cheng M-M, Jiang H, Li J (2015) Salient object detection: a benchmark. IEEE Trans Image Proc 24(12):5706\u20135722","journal-title":"IEEE Trans Image Proc"},{"key":"16593_CR16","doi-asserted-by":"crossref","unstructured":"Vig E, Dorr M, Cox D (2014) Large-scale optimization of hierarchical features for saliency prediction in natural images. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2798\u20132805","DOI":"10.1109\/CVPR.2014.358"},{"key":"16593_CR17","unstructured":"Liu N, Han J, Zhang D, Wen S, Liu T (2015) Predicting eye fixations using convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 362\u2013370"},{"key":"16593_CR18","doi-asserted-by":"crossref","unstructured":"Huang X, Shen C, Boix X, Zhao Q (2015) Salicon: reducing the semantic gap in saliency prediction by adapting deep neural networks. In: Proceedings of the IEEE international conference on computer vision, pp 262\u2013270","DOI":"10.1109\/ICCV.2015.38"},{"key":"16593_CR19","doi-asserted-by":"crossref","unstructured":"Pan J, Sayrol E, Giro-i-Nieto X, McGuinness K, O\u2019Connor NE (2016) Shallow and deep convolutional networks for saliency prediction. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 598\u2013606","DOI":"10.1109\/CVPR.2016.71"},{"issue":"2","key":"16593_CR20","doi-asserted-by":"publisher","first-page":"392","DOI":"10.1109\/TNNLS.2016.2628878","volume":"29","author":"N Liu","year":"2016","unstructured":"Liu N, Han J, Liu T, Li X (2016) Learning to predict eye fixations via multiresolution convolutional neural networks. IEEE Trans Neur Netw Learn Syst 29(2):392\u2013404","journal-title":"IEEE Trans Neur Netw Learn Syst"},{"issue":"9","key":"16593_CR21","doi-asserted-by":"publisher","first-page":"4446","DOI":"10.1109\/TIP.2017.2710620","volume":"26","author":"SS Kruthiventi","year":"2017","unstructured":"Kruthiventi SS, Ayush K, Babu RV (2017) Deepfix: a fully convolutional neural network for predicting human eye fixations. IEEE Trans Image Proc 26(9):4446\u20134456","journal-title":"IEEE Trans Image Proc"},{"issue":"7","key":"16593_CR22","doi-asserted-by":"publisher","first-page":"3264","DOI":"10.1109\/TIP.2018.2817047","volume":"27","author":"N Liu","year":"2018","unstructured":"Liu N, Han J (2018) A deep spatial contextual long-term recurrent convolutional network for saliency detection. IEEE Trans Image Proc 27(7):3264\u20133274","journal-title":"IEEE Trans Image Proc"},{"issue":"7","key":"16593_CR23","doi-asserted-by":"publisher","first-page":"1408","DOI":"10.1109\/TPAMI.2014.2366154","volume":"37","author":"S Mathe","year":"2014","unstructured":"Mathe S, Sminchisescu C (2014) Actions in the eye: dynamic gaze datasets and learnt saliency models for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(7):1408\u20131424","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"11","key":"16593_CR24","doi-asserted-by":"publisher","first-page":"1254","DOI":"10.1109\/34.730558","volume":"20","author":"L Itti","year":"1998","unstructured":"Itti L, Koch C, Niebur E (1998) A model of saliency-based visual attention for rapid scene analysis. IEEE Trans Pattern Anal Mach Intell 20(11):1254\u20131259","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"16593_CR25","doi-asserted-by":"crossref","unstructured":"Gibson JJ (1950) The perception of the visual world","DOI":"10.2307\/1418003"},{"key":"16593_CR26","doi-asserted-by":"crossref","unstructured":"Teed Z, Deng J (2020) Raft: recurrent all-pairs field transforms for optical flow. In: European conference on computer vision. Springer, pp 402\u2013419","DOI":"10.1007\/978-3-030-58536-5_24"},{"key":"16593_CR27","doi-asserted-by":"crossref","unstructured":"Cong R, Song W, Lei J, Yue G, Zhao Y, Kwong S (2022) Psnet: parallel symmetric network for video salient object detection. IEEE Trans Emerg Top Comput Intell","DOI":"10.1109\/TETCI.2022.3220250"},{"issue":"7","key":"16593_CR28","doi-asserted-by":"publisher","first-page":"1688","DOI":"10.1109\/TMM.2017.2777665","volume":"20","author":"C Bak","year":"2017","unstructured":"Bak C, Kocak A, Erdem E, Erdem A (2017) Spatio-temporal saliency networks for dynamic saliency prediction. IEEE Trans Multimed 20(7):1688\u20131698","journal-title":"IEEE Trans Multimed"},{"key":"16593_CR29","doi-asserted-by":"crossref","unstructured":"Jiang L, Xu M, Liu T, Qiao M, Wang Z (2018) Deepvs: a deep learning based video saliency prediction approach. In: Proceedings of the European conference on computer vision (eccv), pp 602\u2013617","DOI":"10.1007\/978-3-030-01264-9_37"},{"issue":"12","key":"16593_CR30","doi-asserted-by":"publisher","first-page":"3544","DOI":"10.1109\/TCSVT.2018.2883305","volume":"29","author":"K Zhang","year":"2018","unstructured":"Zhang K, Chen Z (2018) Video saliency prediction based on spatial-temporal two-stream network. IEEE Trans Circ Syst Video Technol 29(12):3544\u20133557","journal-title":"IEEE Trans Circ Syst Video Technol"},{"key":"16593_CR31","doi-asserted-by":"publisher","first-page":"1113","DOI":"10.1109\/TIP.2019.2936112","volume":"29","author":"Q Lai","year":"2019","unstructured":"Lai Q, Wang W, Sun H, Shen J (2019) Video saliency prediction using spatiotemporal residual attentive networks. IEEE Trans Image Proc 29:1113\u20131126","journal-title":"IEEE Trans Image Proc"},{"key":"16593_CR32","unstructured":"Ballas N, Yao L, Pal C, Courville A (2015) Delving deeper into convolutional networks for learning video representations. arXiv preprint arXiv:1511.06432"},{"issue":"5","key":"16593_CR33","doi-asserted-by":"publisher","first-page":"1773","DOI":"10.1007\/s11554-021-01122-x","volume":"18","author":"PN Srinivasu","year":"2021","unstructured":"Srinivasu PN, Bhoi AK, Jhaveri RH, Reddy GT, Bilal M (2021) Probabilistic deep q network for real-time path planning in censorious robotic procedures using force sensors. J Real-Time Image Proc 18(5):1773\u20131785","journal-title":"J Real-Time Image Proc"},{"key":"16593_CR34","doi-asserted-by":"crossref","unstructured":"Craye C, Filliat D, Goudou J-F (2016) Environment exploration for object-based visual saliency learning. In: 2016 IEEE international conference on robotics and automation (ICRA), pp 2303\u20132309. IEEE","DOI":"10.1109\/ICRA.2016.7487379"},{"issue":"5","key":"16593_CR35","doi-asserted-by":"publisher","first-page":"802","DOI":"10.1109\/TPAMI.2006.86","volume":"28","author":"O Le Meur","year":"2006","unstructured":"Le Meur O, Le Callet P, Barba D, Thoreau D (2006) A coherent computational approach to model bottom-up visual attention. IEEE Trans Pattern Ana Mach Intell 28(5):802\u2013817","journal-title":"IEEE Trans Pattern Ana Mach Intell"},{"issue":"7","key":"16593_CR36","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1167\/8.7.32","volume":"8","author":"L Zhang","year":"2008","unstructured":"Zhang L, Tong MH, Marks TK, Shan H, Cottrell GW (2008) Sun: a bayesian framework for saliency using natural statistics. J Vision 8(7):32\u201332","journal-title":"J Vision"},{"key":"16593_CR37","unstructured":"Gao D, Vasconcelos N (2005) Discriminant saliency for visual recognition from cluttered scenes. In: Adv Neural Inf Proc Syst, pp 481\u2013488"},{"key":"16593_CR38","first-page":"155","volume":"18","author":"N Bruce","year":"2005","unstructured":"Bruce N, Tsotsos J (2005) Saliency based on information maximization. Adv Neural Inf Proc Syst 18:155\u2013162","journal-title":"Adv Neural Inf Proc Syst"},{"issue":"3","key":"16593_CR39","doi-asserted-by":"publisher","first-page":"569","DOI":"10.1109\/TPAMI.2014.2345401","volume":"37","author":"M-M Cheng","year":"2014","unstructured":"Cheng M-M, Mitra NJ, Huang X, Torr PH, Hu S-M (2014) Global contrast based salient region detection. IEEE Trans Pattern Anal Machi Intell 37(3):569\u2013582","journal-title":"IEEE Trans Pattern Anal Machi Intell"},{"key":"16593_CR40","doi-asserted-by":"crossref","unstructured":"Xu M, Ren Y, Wang Z (2015) Learning to predict saliency on face images. In: Proceedings of the IEEE international conference on computer vision, pp 3907\u20133915","DOI":"10.1109\/ICCV.2015.445"},{"issue":"3","key":"16593_CR41","doi-asserted-by":"publisher","first-page":"273","DOI":"10.1007\/BF00994018","volume":"20","author":"C Cortes","year":"1995","unstructured":"Cortes C, Vapnik V (1995) Support-vector networks. Mach Learn 20(3):273\u2013297","journal-title":"Mach Learn"},{"key":"16593_CR42","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556"},{"issue":"10","key":"16593_CR43","doi-asserted-by":"publisher","first-page":"5142","DOI":"10.1109\/TIP.2018.2851672","volume":"27","author":"M Cornia","year":"2018","unstructured":"Cornia M, Baraldi L, Serra G, Cucchiara R (2018) Predicting human eye fixations via an lstm-based saliency attentive model. IEEE Trans Image Processing 27(10):5142\u20135154","journal-title":"IEEE Trans Image Processing"},{"key":"16593_CR44","unstructured":"Shi X, Chen Z, Wang H, Yeung D-Y, Wong W-K, Woo W-c (2015) Convolutional lstm network: a machine learning approach for precipitation nowcasting. Adv Neural Inf Proc Syst 28"},{"issue":"5","key":"16593_CR45","doi-asserted-by":"publisher","first-page":"2368","DOI":"10.1109\/TIP.2017.2787612","volume":"27","author":"W Wang","year":"2017","unstructured":"Wang W, Shen J (2017) Deep visual attention prediction. IEEE Trans Image Proc 27(5):2368\u20132378","journal-title":"IEEE Trans Image Proc"},{"key":"16593_CR46","unstructured":"Guo C, Ma Q, Zhang L (2008) Spatio-temporal saliency detection using phase spectrum of quaternion fourier transform. In: 2008 IEEE conference on computer vision and pattern recognition, pp 1\u20138 IEEE"},{"key":"16593_CR47","doi-asserted-by":"crossref","unstructured":"Itti L, Dhavale N, Pighin F (2003) Realistic avatar eye and head animation using a neurobiological model of visual attention. In: Applications and science of neural networks, fuzzy systems, and evolutionary computation VI, vol 5200, pp 64\u201378. SPIE","DOI":"10.1117\/12.512618"},{"key":"16593_CR48","doi-asserted-by":"publisher","unstructured":"Wang W, Shen J, Xie J, Cheng M-M, Ling H, Borji A (2019) Revisiting video saliency prediction in the deep learning era. IEEE Trans Pattern Anal Mach Intell 1\u20131. https:\/\/doi.org\/10.1109\/TPAMI.2019.2924417","DOI":"10.1109\/TPAMI.2019.2924417"},{"key":"16593_CR49","doi-asserted-by":"crossref","unstructured":"Zhu S, Chang Q, L, Q (2022) Video saliency aware intelligent hd video compression with the improvement of visual quality and the reduction of coding complexity. Neural Computing and Applications 1\u201320","DOI":"10.1007\/s00521-022-06895-1"},{"key":"16593_CR50","doi-asserted-by":"publisher","first-page":"3995","DOI":"10.1109\/TIP.2021.3068644","volume":"30","author":"C Chen","year":"2021","unstructured":"Chen C, Wang G, Peng C, Fang Y, Zhang D, Qin H (2021) Exploring rich and efficient spatial temporal interactions for real-time video salient object detection. IEEE Trans Image Proc 30:3995\u20134007","journal-title":"IEEE Trans Image Proc"},{"key":"16593_CR51","doi-asserted-by":"crossref","unstructured":"Zhang F, Woodford OJ, Prisacariu VA, Torr PH (2021) Separable flow: Learning motion cost volumes for optical flow estimation. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 10807\u201310817","DOI":"10.1109\/ICCV48922.2021.01063"},{"key":"16593_CR52","doi-asserted-by":"crossref","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779\u2013788","DOI":"10.1109\/CVPR.2016.91"},{"key":"16593_CR53","doi-asserted-by":"crossref","unstructured":"Dosovitskiy A, Fischer P, Ilg E, Hausser P, Hazirbas C, Golkov V, Van Der\u00a0Smagt P, Cremers D, Brox T (2015) Flownet: learning optical flow with convolutional networks. In: Proceedings of the IEEE international conference on computer vision, pp 2758\u20132766","DOI":"10.1109\/ICCV.2015.316"},{"key":"16593_CR54","doi-asserted-by":"crossref","unstructured":"Mital P, mith TJ, Luke S, Henderson J (2013) Do low-level visual features have a causal influence on gaze during dynamic scene viewing? J Vision 13(9):144\u2013144","DOI":"10.1167\/13.9.144"},{"issue":"5","key":"16593_CR55","doi-asserted-by":"publisher","first-page":"427","DOI":"10.1111\/1467-9280.01458","volume":"14","author":"RA Abrams","year":"2003","unstructured":"Abrams RA (2003) Christ SE (2003) Motion onset captures attention. Psychol Sci 14(5):427\u2013432","journal-title":"Psychol Sci"},{"key":"16593_CR56","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"16593_CR57","doi-asserted-by":"crossref","unstructured":"Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8): 1735\u20131780","DOI":"10.1162\/neco.1997.9.8.1735"},{"issue":"1","key":"16593_CR58","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1007\/s12559-010-9074-z","volume":"3","author":"PK Mital","year":"2011","unstructured":"Mital PK, Smith TJ, Hill RL, Henderson JM (2011) Clustering of gaze during dynamic scene viewing is predicted by motion. Cogn Comput 3(1):5\u201324","journal-title":"Cogn Comput"},{"key":"16593_CR59","doi-asserted-by":"crossref","unstructured":"Judd T, Ehinger K, Durand F, Torralba A (2009) Learning to predict where humans look. In: 2009 IEEE 12th international conference on computer vision, pp 2106\u20132113. IEEE","DOI":"10.1109\/ICCV.2009.5459462"},{"key":"16593_CR60","doi-asserted-by":"crossref","unstructured":"Borji A, Tavakoli HR, Sihite DN, Itti L (2013) Analysis of scores, datasets, and models in visual saliency prediction. In: Proceedings of the IEEE international conference on computer vision, pp 921\u2013928","DOI":"10.1109\/ICCV.2013.118"},{"key":"16593_CR61","unstructured":"Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980"},{"key":"16593_CR62","unstructured":"Linardos P, Mohedano E, Nieto JJ, O\u2019Connor NE, Giro-i-Nieto X, McGuinness K (2019) Simple vs complex temporal recurrences for video saliency prediction. arXiv preprint arXiv:1907.01869"},{"key":"16593_CR63","doi-asserted-by":"crossref","unstructured":"Min K, Corso JJ (2019) Tased-net: temporally-aggregating spatial encoder-decoder network for video saliency detection. In: Proceedings of the IEEE international conference on computer vision, pp 2394\u20132403","DOI":"10.1109\/ICCV.2019.00248"},{"key":"16593_CR64","unstructured":"Wang Z, Zhou Z, Lu H, Hu Q, Jiang J (2020) Video saliency prediction via joint discrimination and local consistency. IEEE Transactions on Cybernetics"},{"key":"16593_CR65","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107275","volume":"103","author":"Z Wang","year":"2020","unstructured":"Wang Z, Zhou Z, Lu H, Jiang J (2020) Global and local sensitivity guided key salient object re-augmentation for video saliency detection. Pattern Recogn 103:107275","journal-title":"Pattern Recogn"},{"key":"16593_CR66","doi-asserted-by":"crossref","unstructured":"Jiang L, Xu M, Zhang S, Sigal L (2020) Deepct: a novel deep complex-valued network with learnable transform for video saliency prediction. Pattern Recogn 102:107234","DOI":"10.1016\/j.patcog.2020.107234"},{"key":"16593_CR67","doi-asserted-by":"publisher","first-page":"78","DOI":"10.1016\/j.patrec.2021.04.010","volume":"147","author":"W Zou","year":"2021","unstructured":"Zou W, Zhuo S, Tang Y, Tian S, Li X, Xu C (2021) Sta3d: spatiotemporally attentive 3d network for video saliency prediction. Pattern Recognition Letters 147:78\u201384","journal-title":"Pattern Recognition Letters"},{"key":"16593_CR68","doi-asserted-by":"publisher","first-page":"233","DOI":"10.1016\/j.neucom.2021.10.024","volume":"468","author":"H Xue","year":"2022","unstructured":"Xue H, Sun M, Liang Y (2022) Ecanet: explicit cyclic attention-based network for video saliency prediction. Neurocomput 468:233\u2013244","journal-title":"Neurocomput"},{"key":"16593_CR69","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1016\/j.neucom.2021.07.088","volume":"462","author":"J Chen","year":"2021","unstructured":"Chen J, Li Z, Jin Y, Ren D, Ling H (2021) Video saliency prediction via spatio-temporal reasoning. Neurocomput 462:59\u201368","journal-title":"Neurocomput"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16593-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-16593-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16593-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,8]],"date-time":"2024-03-08T06:53:59Z","timestamp":1709880839000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-16593-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,20]]},"references-count":69,"journal-issue":{"issue":"11","published-online":{"date-parts":[[2024,3]]}},"alternative-id":["16593"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-16593-2","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2023,9,20]]},"assertion":[{"value":"11 July 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 June 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 August 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 September 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there are no conflicts of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}]}}