{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T20:54:36Z","timestamp":1768596876522,"version":"3.49.0"},"reference-count":35,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Engineering Applications of Artificial Intelligence"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1016\/j.engappai.2025.113491","type":"journal-article","created":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T13:15:02Z","timestamp":1765199702000},"page":"113491","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"PB","title":["No-reference video quality assessment based on spatio-temporal perception information"],"prefix":"10.1016","volume":"165","author":[{"given":"Zhiming","family":"Shi","sequence":"first","affiliation":[]},{"given":"Xinrong","family":"Wu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"10","key":"10.1016\/j.engappai.2025.113491_bib1","doi-asserted-by":"crossref","first-page":"15677","DOI":"10.1007\/s11042-020-10286-w","article-title":"PIQI: perceptual image quality index based on ensemble of Gaussian process regression","volume":"80","author":"Ahmed","year":"2021","journal-title":"Multimed. Tool. Appl."},{"issue":"103259","key":"10.1016\/j.engappai.2025.113491_bib2","first-page":"141","article-title":"Spatio-temporal attention feature fusion: a video quality assessment method for User-Generated Content","volume":"91","author":"Ai","year":"2025","journal-title":"Displays"},{"issue":"6","key":"10.1016\/j.engappai.2025.113491_bib3","doi-asserted-by":"crossref","DOI":"10.1117\/1.JEI.30.6.063001","article-title":"CNN-based no-reference video quality assessment method using a spatiotemporal saliency patch selection procedure","volume":"30","author":"Alamgeer","year":"2021","journal-title":"J. Electron. Imag."},{"key":"10.1016\/j.engappai.2025.113491_bib4","doi-asserted-by":"crossref","first-page":"1882","DOI":"10.1109\/TIP.2023.3251695","article-title":"Attention-Guided neural networks for full-reference and No-Reference audio-visual quality assessment","volume":"32","author":"Cao","year":"2023","journal-title":"IEEE Trans. Image Process."},{"issue":"4","key":"10.1016\/j.engappai.2025.113491_bib5","doi-asserted-by":"crossref","first-page":"1903","DOI":"10.1109\/TCSVT.2021.3088505","article-title":"Learning generalized spatial-temporal deep feature representation for No-Reference video quality assessment","volume":"32","author":"Chen","year":"2022","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"issue":"3","key":"10.1016\/j.engappai.2025.113491_bib6","doi-asserted-by":"crossref","first-page":"844","DOI":"10.1109\/TBC.2024.3399479","article-title":"No-Reference VMAF: a deep neural network-based approach to blind video quality assessment","volume":"70","author":"DeDecker","year":"2024","journal-title":"IEEE Trans. Broadcast."},{"key":"10.1016\/j.engappai.2025.113491_bib7","doi-asserted-by":"crossref","first-page":"8059","DOI":"10.1109\/TIP.2021.3112055","article-title":"No-Reference Video quality prediction via space-time chips","volume":"30","author":"Ebenezer","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.engappai.2025.113491_bib8","doi-asserted-by":"crossref","first-page":"2693","DOI":"10.1109\/TIP.2023.3272480","article-title":"Study of spatio-temporal modeling in video quality assessment","volume":"32","author":"Fang","year":"2023","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.engappai.2025.113491_bib9","series-title":"2021 8th International Conference on Future Internet of Things and Cloud (FiCloud)","first-page":"19","article-title":"Video streaming analysis in multi-tier edge-cloud networks","author":"Gama","year":"2021"},{"key":"10.1016\/j.engappai.2025.113491_bib10","series-title":"Data Traffic Monitoring and Analysis","first-page":"264","article-title":"Internet video delivery in YouTube: from traffic measurements to quality of experience","author":"Ho\u00dffeld","year":"2013"},{"issue":"9","key":"10.1016\/j.engappai.2025.113491_bib11","doi-asserted-by":"crossref","first-page":"5944","DOI":"10.1109\/TCSVT.2022.3164467","article-title":"Blindly assess quality of in-the-wild videos via quality-aware pre-training and motion perception","volume":"32","author":"Li","year":"2022","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"issue":"9","key":"10.1016\/j.engappai.2025.113491_bib12","doi-asserted-by":"crossref","first-page":"5944","DOI":"10.1109\/TCSVT.2022.3164467","article-title":"Blindly assess quality of in-the-wild videos via quality-aware pre-training and motion perception","volume":"32","author":"Li","year":"2022","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"key":"10.1016\/j.engappai.2025.113491_bib13","doi-asserted-by":"crossref","first-page":"2939","DOI":"10.1109\/TMM.2022.3153208","article-title":"Optimal volumetric video streaming with hybrid saliency based tiling","volume":"25","author":"Li","year":"2023","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.engappai.2025.113491_bib14","series-title":"Proc. ACM SIGCOMM Conf. Appl., Technol., Architectures, Protocols Comput. Commun.","first-page":"359","article-title":"A case for a coordinated internet video control plane","author":"Liu","year":"2012"},{"key":"10.1016\/j.engappai.2025.113491_bib15","doi-asserted-by":"crossref","first-page":"6623","DOI":"10.1109\/TIP.2021.3096060","article-title":"Reduced reference perceptual quality model with application to rate control for video-based point cloud compression","volume":"30","author":"Liu","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.engappai.2025.113491_bib16","doi-asserted-by":"crossref","first-page":"5138","DOI":"10.1109\/TIP.2023.3310344","article-title":"Contrastive Video quality estimator","volume":"32","author":"Madhusudana","year":"2023","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.engappai.2025.113491_bib17","series-title":"Proc. 3rd Multimedia Syst","first-page":"11","article-title":"QDASH: a QoE-aware DASH system","author":"Mok","year":"2012"},{"issue":"11","key":"10.1016\/j.engappai.2025.113491_bib18","doi-asserted-by":"crossref","first-page":"3440","DOI":"10.1109\/TIP.2006.881959","article-title":"A statistical evaluation of recent full reference image quality assessment algorithms","volume":"15","author":"Sheikh","year":"2006","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.engappai.2025.113491_bib19","article-title":"Synchronization of markovian jump neural networks for sampled data control systems with additive delay components: analysis of image encryption technique","author":"Tamil Thendral","year":"2022","journal-title":"Math. Methods Appl. Sci."},{"issue":"4","key":"10.1016\/j.engappai.2025.113491_bib20","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3632178","article-title":"2BiVQA: double Bi-LSTM-based video quality assessment of UGC videos","volume":"20","author":"Telili","year":"2022","journal-title":"ACM Trans. Multimed Comput. Commun. Appl"},{"key":"10.1016\/j.engappai.2025.113491_bib21","doi-asserted-by":"crossref","first-page":"4449","DOI":"10.1109\/TIP.2021.3072221","article-title":"UGC-VQA: benchmarking blind video quality assessment for user generated content","volume":"30","author":"Tu","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.engappai.2025.113491_bib22","doi-asserted-by":"crossref","first-page":"425","DOI":"10.1109\/OJSP.2021.3090333","article-title":"RAPIQUE: rapid and accurate video quality prediction of user generated content","volume":"2","author":"Tu","year":"2021","journal-title":"IEEE Open Journal of Signal Processing"},{"key":"10.1016\/j.engappai.2025.113491_bib23","doi-asserted-by":"crossref","first-page":"425","DOI":"10.1109\/OJSP.2021.3090333","article-title":"RAPIQUE: rapid and accurate video quality prediction of user generated content","volume":"2","author":"Tu","year":"2021","journal-title":"IEEE Open Journal of Signal Processing"},{"issue":"4","key":"10.1016\/j.engappai.2025.113491_bib24","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: from error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.engappai.2025.113491_bib25","article-title":"ReLaX-VQA: residual fragment and layer stack extraction for enhancing video quality assessment","author":"Wang","year":"2024","journal-title":"arxiv preprint arxiv:2407.11496"},{"key":"10.1016\/j.engappai.2025.113491_bib26","series-title":"European Conference on Computer Vision","article-title":"FAST-VQA: efficient end-to-end video quality assessment with Fragment Sampling[C]","author":"Wu","year":"2022"},{"issue":"9","key":"10.1016\/j.engappai.2025.113491_bib27","doi-asserted-by":"crossref","first-page":"4840","DOI":"10.1109\/TCSVT.2023.3249741","article-title":"DisCoVQA: temporal distortion-content transformers for video quality assessment","volume":"33","author":"Wu","year":"2023","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"issue":"9","key":"10.1016\/j.engappai.2025.113491_bib28","doi-asserted-by":"crossref","first-page":"4840","DOI":"10.1109\/TCSVT.2023.3249741","article-title":"DisCoVQA: temporal distortion-content transformers for video quality assessment","volume":"33","author":"Wu","year":"2023","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"key":"10.1016\/j.engappai.2025.113491_bib29","series-title":"Exploring Video Quality Assessment on User Generated Contents from Aesthetic and Technical Perspectives[C]\/IEEE\/CVFInternational Conference on Computer Vision","first-page":"20087","author":"Wu","year":"2023"},{"key":"10.1016\/j.engappai.2025.113491_bib30","series-title":"Proc. IEEE Int. Conf. Multimedia Expo. Workshops (ICMEW)","first-page":"1","article-title":"Assessing quality of experience for adaptive HTTP video streaming","author":"Xue","year":"2014"},{"key":"10.1016\/j.engappai.2025.113491_bib31","doi-asserted-by":"crossref","DOI":"10.1016\/j.displa.2023.102585","article-title":"Revisiting the robustness of spatio-temporal modeling in video quality assessment","volume":"81","author":"Yan","year":"2024","journal-title":"Displays"},{"key":"10.1016\/j.engappai.2025.113491_bib32","series-title":"Proc. ACM Conf. Special Interest Group Data Commun.","first-page":"325","article-title":"A control-theoretic approach for dynamic adaptive video streaming over HTTP","author":"Yin","year":"2015"},{"key":"10.1016\/j.engappai.2025.113491_bib33","series-title":"Patch-VQ: \u2018Patching Up\u2019 the Video Quality Problem[C]","first-page":"14014","author":"Ying","year":"2021"},{"key":"10.1016\/j.engappai.2025.113491_bib34","series-title":"HVS Revisited: a Comprehensive Video Quality Assessment Framework","author":"Zhang","year":"2022"},{"issue":"103232","key":"10.1016\/j.engappai.2025.113491_bib35","first-page":"141","article-title":"Dual-stage attention based symmetric framework for stereo video quality assessment","volume":"91","author":"Zhang","year":"2025","journal-title":"Displays"}],"container-title":["Engineering Applications of Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625035225?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0952197625035225?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T08:20:45Z","timestamp":1768551645000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0952197625035225"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2]]},"references-count":35,"alternative-id":["S0952197625035225"],"URL":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113491","relation":{},"ISSN":["0952-1976"],"issn-type":[{"value":"0952-1976","type":"print"}],"subject":[],"published":{"date-parts":[[2026,2]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"No-reference video quality assessment based on spatio-temporal perception information","name":"articletitle","label":"Article Title"},{"value":"Engineering Applications of Artificial Intelligence","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.engappai.2025.113491","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"113491"}}