{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T03:45:57Z","timestamp":1775533557670,"version":"3.50.1"},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,7,6]],"date-time":"2024-07-06T00:00:00Z","timestamp":1720224000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,7,6]],"date-time":"2024-07-06T00:00:00Z","timestamp":1720224000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No. 61971347, No. 62272383"],"award-info":[{"award-number":["No. 61971347, No. 62272383"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s00530-024-01403-y","type":"journal-article","created":{"date-parts":[[2024,7,6]],"date-time":"2024-07-06T13:01:25Z","timestamp":1720270885000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["A multi-scale no-reference video quality assessment method based on transformer"],"prefix":"10.1007","volume":"30","author":[{"given":"Yingan","family":"Cui","sequence":"first","affiliation":[]},{"given":"Zonghua","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Yuqin","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Huaijun","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Junhuai","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,6]]},"reference":[{"key":"1403_CR1","first-page":"1","volume":"72","author":"AK Vishwakarma","year":"2023","unstructured":"Vishwakarma, A.K., Bhurchandi, K.M.: No-reference video quality assessment using local structural and quality-aware deep features. IEEE Trans. Instrum. Meas. 72, 1\u201312 (2023)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1403_CR2","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1109\/TMM.2021.3122347","volume":"25","author":"Y Li","year":"2023","unstructured":"Li, Y., Meng, S., Zhang, X., Wang, M., Wang, S., Wang, Y., Ma, S.: User-generated video quality assessment: a subjective and objective study. IEEE Trans. Multimed. 25, 154\u2013166 (2023)","journal-title":"IEEE Trans. Multimed."},{"key":"1403_CR3","doi-asserted-by":"crossref","unstructured":"Chen, Y., Li, S.: Multi-level feature-guided stereoscopic video quality assessment based on transformer and convolutional neural network. In: Proc. IEEE International Conference on Multimedia and Expo (ICME), pp. 2513\u20132518 (2023)","DOI":"10.1109\/ICME55011.2023.00428"},{"issue":"2","key":"1403_CR4","doi-asserted-by":"publisher","first-page":"89","DOI":"10.1007\/s00530-024-01285-0","volume":"30","author":"F Yang","year":"2024","unstructured":"Yang, F., Yang, C., An, P., Huang, X.: 360$$^\\circ$$ video quality assessment based on saliency-guided viewport extraction. Multimedia Syst. 30(2), 89 (2024)","journal-title":"Multimedia Syst."},{"issue":"5","key":"1403_CR5","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1007\/s00530-023-01129-3","volume":"29","author":"J Li","year":"2023","unstructured":"Li, J., Zhang, H., Ma, H.: Drl-based transmission control for QOE guaranteed transmission efficiency optimization in tile-based panoramic video streaming. Multimed. Syst. 29(5), 2761\u20132777 (2023)","journal-title":"Multimed. Syst."},{"key":"1403_CR6","doi-asserted-by":"publisher","first-page":"573","DOI":"10.1007\/s00530-021-00858-7","volume":"28","author":"H Zhang","year":"2022","unstructured":"Zhang, H., Guo, D., Liu, W., Nie, J., Li, S.: An improved algorithm of video quality assessment by danmaku analysis. Multimed. Syst. 28, 573\u2013582 (2022)","journal-title":"Multimed. Syst."},{"issue":"8","key":"1403_CR7","doi-asserted-by":"publisher","first-page":"2256","DOI":"10.1109\/TCSVT.2018.2868262","volume":"29","author":"CG Bampis","year":"2019","unstructured":"Bampis, C.G., Li, Z., Bovik, A.C.: Spatiotemporal feature integration and model fusion for full reference video quality assessment. IEEE Trans. Circuits Syst. Video Technol. 29(8), 2256\u20132270 (2019). https:\/\/doi.org\/10.1109\/TCSVT.2018.2868262","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1403_CR8","doi-asserted-by":"publisher","unstructured":"Xu, M., Chen, J., Wang, H., Liu, S., Li, G., Bai, Z.: C3DVQA: Full-reference video quality assessment with 3d convolutional neural network. In: Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4447\u20134451 (2020). https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9053031","DOI":"10.1109\/ICASSP40776.2020.9053031"},{"key":"1403_CR9","doi-asserted-by":"crossref","unstructured":"Fu, H., Pan, D., Shi, P.: Full-reference video quality assessment based on spatiotemporal visual sensitivity. In: Proc. International Conference on Culture-oriented Science & Technology (ICCST), pp. 305\u2013309 (2021)","DOI":"10.1109\/ICCST53801.2021.00071"},{"issue":"4","key":"1403_CR10","doi-asserted-by":"publisher","first-page":"684","DOI":"10.1109\/TCSVT.2012.2214933","volume":"23","author":"R Soundararajan","year":"2013","unstructured":"Soundararajan, R., Bovik, A.C.: Video quality assessment by reduced reference spatio-temporal entropic differencing. IEEE Trans. Circuits Syst. Video Technol. 23(4), 684\u2013694 (2013). https:\/\/doi.org\/10.1109\/TCSVT.2012.2214933","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1403_CR11","doi-asserted-by":"crossref","unstructured":"Wu, H., Chen, C., Hou, J., Liao, L., Wang, A., Sun, W., Yan, Q., Lin, W.: FAST-VQA: Efficient end-to-end video quality assessment with fragment sampling. In: Proc. European Conference on Computer Vision (ECCV), pp. 538\u2013554 (2022)","DOI":"10.1007\/978-3-031-20068-7_31"},{"issue":"5","key":"1403_CR12","doi-asserted-by":"publisher","first-page":"1316","DOI":"10.1109\/TNN.2006.879766","volume":"17","author":"P Le Callet","year":"2006","unstructured":"Le Callet, P., Viard-Gaudin, C., Barba, D.: A convolutional neural network approach for objective video quality assessment. IEEE. Trans. Neural. Netw. Learn. Syst. 17(5), 1316\u20131327 (2006)","journal-title":"IEEE. Trans. Neural. Netw. Learn. Syst."},{"key":"1403_CR13","doi-asserted-by":"crossref","unstructured":"Li, D., Jiang, T., Jiang, M.: Quality assessment of in-the-wild videos. In: Proc. ACM International Conference on Multimedia (ACM MM), pp. 2351\u20132359 (2019)","DOI":"10.1145\/3343031.3351028"},{"key":"1403_CR14","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1403_CR15","doi-asserted-by":"crossref","unstructured":"Cho, K., Merri\u00ebnboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., Bengio, Y.: Learning phrase representations using RNN encoder\u2013decoder for statistical machine translation. In: Proc. Empirical Methods in Natural Language Processing (EMNLP), pp. 1724\u20131734 (2014)","DOI":"10.3115\/v1\/D14-1179"},{"issue":"12","key":"1403_CR16","doi-asserted-by":"publisher","first-page":"4695","DOI":"10.1109\/TIP.2012.2214050","volume":"21","author":"A Mittal","year":"2012","unstructured":"Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 21(12), 4695\u20134708 (2012)","journal-title":"IEEE Trans. Image Process."},{"issue":"3","key":"1403_CR17","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/LSP.2012.2227726","volume":"20","author":"A Mittal","year":"2013","unstructured":"Mittal, A., Soundararajan, R., Bovik, A.C.: Making a \u201ccompletely blind\u2019\u2019 image quality analyzer. IEEE Signal Process. Lett. 20(3), 209\u2013212 (2013)","journal-title":"IEEE Signal Process. Lett."},{"issue":"3","key":"1403_CR18","doi-asserted-by":"publisher","first-page":"1352","DOI":"10.1109\/TIP.2014.2299154","volume":"23","author":"MA Saad","year":"2014","unstructured":"Saad, M.A., Bovik, A.C., Charrier, C.: Blind prediction of natural video quality. IEEE Trans. Image Process. 23(3), 1352\u20131365 (2014)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"1403_CR19","doi-asserted-by":"publisher","first-page":"289","DOI":"10.1109\/TIP.2015.2502725","volume":"25","author":"A Mittal","year":"2016","unstructured":"Mittal, A., Saad, M.A., Bovik, A.C.: A completely blind video integrity oracle. IEEE Trans. Image Process. 25(1), 289\u2013300 (2016)","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"1403_CR20","doi-asserted-by":"publisher","first-page":"5923","DOI":"10.1109\/TIP.2019.2923051","volume":"28","author":"J Korhonen","year":"2019","unstructured":"Korhonen, J.: Two-level approach for no-reference consumer video quality assessment. IEEE Trans. Image Process. 28(12), 5923\u20135938 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"1403_CR21","doi-asserted-by":"publisher","first-page":"4449","DOI":"10.1109\/TIP.2021.3072221","volume":"30","author":"Z Tu","year":"2021","unstructured":"Tu, Z., Wang, Y., Birkbeck, N., Adsumilli, B., Bovik, A.C.: UGC-VQA: benchmarking blind video quality assessment for user generated content. IEEE Trans. Image Process. 30, 4449\u20134464 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"1403_CR22","first-page":"425","volume":"2","author":"Z Tu","year":"2021","unstructured":"Tu, Z., Yu, X., Wang, Y., Birkbeck, N., Adsumilli, B., Bovik, A.C.: RAPIQUE: rapid and accurate video quality prediction of user generated content. IEEE Trans. Nucl. Sci. 2, 425\u2013440 (2021)","journal-title":"IEEE Trans. Nucl. Sci."},{"key":"1403_CR23","doi-asserted-by":"publisher","first-page":"1238","DOI":"10.1007\/s11263-020-01408-w","volume":"129","author":"D Li","year":"2021","unstructured":"Li, D., Jiang, T., Jiang, M.: Unified quality assessment of in-the-wild videos with mixed datasets training. Int. J. Comput. Vis. 129, 1238\u20131257 (2021)","journal-title":"Int. J. Comput. Vis."},{"issue":"6","key":"1403_CR24","doi-asserted-by":"publisher","first-page":"3500","DOI":"10.1109\/TCSVT.2021.3114509","volume":"32","author":"Y Liu","year":"2022","unstructured":"Liu, Y., Wu, J., Li, L., Dong, W., Zhang, J., Shi, G.: Spatiotemporal representation learning for blind video quality assessment. IEEE Trans. Circuits Syst. Video Technol. 32(6), 3500\u20133513 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"3","key":"1403_CR25","doi-asserted-by":"publisher","first-page":"651","DOI":"10.1109\/TBC.2022.3164332","volume":"68","author":"W Shen","year":"2022","unstructured":"Shen, W., Zhou, M., Liao, X., Jia, W., Xiang, T., Fang, B., Shang, Z.: An end-to-end no-reference video quality assessment method with hierarchical spatiotemporal feature representation. IEEE Trans. Broadcast. 68(3), 651\u2013660 (2022)","journal-title":"IEEE Trans. Broadcast."},{"key":"1403_CR26","doi-asserted-by":"crossref","unstructured":"Hara, K., Kataoka, H., Satoh, Y.: Can spatiotemporal 3d CNNS retrace the history of 2d CNNS and imagenet? In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6546\u20136555 (2018)","DOI":"10.1109\/CVPR.2018.00685"},{"issue":"8","key":"1403_CR27","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural. Comput. 9(8), 1735\u20131780 (1997)","journal-title":"Neural. Comput."},{"key":"1403_CR28","doi-asserted-by":"crossref","unstructured":"Zhang, K., Chao, W.-L., Sha, F., Grauman, K.: Video summarization with long short-term memory. In: Proc. European Conference on Computer Vision (ECCV), pp. 766\u2013782 (2016)","DOI":"10.1007\/978-3-319-46478-7_47"},{"key":"1403_CR29","doi-asserted-by":"crossref","unstructured":"Hosu, V., Hahn, F., Jenadeleh, M., Lin, H., Men, H., Szir\u00e1nyi, T., Li, S., Saupe, D.: The konstanz natural video database (KoNViD-1k). In: Proc. International Conference on Quality of Multimedia Experience (QoMEX), pp. 1\u20136 (2017)","DOI":"10.1109\/QoMEX.2017.7965673"},{"issue":"7","key":"1403_CR30","doi-asserted-by":"publisher","first-page":"3073","DOI":"10.1109\/TIP.2016.2562513","volume":"25","author":"M Nuutinen","year":"2016","unstructured":"Nuutinen, M., Virtanen, T., Vaahteranoksa, M., Vuori, T., Oittinen, P., H\u00e4kkinen, J.: CVD2014\u2014a database for evaluating no-reference video quality assessment algorithms. IEEE Trans. Image Process. 25(7), 3073\u20133086 (2016)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"1403_CR31","doi-asserted-by":"publisher","first-page":"612","DOI":"10.1109\/TIP.2018.2869673","volume":"28","author":"Z Sinno","year":"2019","unstructured":"Sinno, Z., Bovik, A.C.: Large-scale study of perceptual video quality. IEEE Trans. Image Process. 28(2), 612\u2013627 (2019)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"1403_CR32","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1145\/2812802","volume":"59","author":"B Thomee","year":"2016","unstructured":"Thomee, B., Shamma, D.A., Friedland, G., Elizalde, B., Ni, K., Poland, D., Borth, D., Li, L.-J.: YFCC100M: the new data in multimedia research. Commun. ACM 59(2), 64\u201373 (2016)","journal-title":"Commun. ACM"},{"key":"1403_CR33","doi-asserted-by":"crossref","unstructured":"Liao, L., Xu, K., Wu, H., Chen, C., Sun, W., Yan, Q., Lin, W.: Exploring the effectiveness of video perceptual representation in blind video quality assessment. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 837\u2013846 (2022)","DOI":"10.1145\/3503161.3547849"},{"key":"1403_CR34","first-page":"2555","volume":"37","author":"J Wang","year":"2023","unstructured":"Wang, J., Chan, K.C., Loy, C.C.: Exploring clip for assessing the look and feel of images. Proc. AAAI Conf. Artif. Intell. 37, 2555\u20132563 (2023)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"1403_CR35","doi-asserted-by":"crossref","unstructured":"Zhang, W., Zhai, G., Wei, Y., Yang, X., Ma, K.: Blind image quality assessment via vision-language correspondence: A multitask learning perspective. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14071\u201314081 (2023)","DOI":"10.1109\/CVPR52729.2023.01352"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01403-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01403-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01403-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,22]],"date-time":"2024-08-22T08:29:59Z","timestamp":1724315399000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01403-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,6]]},"references-count":35,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["1403"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01403-y","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,6]]},"assertion":[{"value":"8 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 June 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no confict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"201"}}