{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,20]],"date-time":"2025-12-20T22:25:03Z","timestamp":1766269503940},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"18","license":[{"start":{"date-parts":[[2021,5,21]],"date-time":"2021-05-21T00:00:00Z","timestamp":1621555200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,5,21]],"date-time":"2021-05-21T00:00:00Z","timestamp":1621555200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2021,7]]},"DOI":"10.1007\/s11042-021-10862-8","type":"journal-article","created":{"date-parts":[[2021,5,21]],"date-time":"2021-05-21T03:47:25Z","timestamp":1621568845000},"page":"27531-27552","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["No-reference omnidirectional video quality assessment based on generative adversarial networks"],"prefix":"10.1007","volume":"80","author":[{"given":"Jiefeng","family":"Guo","sequence":"first","affiliation":[]},{"given":"Yao","family":"Luo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,5,21]]},"reference":[{"key":"10862_CR1","unstructured":"Arjovsky M, Chintala S, Bottou L (2017) Wasserstein generative adversarial networks. In: Proc. int. conf. mach. learn. (ICML), vol 70, pp 214\u2013223, Sydney, Australia"},{"issue":"1","key":"10862_CR2","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1109\/TIP.2017.2760518","volume":"27","author":"S Bosse","year":"2017","unstructured":"Bosse S, Maniry D, M\u00fcller KR, Wiegand T, Samek W (2017) Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans. Image Process 27(1):206\u2013219","journal-title":"IEEE Trans. Image Process"},{"key":"10862_CR3","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li LJ, Li K, Li FF (2009) ImageNet: A large-scale hierarchical image database. In: Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp 248\u2013255, Miami, FL, USA","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"10862_CR4","doi-asserted-by":"crossref","unstructured":"Dong X, Shen J (2018) Triplet loss in siamese network for object tracking. In: Proc. Eur. Conf. Comput. Vis. (ECCV). pp 472\u2013488, Munich, Germany","DOI":"10.1007\/978-3-030-01261-8_28"},{"issue":"7","key":"10862_CR5","doi-asserted-by":"publisher","first-page":"3516","DOI":"10.1109\/TIP.2019.2898567","volume":"28","author":"X Dong","year":"2019","unstructured":"Dong X, Shen J, Wu D, Guo K, Jin X, Porikli F (2019) Quadruplet network with one-shot learning for fast visual object tracking. IEEE Trans. Image Process 28(7):3516\u20133527","journal-title":"IEEE Trans. Image Process"},{"key":"10862_CR6","unstructured":"Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In: Proc. Adv. Neural Inf. Process. Syst. (NIPS), vol 27, pp 2672\u20132680, Montr\u00e9al, Canada"},{"key":"10862_CR7","unstructured":"Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville A (2017) Improved training of wasserstein gans. In: Proc. Adv. Neural Inf. Process. Syst. (NIPS), vol 30, pp 5767\u20135777, Long Beach, CA, USA"},{"key":"10862_CR8","unstructured":"Huawei iLab (2019) Cloud VR service quality monitoring white paper. Huawei report, Huawei Technologies Co., Ltd."},{"key":"10862_CR9","doi-asserted-by":"crossref","unstructured":"Justin J, Alexandre A, Li FF (2016) Perceptual losses for real-time style transfer and super-resolution. In: Proc. Eur. Conf. Comput. Vis. (ECCV), 694\u2013711, Amsterdam, The Netherlands","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"10862_CR10","doi-asserted-by":"crossref","unstructured":"Kang L, Ye P, Li Y, Doermann D (2014) Convolutional neural networks for no-reference image quality assessment. In: Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp 1733\u20131740, Columbus, OH, United States","DOI":"10.1109\/CVPR.2014.224"},{"issue":"4","key":"10862_CR11","doi-asserted-by":"publisher","first-page":"917","DOI":"10.1109\/TCSVT.2019.2898732","volume":"30","author":"HG Kim","year":"2019","unstructured":"Kim HG, Lim HT, Ro YM (2019) Deep virtual reality image quality assessment with human perception guider for omnidirectional image. IEEE Trans. Circuits Syst. Video Technol. 30(4):917\u2013928","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10862_CR12","doi-asserted-by":"crossref","unstructured":"Kupyn O, Budzan V, Mykhailych M, Mishkin D, Matas J (2018) DeblurGAN: Blind motion deblurring using conditional adversarial networks. In: Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR). pp 8183\u20138192, Salt Lake City, UT, USA","DOI":"10.1109\/CVPR.2018.00854"},{"key":"10862_CR13","doi-asserted-by":"crossref","unstructured":"Ledig C, Theis L, Huszar F, Caballero J, Cunningham A, Acosta A, Aitken A, Tejani A, Totz J, Wang Z, Shi W (2017) Photo-realistic single image super-resolution using a generative adversarial network. In: Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), vol 27, pp 2672\u20132680, Honolulu, Hawaii, USA","DOI":"10.1109\/CVPR.2017.19"},{"key":"10862_CR14","doi-asserted-by":"crossref","unstructured":"Li C, Xu M, Du X, Wang Z (2018) Bridge the gap between VQA and human behavior on omnidirectional video. In: Proc. 26th ACM Int. Conf. Multimedia (MM \u201918), pp 932\u2013940, Seoul, Republic of Korea","DOI":"10.1145\/3240508.3240581"},{"key":"10862_CR15","doi-asserted-by":"publisher","first-page":"3351","DOI":"10.1109\/TIP.2019.2959256","volume":"29","author":"Z Liang","year":"2020","unstructured":"Liang Z, Shen J (2020) Local semantic siamese networks for fast tracking. IEEE Trans. Image Process 29:3351\u20133364","journal-title":"IEEE Trans. Image Process"},{"key":"10862_CR16","doi-asserted-by":"crossref","unstructured":"Lin KY, Wang G (2018) Hallucinated-IQA: No-reference image quality assessment via adversarial learning. In: Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp 732\u2013741, Salt Lake City, UT, USA","DOI":"10.1109\/CVPR.2018.00083"},{"issue":"10","key":"10862_CR17","doi-asserted-by":"publisher","first-page":"4818","DOI":"10.1109\/TIP.2017.2718185","volume":"26","author":"Z Ni","year":"2017","unstructured":"Ni Z, Ma L, Zeng H, Chen J, Cai C, Ma KK (2017) ESIM: Edge similarity for screen content image quality assessment. IEEE Trans. Image Process 26(10):4818\u20134831","journal-title":"IEEE Trans. Image Process"},{"key":"10862_CR18","doi-asserted-by":"crossref","unstructured":"Orduna M, P\u00e9rez P, D\u00edaz C, Garc\u00eda N (2020) Evaluating the influence of the HMD, usability, and fatigue in 360VR video quality assessments. In: Proceedings of IEEE Conference on Virtual Reality and 3D User Interfaces Workshops (VRW). pp 683\u2013684, Atlanta, GA, USA","DOI":"10.1109\/VRW50115.2020.00192"},{"key":"10862_CR19","doi-asserted-by":"publisher","first-page":"5612","DOI":"10.1109\/TIP.2020.2984879","volume":"29","author":"SV Reddy Dendi","year":"2020","unstructured":"Reddy Dendi SV, Channappayya SS (2020) No-reference video quality assessment using natural spatiotemporal scene statistics. IEEE Trans. Image Process 29:5612\u20135624","journal-title":"IEEE Trans. Image Process"},{"key":"10862_CR20","doi-asserted-by":"crossref","unstructured":"Ren H, Chen D, Wang Y (2018) RAN4IQA: Restorative adversarial nets for no-reference image quality assessment. In: Proc. AAAI Conf. Artif. Intell. (AAAI). New orleans, Louisiana, USA","DOI":"10.1609\/aaai.v32i1.12258"},{"key":"10862_CR21","unstructured":"Salimans T, Goodfellow I, Zaremba W, Cheung V, Radford A, Chen X (2016) Improved techniques for training GANs. In: Proc. Adv. Neural Inf. Process. Syst. (NIPS). pp 2234\u20132242, Barcelona, Spain"},{"issue":"6","key":"10862_CR22","doi-asserted-by":"publisher","first-page":"1427","DOI":"10.1109\/TIP.2010.2042111","volume":"19","author":"K Seshadrinathan","year":"2010","unstructured":"Seshadrinathan K, Soundararajan R, Conrad AB, Cormack LK (2010) Study of subjective and objective quality assessment of video. IEEE Trans. Image Process 19(6):1427\u20131441","journal-title":"IEEE Trans. Image Process"},{"issue":"7","key":"10862_CR23","doi-asserted-by":"publisher","first-page":"3068","DOI":"10.1109\/TCYB.2019.2936503","volume":"50","author":"J Shen","year":"2020","unstructured":"Shen J, Tang X, Dong X, Shao L (2020) Visual object tracking by hierarchical attention siamese network. IEEE Trans. Cybern. 50(7):3068\u20133080","journal-title":"IEEE Trans. Cybern."},{"issue":"1","key":"10862_CR24","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1109\/JSTSP.2019.2955024","volume":"14","author":"W Sun","year":"2020","unstructured":"Sun W, Min X, Zhai G, Gu K, Duan H, Ma S (2020) MC360IQA: A multi-channel CNN for blind 360-degree image quality assessment. IEEE J. Sel. Topics Signal Process 14(1):64\u201377","journal-title":"IEEE J. Sel. Topics Signal Process"},{"key":"10862_CR25","doi-asserted-by":"crossref","unstructured":"Tran D, Bourdev L, Fergus R, Torresani L, Paluri M (2015) Learning spatiotemporal features with 3D convolutional networks. In: Proc. IEEE Int. Conf. Comput. Vis. (ICCV), pp 4489\u20134497, Santiago, Chile","DOI":"10.1109\/ICCV.2015.510"},{"issue":"1","key":"10862_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TCSVT.2016.2602764","volume":"28","author":"S Wang","year":"2016","unstructured":"Wang S, Gu K, Zhang X, Lin W, Ma S, Gao W (2016) Reduced-reference quality assessment of screen content images. IEEE Trans. Circuits Syst. Video Technol. 28(1):1\u201314","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"5","key":"10862_CR27","doi-asserted-by":"publisher","first-page":"2368","DOI":"10.1109\/TIP.2017.2787612","volume":"27","author":"W Wang","year":"2018","unstructured":"Wang W, Shen J (2018) Deep visual attention prediction. IEEE Trans. Image Process 27(5):2368\u20132378","journal-title":"IEEE Trans. Image Process"},{"issue":"7","key":"10862_CR28","doi-asserted-by":"publisher","first-page":"1531","DOI":"10.1109\/TPAMI.2018.2840724","volume":"41","author":"W Wang","year":"2019","unstructured":"Wang W, Shen J, Ling H (2019) A deep network solution for attention and aesthetics aware photo cropping. IEEE Trans. Pattern Anal. Mach. Intell. 41(7):1531\u20131544","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"12","key":"10862_CR29","doi-asserted-by":"publisher","first-page":"3516","DOI":"10.1109\/TCSVT.2018.2886277","volume":"29","author":"M Xu","year":"2019","unstructured":"Xu M, Li C, Chen Z, Wang Z, Guan Z (2019) Assessing visual quality of omnidirectional videos. IEEE Trans. Circuits Syst. Video Technol. 29 (12):3516\u20133530","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10862_CR30","doi-asserted-by":"publisher","first-page":"38669","DOI":"10.1109\/ACCESS.2018.2854922","volume":"6","author":"J Yang","year":"2018","unstructured":"Yang J, Liu T, Jiang B, Song H, Lu W (2018) 3D panoramic virtual reality video quality assessment based on 3D convolutional neural networks. IEEE Access 6:38669\u201338682","journal-title":"IEEE Access"},{"key":"10862_CR31","doi-asserted-by":"publisher","first-page":"83","DOI":"10.1016\/j.neucom.2018.04.072","volume":"309","author":"J Yang","year":"2018","unstructured":"Yang J, Zhu Y, Ma C, Lu W, Meng Q (2018) Stereoscopic video quality assessment based on 3D convolutional neural networks. Neurocomputing 309:83\u201393","journal-title":"Neurocomputing"},{"key":"10862_CR32","doi-asserted-by":"crossref","unstructured":"Yu M, Lakshman H, Girod B (2015) A framework to evaluate omnidirectional video coding schemes. In: International Symposium on Mixed and Augmented Reality (ISMAR), pp 31\u201336, Fukuoka, Japan","DOI":"10.1109\/ISMAR.2015.12"},{"key":"10862_CR33","doi-asserted-by":"crossref","unstructured":"Zakharchenko V, Choi KP, Park JH (2016) Quality metric for spherical panoramic video. In: Optics and Photonics for Information Processing X. vol 9970, pp 57\u201365, San Diego, CA, United States","DOI":"10.1117\/12.2235885"},{"issue":"1","key":"10862_CR34","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1109\/TCSVT.2018.2886771","volume":"30","author":"W Zhang","year":"2018","unstructured":"Zhang W, Ma K, Yan J, Deng D, Wang Z (2018) Blind image quality assessment using a deep bilinear convolutional neural network. IEEE Trans. Circuits Syst. Video Technol. 30(1):36\u201347","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"8","key":"10862_CR35","doi-asserted-by":"publisher","first-page":"2716","DOI":"10.1109\/TNNLS.2018.2890310","volume":"31","author":"Y Zhang","year":"2019","unstructured":"Zhang Y, Gao X, He L, Lu W, He R (2019) Objective video quality assessment combining transfer learning with CNN. IEEE Trans. Neural Netw. Learn. Syst. 31(8):2716\u20132730","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-021-10862-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-021-10862-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-021-10862-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,28]],"date-time":"2022-12-28T05:27:00Z","timestamp":1672205220000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-021-10862-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,5,21]]},"references-count":35,"journal-issue":{"issue":"18","published-print":{"date-parts":[[2021,7]]}},"alternative-id":["10862"],"URL":"https:\/\/doi.org\/10.1007\/s11042-021-10862-8","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,5,21]]},"assertion":[{"value":"13 October 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 January 2021","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 March 2021","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 May 2021","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}