{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T12:09:25Z","timestamp":1771502965443,"version":"3.50.1"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2023,6,22]],"date-time":"2023-06-22T00:00:00Z","timestamp":1687392000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,6,22]],"date-time":"2023-06-22T00:00:00Z","timestamp":1687392000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61972097"],"award-info":[{"award-number":["61972097"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003392","name":"Natural Science Foundation of Fujian Province","doi-asserted-by":"publisher","award":["2020J01494"],"award-info":[{"award-number":["2020J01494"]}],"id":[{"id":"10.13039\/501100003392","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2024,1]]},"DOI":"10.1007\/s11042-023-15739-6","type":"journal-article","created":{"date-parts":[[2023,6,22]],"date-time":"2023-06-22T10:14:52Z","timestamp":1687428892000},"page":"12139-12157","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Two-stream network with viewport selection for blind omnidirectional video quality assessment"],"prefix":"10.1007","volume":"83","author":[{"given":"Junhao","family":"Chen","sequence":"first","affiliation":[]},{"given":"Yuzhen","family":"Niu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,22]]},"reference":[{"key":"15739_CR1","unstructured":"Antkowiak J, Jamal Baina T, Baroncini FV, Chateau N, FranceTelecom F, Pessoa ACF, Stephanie Colonnese F, Contin IL, Caviedes J, Philips F (2000) Final report from the video quality experts group on the validation of objective models of video quality assessment March 2000"},{"issue":"1","key":"15739_CR2","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1109\/TIP.2017.2760518","volume":"27","author":"S Bosse","year":"2017","unstructured":"Bosse S, Maniry D, M\u00fcller K-R, Wiegand T, Samek W (2017) Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans Image Process 27(1):206\u2013219. https:\/\/doi.org\/10.1109\/TIP.2017.2760518","journal-title":"IEEE Trans Image Process"},{"key":"15739_CR3","unstructured":"Boyce J, Alshina E, Abbas A, Yan Y (2018) JVET-E1030: JVET common test conditions and evaluation procedures for 360$$^{\\circ }$$ video"},{"key":"15739_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijleo.2020.165887","volume":"226","author":"X Chai","year":"2021","unstructured":"Chai X, Shao F (2021) Blind quality assessment of omnidirectional videos using spatio-temporal convolutional neural networks. Optik 226:165887. https:\/\/doi.org\/10.1016\/j.ijleo.2020.165887","journal-title":"Optik"},{"key":"15739_CR5","doi-asserted-by":"crossref","unstructured":"Chen S, Zhang Y, Li Y, Chen Z, Wang Z (2018) Spherical structural similarity index for objective omnidirectional video quality assessment. In: 2018 IEEE International Conference on Multimedia and Expo (ICME). pp 1\u20136. IEEE","DOI":"10.1109\/ICME.2018.8486584"},{"key":"15739_CR6","doi-asserted-by":"publisher","unstructured":"Chi L, Tian G, Mu Y, Tian Q (2019) Two-stream video classification with cross-modality attention. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops. pp 0\u20130 . https:\/\/doi.org\/10.1109\/ICCVW.2019.00552","DOI":"10.1109\/ICCVW.2019.00552"},{"key":"15739_CR7","unstructured":"Choi B, Wang Y, Hannuksela M, Lim Y, Murtaza A (2017) Information technology-coded representation of immersive media (mpeg-i)-part 2: Omnidirectional media format. ISO\/IEC 23090\u20132"},{"key":"15739_CR8","unstructured":"der Auwera GV, Coban M, Karczewicz M (2016) AHG8: Truncated square pyramid projection (tsp) for 360 video. JVET Doc 0071"},{"key":"15739_CR9","doi-asserted-by":"publisher","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proc IEEE Conf Comput Vis Pattern Recognit. pp 770\u2013778 . https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"15739_CR10","doi-asserted-by":"publisher","unstructured":"Jain SD, Xiong B, Grauman K (2017) Fusionseg: Learning to combine motion and appearance for fully automatic segmentation of generic objects in videos. In: 2017 IEEE Conf Comput Vis Pattern Recognit (CVPR). pp 2117\u20132126. https:\/\/doi.org\/10.1109\/CVPR.2017.228. IEEE","DOI":"10.1109\/CVPR.2017.228"},{"key":"15739_CR11","doi-asserted-by":"publisher","unstructured":"Kendall A, Gal Y, Cipolla R (2018) Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In: Proc IEEE Conf Comput Vis Pattern Recognit. pp 7482\u20137491. https:\/\/doi.org\/10.1109\/CVPR.2018.00781","DOI":"10.1109\/CVPR.2018.00781"},{"issue":"4","key":"15739_CR12","doi-asserted-by":"publisher","first-page":"1646","DOI":"10.1109\/TIP.2018.2880509","volume":"28","author":"HG Kim","year":"2018","unstructured":"Kim HG, Lim H-T, Lee S, Ro YM (2018) Vrsa net: Vr sickness assessment considering exceptional motion for 360 vr video. IEEE Trans Image Process 28(4):1646\u20131660. https:\/\/doi.org\/10.1109\/TIP.2018.2880509","journal-title":"IEEE Trans Image Process"},{"issue":"4","key":"15739_CR13","doi-asserted-by":"publisher","first-page":"917","DOI":"10.1109\/TCSVT.2019.2898732","volume":"30","author":"HG Kim","year":"2019","unstructured":"Kim HG, Lim H-T, Ro YM (2019) Deep virtual reality image quality assessment with human perception guider for omnidirectional image. IEEE Transactions on Circuits and Systems for Video Technology 30(4):917\u2013928. https:\/\/doi.org\/10.1109\/TCSVT.2019.2898732","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"key":"15739_CR14","doi-asserted-by":"publisher","unstructured":"Kim W, Kim J, Ahn S, Kim J, Lee S (2018) Deep video quality assessor: From spatio-temporal visual sensitivity to a convolutional neural aggregation network. In: Proceedings of the European Conference on Computer Vision (ECCV). pp 219\u2013234. https:\/\/doi.org\/10.1007\/978-3-030-01246-5_14","DOI":"10.1007\/978-3-030-01246-5_14"},{"key":"15739_CR15","doi-asserted-by":"publisher","unstructured":"Kim J, Lee S (2017) Deep learning of human visual sensitivity in image quality assessment framework. In: Proc IEEE Conf Comput Vis Pattern Recognit. pp 1676\u20131684. https:\/\/doi.org\/10.1109\/CVPR.2017.213","DOI":"10.1109\/CVPR.2017.213"},{"key":"15739_CR16","unstructured":"Kingma DP, Ba J (2014) Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980"},{"key":"15739_CR17","doi-asserted-by":"publisher","unstructured":"Lim H-T, Kim HG, Ra, YM (2018) Vr iqa net: Deep virtual reality image quality assessment using adversarial learning. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). pp 6737\u20136741. https:\/\/doi.org\/10.1109\/ICASSP.2018.8461317. IEEE","DOI":"10.1109\/ICASSP.2018.8461317"},{"key":"15739_CR18","doi-asserted-by":"publisher","unstructured":"Li C, Xu M, Du X, Wang Z (2018) Bridge the gap between vqa and human behavior on omnidirectional video: A large-scale dataset and a deep learning model. In: Proceedings of the 26th ACM International Conference on Multimedia. pp 932\u2013940. https:\/\/doi.org\/10.1145\/3240508.3240581","DOI":"10.1145\/3240508.3240581"},{"key":"15739_CR19","doi-asserted-by":"publisher","unstructured":"Li C, Xu M, Jiang L, Zhang S, Tao X (2019) Viewport proposal cnn for 360$$^{\\circ }$$ video quality assessment. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp 10169\u201310178. https:\/\/doi.org\/10.1109\/CVPR.2019.01042. IEEE","DOI":"10.1109\/CVPR.2019.01042"},{"key":"15739_CR20","doi-asserted-by":"crossref","unstructured":"Mahmoudpour S, Schelkens P (2019) Visual quality analysis of judder effect on head mounted displays. In: 2019 27th European Signal Processing Conference (EUSIPCO). pp 1\u20135. IEEE","DOI":"10.23919\/EUSIPCO.2019.8902665"},{"key":"15739_CR21","doi-asserted-by":"publisher","unstructured":"Mangiante S, Klas G, Navon A, Guanhua Z, Ran J, Silva MD (2017) VR is on the edge: How to deliver 360$$^{\\circ }$$ videos in mobile networks. In: Proceedings of the Workshop on Virtual Reality and Augmented Reality Network . https:\/\/doi.org\/10.1145\/3097895.3097901","DOI":"10.1145\/3097895.3097901"},{"issue":"1","key":"15739_CR22","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1109\/TCSVT.2004.839989","volume":"15","author":"K-T Ng","year":"2005","unstructured":"Ng K-T, Chan S-C, Shum H-Y (2005) Data compression and transmission aspects of panoramic videos. IEEE Transactions on Circuits and Systems for Video Technology 15(1):82\u201395. https:\/\/doi.org\/10.1109\/TCSVT.2004.839989","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"key":"15739_CR23","doi-asserted-by":"publisher","unstructured":"Qiu Z, Yao T, Mei T (2017) Learning spatio-temporal representation with pseudo-3d residual networks. In: Proceedings of the IEEE International Conference on Computer Vision. pp 5533\u20135541. https:\/\/doi.org\/10.1109\/ICCV.2017.590","DOI":"10.1109\/ICCV.2017.590"},{"issue":"2","key":"15739_CR24","doi-asserted-by":"publisher","first-page":"335","DOI":"10.1109\/TIP.2009.2034992","volume":"19","author":"K Seshadrinathan","year":"2009","unstructured":"Seshadrinathan K, Bovik AC (2009) Motion tuned spatio-temporal quality assessment of natural videos. IEEE Trans Image Process 19(2):335\u2013350. https:\/\/doi.org\/10.1109\/TIP.2009.2034992","journal-title":"IEEE Trans Image Process"},{"issue":"6","key":"15739_CR25","doi-asserted-by":"publisher","first-page":"1427","DOI":"10.1109\/TIP.2010.2042111","volume":"19","author":"K Seshadrinathan","year":"2010","unstructured":"Seshadrinathan K, Soundararajan R, Bovik AC, Cormack LK (2010) Study of subjective and objective quality assessment of video. IEEE Trans Image Process 19(6):1427\u20131441. https:\/\/doi.org\/10.1109\/TIP.2010.2042111","journal-title":"IEEE Trans Image Process"},{"key":"15739_CR26","doi-asserted-by":"publisher","unstructured":"Shi L, Zhang Y, Cheng J, Lu H (2019) Two-stream adaptive graph convolutional networks for skeletonbased action recognition. In: Proc IEEE\/CVF Conf Comput Vis Pattern Recognit. pp 12026\u201312035. https:\/\/doi.org\/10.1109\/CVPR.2019.01230","DOI":"10.1109\/CVPR.2019.01230"},{"key":"15739_CR27","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556"},{"issue":"9","key":"15739_CR28","doi-asserted-by":"publisher","first-page":"1408","DOI":"10.1109\/LSP.2017.2720693","volume":"24","author":"Y Sun","year":"2017","unstructured":"Sun Y, Lu A, Yu L (2017) Weighted-to-spherically-uniform quality evaluation for omnidirectional video. IEEE signal processing letters 24(9):1408\u20131412. https:\/\/doi.org\/10.1109\/LSP.2017.2720693","journal-title":"IEEE signal processing letters"},{"issue":"1","key":"15739_CR29","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1109\/JSTSP.2019.2955024","volume":"14","author":"W Sun","year":"2019","unstructured":"Sun W, Min X, Zhai G, Gu K, Duan H, Ma S (2019) Mc360iqa: A multi-channel cnn for blind 360-degree image quality assessment. IEEE Journal of Selected Topics in Signal Processing 14(1):64\u201377","journal-title":"IEEE Journal of Selected Topics in Signal Processing"},{"key":"15739_CR30","unstructured":"Tan M, Le Q (2019) Efficientnet: Rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning. pp 6105\u20136114. PMLR"},{"key":"15739_CR31","doi-asserted-by":"publisher","unstructured":"Tran D, Wang H, Torresani L, Ray J, LeCun Y, Paluri M (2018) A closer look at spatiotemporal convolutions for action recognition. In: Proc IEEE Conf Comput Vis Pattern Recognit. pp 6450\u20136459. https:\/\/doi.org\/10.1109\/CVPR.2018.00675","DOI":"10.1109\/CVPR.2018.00675"},{"key":"15739_CR32","doi-asserted-by":"publisher","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y, Lin D, Tang X, Van Gool L (2016) Temporal segment networks: Towards good practices for deep action recognition. In: European Conference on Computer Vision. pp 20\u201336. Springer. https:\/\/doi.org\/10.1007\/978-3-319-46484-8_2","DOI":"10.1007\/978-3-319-46484-8_2"},{"issue":"12","key":"15739_CR33","doi-asserted-by":"publisher","first-page":"3516","DOI":"10.1109\/TCSVT.2018.2886277","volume":"29","author":"M Xu","year":"2018","unstructured":"Xu M, Li C, Chen Z, Wang Z, Guan Z (2018) Assessing visual quality of omnidirectional videos. IEEE transactions on circuits and systems for video technology 29(12):3516\u20133530. https:\/\/doi.org\/10.1109\/TCSVT.2018.2886277","journal-title":"IEEE transactions on circuits and systems for video technology"},{"issue":"1","key":"15739_CR34","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1109\/JSTSP.2020.2966864","volume":"14","author":"M Xu","year":"2020","unstructured":"Xu M, Li C, Zhang S, Le Callet P (2020) State-of-the-art in 360$$^{\\circ }$$ video\/image processing: Perception, assessment and compression. IEEE Journal of Selected Topics in Signal Processing 14(1):5\u201326. https:\/\/doi.org\/10.1109\/JSTSP.2020.2966864","journal-title":"IEEE Journal of Selected Topics in Signal Processing"},{"issue":"5","key":"15739_CR35","doi-asserted-by":"publisher","first-page":"1724","DOI":"10.1109\/TCSVT.2020.3015186","volume":"31","author":"J Xu","year":"2020","unstructured":"Xu J, Zhou W, Chen Z (2020) Blind omnidirectional image quality assessment with viewport oriented graph convolutional networks. IEEE Transactions on Circuits and Systems for Video Technology 31(5):1724\u20131737. https:\/\/doi.org\/10.1109\/TCSVT.2020.3015186","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"key":"15739_CR36","doi-asserted-by":"crossref","unstructured":"Xu J, Luo Z, Zhou W, Zhang W, Chen Z (2019) Quality assessment of stereoscopic 360-degree images from multi-viewports. In: 2019 Picture Coding Symposium (PCS). pp 1\u20135. IEEE","DOI":"10.1109\/PCS48520.2019.8954555"},{"key":"15739_CR37","doi-asserted-by":"crossref","unstructured":"Yang S, Zhao J, Jiang T, Wang J, Rahim T, Zhang B, Xu Z, Fei Z (2017) An objective assessment method based on multi-level factors for panoramic videos. In: 2017 IEEE Visual Communications and Image Processing (VCIP). pp 1\u20134. IEEE","DOI":"10.1109\/VCIP.2017.8305133"},{"key":"15739_CR38","doi-asserted-by":"publisher","unstructured":"Yu M, Lakshman H, Girod B (2015) A framework to evaluate omnidirectional video coding schemes. In: 2015 IEEE International Symposium on Mixed and Augmented Reality. pp 31\u201336. https:\/\/doi.org\/10.1109\/ISMAR.2015.12. IEEE","DOI":"10.1109\/ISMAR.2015.12"},{"key":"15739_CR39","doi-asserted-by":"publisher","unstructured":"Zach C, Pock T, Bischof H (2007) A duality based approach for realtime tv-l 1 optical flow. In: Joint Pattern Recognition Symposium, pp. 214-223 . Springer. https:\/\/doi.org\/10.1007\/978-3-540-74936-3_22","DOI":"10.1007\/978-3-540-74936-3_22"},{"key":"15739_CR40","doi-asserted-by":"publisher","unstructured":"Zakharchenko V, Choi KP, Park JH (2016) Quality metric for spherical panoramic video. In: Optics and Photonics for Information Processing X, vol 9970. p 99700 . https:\/\/doi.org\/10.1117\/12.2235885. International Society for Optics and Photonics","DOI":"10.1117\/12.2235885"},{"key":"15739_CR41","doi-asserted-by":"publisher","first-page":"8326","DOI":"10.1109\/TIP.2020.3013162","volume":"29","author":"T Zhou","year":"2020","unstructured":"Zhou T, Li J, Wang S, Tao R, Shen J (2020) Matnet: Motion-attentive transition network for zero-shot video object segmentation. IEEE Trans Image Process 29:8326\u20138338. https:\/\/doi.org\/10.1109\/TIP.2020.3013162","journal-title":"IEEE Trans Image Process"},{"key":"15739_CR42","doi-asserted-by":"publisher","unstructured":"Zhou Y, Yu M, Ma H, Shao H, Jiang G (2018) Weighted-to-spherically-uniform ssim objective quality evaluation for panoramic video. In: 2018 14th IEEE International Conference on Signal Processing (ICSP). pp 54\u201357. https:\/\/doi.org\/10.1109\/ICSP.2018.8652269. IEEE","DOI":"10.1109\/ICSP.2018.8652269"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-15739-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-15739-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-15739-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,10]],"date-time":"2024-01-10T09:20:15Z","timestamp":1704878415000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-15739-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,22]]},"references-count":42,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,1]]}},"alternative-id":["15739"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-15739-6","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,22]]},"assertion":[{"value":"28 November 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 March 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 April 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 June 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}]}}