{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,14]],"date-time":"2025-10-14T14:47:06Z","timestamp":1760453226443,"version":"3.37.3"},"reference-count":138,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2023,5,16]],"date-time":"2023-05-16T00:00:00Z","timestamp":1684195200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,5,16]],"date-time":"2023-05-16T00:00:00Z","timestamp":1684195200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100003593","name":"Conselho Nacional de Desenvolvimento Cient\u00edfico e Tecnol\u00e1gico","doi-asserted-by":"publisher","award":["304836\/2022-2"],"award-info":[{"award-number":["304836\/2022-2"]}],"id":[{"id":"10.13039\/501100003593","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001807","name":"Funda\u00e7\u00e3o de Amparo \u00e0 Pesquisa do Estado de S\u00e3o Paulo","doi-asserted-by":"publisher","award":["2022\/12294-8"],"award-info":[{"award-number":["2022\/12294-8"]}],"id":[{"id":"10.13039\/501100001807","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Sign Process Syst"],"published-print":{"date-parts":[[2023,6]]},"DOI":"10.1007\/s11265-023-01872-w","type":"journal-article","created":{"date-parts":[[2023,5,16]],"date-time":"2023-05-16T14:02:53Z","timestamp":1684245773000},"page":"679-702","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Survey on Video Colorization: Concepts, Methods and Applications"],"prefix":"10.1007","volume":"95","author":[{"given":"Leandro","family":"Stival","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0125-630X","authenticated-orcid":false,"given":"Helio","family":"Pedrini","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,5,16]]},"reference":[{"key":"1872_CR1","doi-asserted-by":"crossref","unstructured":"Smith, L. (2020). Colouring Photographs - No.I. In color and Victorian Photography (pp. 151\u2013155). Routledge.","DOI":"10.4324\/9781003084976-24"},{"key":"1872_CR2","first-page":"1","volume":"114","author":"S Huang","year":"2022","unstructured":"Huang, S., Jin, X., Jiang, Q., & Liu, L. (2022). Deep learning for image colorization: Current and future prospects. Engineering Applications of Artificial Intelligence, 114, 1\u201327.","journal-title":"Engineering Applications of Artificial Intelligence"},{"key":"1872_CR3","unstructured":"Kumar,\u00a0M., Weissenborn,\u00a0D., & Kalchbrenner, N. (2021) Colorization transformer (pp. 1\u201324). Preprint retrieved from https:\/\/arxiv.org\/abs\/2102.04432"},{"key":"1872_CR4","doi-asserted-by":"crossref","unstructured":"Pierre, F., & Aujol, J.-F. (2021). Recent approaches for image colorization. Handbook of Mathematical Models and Algorithms in Computer Vision and Imaging: Mathematical Imaging and Vision (pp. 1\u201338).","DOI":"10.1007\/978-3-030-03009-4_55-1"},{"key":"1872_CR5","doi-asserted-by":"crossref","unstructured":"Pucci, R., Micheloni, C., & Martinel, N. (2021). Collaboration among image and object level features for image colourisation (pp. 1\u20139). Preprint retrieved from http:\/\/arxiv.org\/abs\/2101.07576","DOI":"10.1109\/CVPRW53098.2021.00245"},{"issue":"7","key":"1872_CR6","doi-asserted-by":"crossref","first-page":"1765","DOI":"10.1049\/ipr2.12452","volume":"16","author":"X Qin","year":"2022","unstructured":"Qin, X., Li, M., Liu, Y., Zheng, H., Chen, J., & Zhang, M. (2022). An efficient coding-based grayscale image automatic colorization method combined with attention mechanism. IET Image Processing, 16(7), 1765\u20131777.","journal-title":"IET Image Processing"},{"key":"1872_CR7","doi-asserted-by":"crossref","first-page":"347","DOI":"10.5201\/ipol.2022.403","volume":"12","author":"A Salmona","year":"2022","unstructured":"Salmona, A., Bouza, L., & Delon, J. (2022). Deoldify: A review and implementation of an automatic colorization method. Image Processing On Line, 12, 347\u2013368.","journal-title":"Image Processing On Line"},{"key":"1872_CR8","doi-asserted-by":"crossref","unstructured":"Su, J.-W., Chu, H.-K., & Huang, J.-B. (2020). Instance-aware image colorization. IEEE \/ CVF Computer Vision and Pattern Recognition Conference (pp. 1\u201310).","DOI":"10.1109\/CVPR42600.2020.00799"},{"issue":"5","key":"1872_CR9","doi-asserted-by":"crossref","first-page":"1120","DOI":"10.1109\/TIP.2005.864231","volume":"15","author":"L Yatziv","year":"2006","unstructured":"Yatziv, L., & Sapiro, G. (2006). Fast image and video colorization using chrominance blending. IEEE Transactions on Image Processing, 15(5), 1120\u20131129.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1872_CR10","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., & Efros, A.\u00a0A. (2016). Colorful image colorization. In European Conference on Computer Vision (pp. 649\u2013666). Springer.","DOI":"10.1007\/978-3-319-46487-9_40"},{"issue":"6","key":"1872_CR11","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3355089.3356570","volume":"38","author":"S Iizuka","year":"2019","unstructured":"Iizuka, S., & Simo-Serra, E. (2019). Deepremaster: Temporal source-reference attention networks for comprehensive video enhancement. ACM Transactions on Graphics (TOG), 38(6), 1\u201313.","journal-title":"ACM Transactions on Graphics (TOG)"},{"key":"1872_CR12","doi-asserted-by":"crossref","unstructured":"Jampour, M., Zare, M., & Javidi, M. (2022). Advanced multi-gans towards near to real image and video colorization. Journal of Ambient Intelligence and Humanized Computing, 1\u201318.","DOI":"10.1007\/s12652-022-04206-z"},{"key":"1872_CR13","doi-asserted-by":"crossref","unstructured":"Lei, C., Xing, Y., Ouyang, H., & Chen, Q. (2022). Deep video prior for video consistency and propagation. IEEE Transactions on Pattern Analysis and Machine Intelligence (pp. 1\u201317).","DOI":"10.1109\/TPAMI.2022.3142071"},{"key":"1872_CR14","unstructured":"Liu, Y., Zhao, H., Chan, K.\u00a0C., Wang, X., Loy, C.\u00a0C., Qiao, Y., & Dong, C. (2021c). Temporally consistent video colorization with deep feature propagation and self-regularization learning. Preprint retrieved from https:\/\/arxiv.org\/abs\/2110.04562 (pp. 1\u201317)."},{"key":"1872_CR15","doi-asserted-by":"crossref","unstructured":"Mahajan, A., Patel, N., Kotak, A., & Palkar, B. (2021). An end-to-end approach for automatic and consistent colorization of gray-scale videos using deep-learning techniques. In International Conference on Machine Intelligence and Data Science Applications (pp. 539\u2013551). Springer.","DOI":"10.1007\/978-981-33-4087-9_45"},{"key":"1872_CR16","unstructured":"Shi, M., Zhang, J.-Q., Chen, S.-Y., Gao, L., Lai, Y.-K., & Zhang, F.-L. (2020). Deep line art video colorization with a few references (pp. 1\u201310). Preprint retrieved from http:\/\/arxiv.org\/abs\/2003.10685"},{"key":"1872_CR17","unstructured":"Veluri, B., Saffari, A., Pernu, C., Smith, J., Taylor, M., & Gollakota, S. (2022). Neuricam: Video super-resolution and colorization using key frames (pp. 1\u201317). Preprint retrieved from http:\/\/arxiv.org\/abs\/2207.12496"},{"key":"1872_CR18","doi-asserted-by":"crossref","unstructured":"Yang, Y., Liu, Y., Yuan, H., & Chu, Y. (2022b). Deep colorization: A channel attention-based CNN for video colorization. International Conference on Image and Graphics Processing (pp. 275\u2013280).","DOI":"10.1145\/3512388.3512428"},{"key":"1872_CR19","doi-asserted-by":"crossref","unstructured":"Zhang, B., He, M., Liao, J., Sander, P.\u00a0V., Yuan, L., Bermak, A., & Chen, D. (2019). Deep exemplar-based video colorization. IEEE \/ CVF Computer Vision and Pattern Recognition Conference (pp. 8052\u20138061).","DOI":"10.1109\/CVPR.2019.00824"},{"key":"1872_CR20","first-page":"675","volume":"1","author":"A Survey","year":"2020","unstructured":"Survey, A. (2020). M. E. A-Monem and T. Z. Hammood. Video colorization methods. Iraqi Journal of Science, 1, 675\u2013686.","journal-title":"Iraqi Journal of Science"},{"key":"1872_CR21","first-page":"1","volume":"1","author":"S-Y Chen","year":"2022","unstructured":"Chen, S.-Y., Zhang, J.-Q., Zhao, Y.-Y., Rosin, P. L., Lai, Y.-K., & Gao, L. (2022). A review of image and video colorization: From analogies to deep learning. Visual Informatics, 1, 1\u201318.","journal-title":"Visual Informatics"},{"key":"1872_CR22","unstructured":"Anwar, S., Tahir, M., Li, C., Mian, A., Khan, F.\u00a0S., & Muzaffar, A.\u00a0W. (2020). Image colorization: A survey and dataset.\u00a01, 1\u201320.\u00a0Preprint retrieved from\u00a0http:\/\/arxiv.org\/abs\/2008.10774"},{"key":"1872_CR23","doi-asserted-by":"crossref","unstructured":"Levin, A., Lischinski, D., & Weiss, Y. (2004, August). Colorization using optimization. ACM Transactions on Graphics, 23(3), 689\u2013694.","DOI":"10.1145\/1015706.1015780"},{"issue":"5","key":"1872_CR24","doi-asserted-by":"crossref","first-page":"1120","DOI":"10.1109\/TIP.2005.864231","volume":"15","author":"L Yatziv","year":"2006","unstructured":"Yatziv, L., & Sapiro, G. (2006). Fast image and video colorization using chrominance blending. IEEE Transactions on Image Processing, 15(5), 1120\u20131129.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1872_CR25","doi-asserted-by":"crossref","unstructured":"Heu, J.-H., Hyun, D.-Y., Kim, C.-S., & Lee, S.-U. (2009). Image and video colorization based on prioritized source propagation. In 16th IEEE International Conference on Image Processing (pp. 465\u2013468). IEEE.","DOI":"10.1109\/ICIP.2009.5414371"},{"key":"1872_CR26","unstructured":"Do\u011fan, P., Ayd\u0131n, T.\u00a0O., Stefanoski, N., & Smolic, A. (2015). Key-frame based spatiotemporal scribble propagation. In Eurographics Workshop on Intelligent Cinematography and Editing (pp. 13\u201320)."},{"issue":"8","key":"1872_CR27","doi-asserted-by":"crossref","first-page":"1605","DOI":"10.1109\/TCSVT.2016.2539539","volume":"27","author":"S Paul","year":"2017","unstructured":"Paul, S., Bhattacharya, S., & Gupta, S. (2017). Spatiotemporal colorization of video using 3D steerable pyramids. IEEE Transactions on Circuits and Systems for Video Technology, 27(8), 1605\u20131619.","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"key":"1872_CR28","doi-asserted-by":"crossref","unstructured":"Welsh, T., Ashikhmin, M., & Mueller, K. (2002). Transferring color to greyscale images. In 29th Annual Conference on Computer Graphics and Interactive Techniques (pp. 277\u2013280).","DOI":"10.1145\/566570.566576"},{"key":"1872_CR29","doi-asserted-by":"crossref","unstructured":"Vondrick, C., Shrivastava, A., Fathi, A., Guadarrama, S., & Murphy, K. (2018). Tracking emerges by colorizing videos. In European Conference on Computer Vision (pp. 391\u2013408).","DOI":"10.1007\/978-3-030-01261-8_24"},{"key":"1872_CR30","unstructured":"Meyer, S., Cornill\u00e8re, V., Djelouah, A., Schroers, C., & Gross, M. (2018). Deep video color propagation (pp. 1\u201315). Preprint retrieved from http:\/\/arxiv.org\/abs\/1808.03232"},{"key":"1872_CR31","doi-asserted-by":"crossref","unstructured":"Kouzouglidis, P., Sfikas, G., & Nikou, C. (2019). Automatic video colorization using 3D conditional generative adversarial networks. In International Symposium on Visual Computing (pp. 209\u2013218). Springer.","DOI":"10.1007\/978-3-030-33720-9_16"},{"key":"1872_CR32","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Po, L.-M., Yu, W.-Y., Rehman, Y.\u00a0A.\u00a0U., Liu, M., Zhang, Y., & Ou, W. (2021). VCGAN: Video colorization with hybrid generative adversarial network (pp. 1\u201315). Preprint retrieved from http:\/\/arxiv.org\/abs\/2104.12357","DOI":"10.1109\/TMM.2022.3154600"},{"key":"1872_CR33","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition (pp. 770\u2013778).","DOI":"10.1109\/CVPR.2016.90"},{"issue":"3","key":"1872_CR34","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al. (2015). Imagenet large scale visual recognition challenge. International Journal of Computer Vision, 115(3), 211\u2013252.","journal-title":"International Journal of Computer Vision"},{"key":"1872_CR35","unstructured":"Akimoto, N., Hayakawa, A., Shin, A., & Narihira, T. (2020). Reference-based video colorization with spatiotemporal correspondence. CoRR, abs\/2011.12528, 1\u201314."},{"key":"1872_CR36","doi-asserted-by":"crossref","unstructured":"S\u1ef3kora, D., Buri\u00e1nek, J., & \u017d\u00e1ra, J. (2004). Unsupervised colorization of black-and-white cartoons. In 3rd International Symposium on Non-Photorealistic Animation and Rendering (pp. 121\u2013127).","DOI":"10.1145\/987657.987677"},{"key":"1872_CR37","unstructured":"Shi, M., Zhang, J.-Q., Chen, S.-Y., Gao, L., Lai, Y., & Zhang, F.-L. (2022). Reference-based deep line art video colorization. IEEE Transactions on Visualization and Computer Graphics (pp. 1\u201315)."},{"key":"1872_CR38","doi-asserted-by":"crossref","unstructured":"Thasarathan, H., Nazeri, K., & Ebrahimi, M. (2019). Automatic temporally coherent video colorization. In 16th Conference on Computer and Robot Vision (pp. 189\u2013194). IEEE.","DOI":"10.1109\/CRV.2019.00033"},{"key":"1872_CR39","doi-asserted-by":"crossref","unstructured":"Ilg, E., Mayer, N., Saikia, T., Keuper, M., Dosovitskiy, A., & Brox, T. (2017). Flownet 2.0: Evolution of optical flow estimation with deep networks. In IEEE Conference on Computer Vision and Pattern Recognition (pp 2462\u20132470).","DOI":"10.1109\/CVPR.2017.179"},{"key":"1872_CR40","doi-asserted-by":"crossref","unstructured":"Lei, C., & Chen, Q. (2019). Fully automatic video colorization with self-regularization and diversity. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 3753\u20133761).","DOI":"10.1109\/CVPR.2019.00387"},{"key":"1872_CR41","doi-asserted-by":"crossref","unstructured":"Li, Z., Chen, Q., & Koltun, V. (2018, June). Interactive image segmentation with latent diversity. In IEEE Conference on Computer Vision and Pattern Recognition (pp. 1\u20139).","DOI":"10.1109\/CVPR.2018.00067"},{"issue":"6","key":"1872_CR42","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/2816795.2818107","volume":"34","author":"N Bonneel","year":"2015","unstructured":"Bonneel, N., Tompkin, J., Sunkavalli, K., Sun, D., Paris, S., & Pfister, H. (2015). Blind video temporal consistency. ACM Transactions on Graphics, 34(6), 1\u20139.","journal-title":"ACM Transactions on Graphics"},{"key":"1872_CR43","doi-asserted-by":"crossref","unstructured":"Siyao, L., Zhao, S., Yu, W., Sun, W., Metaxas, D., Loy, C.\u00a0C., & Liu, Z. (2021). Deep animation video interpolation in the wild. In IEEE\/CVF Computer Vision and Pattern Recognition Conference (pp. 1\u20139).","DOI":"10.1109\/CVPR46437.2021.00652"},{"issue":"6","key":"1872_CR44","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2017). Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60(6), 84\u201390.","journal-title":"Communications of the ACM"},{"key":"1872_CR45","doi-asserted-by":"crossref","unstructured":"Graupe, D. (2013). Principles of artificial neural networks (Vol.\u00a07). World Scientific.","DOI":"10.1142\/8868"},{"key":"1872_CR46","unstructured":"G\u00e9ron, A. (2019). Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow: Concepts, tools, and techniques to build intelligent systems. O\u2019Reilly Media, Inc."},{"issue":"11","key":"1872_CR47","doi-asserted-by":"crossref","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y LeCun","year":"1998","unstructured":"LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278\u20132324.","journal-title":"Proceedings of the IEEE"},{"key":"1872_CR48","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., & Brox, T. (2015). U-Net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention (pp. 234\u2013241). Springer.","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"11","key":"1872_CR49","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., & Bengio, Y. (2020). Generative adversarial networks. Communications of the ACM, 63(11), 139\u2013144.","journal-title":"Communications of the ACM"},{"issue":"1","key":"1872_CR50","doi-asserted-by":"crossref","first-page":"53","DOI":"10.1109\/MSP.2017.2765202","volume":"35","author":"A Creswell","year":"2018","unstructured":"Creswell, A., White, T., Dumoulin, V., Arulkumaran, K., Sengupta, B., & Bharath, A. A. (2018). Generative adversarial networks: An overview. IEEE Signal Processing Magazine, 35(1), 53\u201365.","journal-title":"IEEE Signal Processing Magazine"},{"key":"1872_CR51","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.\u00a0N., Kaiser, L.\u00a0U., & Polosukhin, I. (2017). Attention is all you need. In I.\u00a0Guyon, U.\u00a0V. Luxburg, S.\u00a0Bengio, H.\u00a0Wallach, R.\u00a0Fergus, S.\u00a0Vishwanathan, & R.\u00a0Garnett (Eds.), Advances in Neural Information Processing Systems (Vol.\u00a030, pp. 1\u201311). Curran Associates, Inc."},{"key":"1872_CR52","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., & Houlsby, N. (2021). An image is worth 16$$\\times$$16 words: Transformers for image recognition at scale. International Conference on Learning Representations, 1."},{"key":"1872_CR53","doi-asserted-by":"crossref","unstructured":"Marszalek, M., Laptev, I., & Schmid, C. (2009). Actions in context. In IEEE Conference on Computer Vision and Pattern Recognition (pp. 2929\u20132936). IEEE.","DOI":"10.1109\/CVPR.2009.5206557"},{"key":"1872_CR54","doi-asserted-by":"crossref","unstructured":"Perazzi, F., Pont-Tuset, J., McWilliams, B., Van\u00a0Gool, L., Gross, M., & Sorkine-Hornung, A. (2016). A benchmark dataset and evaluation methodology for video object segmentation. In IEEE Conference on Computer Vision and Pattern Recognition (pp. 724\u2013732).","DOI":"10.1109\/CVPR.2016.85"},{"key":"1872_CR55","unstructured":"Kay, W., Carreira, J., Simonyan, K., Zhang, B., Hillier, C., Vijayanarasimhan, S., Viola, F., Green, T., Back, T., Natsev, P., et\u00a0al. (2017). The Kinetics human action video dataset. Preprint retrieved from http:\/\/arxiv.org\/abs\/1705.06950"},{"key":"1872_CR56","first-page":"1998","volume":"1\u201331","author":"A Ford","year":"1998","unstructured":"Ford, A., & Roberts, A. (1998). Colour space conversions. Westminster University, London, 1\u201331, 1998.","journal-title":"Westminster University, London"},{"key":"1872_CR57","doi-asserted-by":"crossref","unstructured":"Ballester, C., Bugeau, A., Carrillo, H., Cl\u00e9ment, M., Giraud, R., Raad, L., & Vitoria, P. (2022b) Influence of color spaces for deep learning image colorization (pp. 1\u201331). Preprint retrieved from http:\/\/arxiv.org\/abs\/2204.02850","DOI":"10.1007\/978-3-030-03009-4_125-1"},{"key":"1872_CR58","doi-asserted-by":"crossref","unstructured":"Riba, E., Mishkin, D., Ponsa, D., Rublee, E., & Bradski, G. (2020). Kornia: An open source differentiable computer vision library for pytorch. In IEEE\/CVF Winter Conference on Applications of Computer Vision (pp. 3674\u20133683).","DOI":"10.1109\/WACV45572.2020.9093363"},{"key":"1872_CR59","unstructured":"Bradski, G., & Kaehler, A. (2008). Learning OpenCV: Computer vision with the OpenCV library. O\u2019Reilly Media, Inc."},{"key":"1872_CR60","doi-asserted-by":"crossref","unstructured":"Podpora, M., Korbas, G.\u00a0P., & Kawala-Janik, A. (2014). YUV vs RGB-choosing a color space for human-machine interaction. In FedCSIS (Position Papers) (pp. 29\u201334).","DOI":"10.15439\/2014F206"},{"key":"1872_CR61","doi-asserted-by":"crossref","unstructured":"Loesdau, M., Chabrier, S., & Gabillon, A. (2014). Hue and saturation in the RGB color space. In International Conference on Image and Signal Processing (pp. 203\u2013212). Springer.","DOI":"10.1007\/978-3-319-07998-1_23"},{"key":"1872_CR62","unstructured":"Luo, M.\u00a0R. (2014). CIELAB\u00a0(pp. 43\u201350). Springer Berlin Heidelberg, Berlin, Heidelberg. ISBN 978-3-642-27851-8."},{"key":"1872_CR63","unstructured":"Yang, H., Nan, G., Lin, M., Chao, F., Shen, Y., Li, K., & Ji, R. (2022a). LAB-Net: LAB color-space oriented lightweight network for shadow removal (pp. 1\u201310). Preprint retrieved from http:\/\/arxiv.org\/abs\/2208.13039"},{"issue":"7","key":"1872_CR64","doi-asserted-by":"crossref","first-page":"1046","DOI":"10.1109\/83.597279","volume":"6","author":"C Connolly","year":"1997","unstructured":"Connolly, C., & Fleiss, T. (1997). A study of efficiency and accuracy in the transformation from RGB to CIELAB color space. IEEE Transactions on Image Processing, 6(7), 1046\u20131048.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1872_CR65","unstructured":"Tong, X., Heeger, D.\u00a0J., & Van\u00a0den Branden\u00a0Lambrecht, C.\u00a0J. (1999). Video quality evaluation using ST-CIELAB. In Human Vision and Electronic Imaging IV (Vol. 3644, pp. 185\u2013196). SPIE."},{"key":"1872_CR66","doi-asserted-by":"crossref","unstructured":"Seymour, J. (2022). Color inconstancy in CIELAB: A red herring? Color Research & Application, 1\u201320.","DOI":"10.1002\/col.22782"},{"key":"1872_CR67","doi-asserted-by":"crossref","unstructured":"\u015eahin, C., Balc\u0131, O., I\u015f\u0131k, M., & G\u00f6ken\u00e7, \u0130. (2022). Artificial neural networks approach for prediction of CIELab values for yarn after dyeing and finishing process. The Journal of The Textile Institute,\u00a01\u201310.","DOI":"10.1080\/00405000.2022.2124629"},{"key":"1872_CR68","unstructured":"Murray,T. (2008). Digital baroque: New media art and cinematic folds (Vol.\u00a026). University of Minnesota Press."},{"issue":"1","key":"1872_CR69","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1002\/col.22291","volume":"44","author":"SY Kahu","year":"2019","unstructured":"Kahu, S. Y., Raut, R. B., & Bhurchandi, K. M. (2019). Review and evaluation of color spaces for image\/video compression. Color Research & Application, 44(1), 8\u201333.","journal-title":"Color Research & Application"},{"issue":"6","key":"1872_CR70","first-page":"78","volume":"2","author":"R Sudhir","year":"2011","unstructured":"Sudhir, R., & Baboo, L. D. S. S. (2011). An efficient CBIR technique with YUV color space and texture features. Computer Engineering and Intelligent Systems, 2(6), 78\u201385.","journal-title":"Computer Engineering and Intelligent Systems"},{"issue":"2","key":"1872_CR71","doi-asserted-by":"crossref","first-page":"187","DOI":"10.1007\/s40745-020-00253-5","volume":"9","author":"Q Wang","year":"2022","unstructured":"Wang, Q., Ma, Y., Zhao, K., & Tian, Y. (2022). A comprehensive survey of loss functions in machine learning. Annals of Data Science, 9(2), 187\u2013212.","journal-title":"Annals of Data Science"},{"key":"1872_CR72","doi-asserted-by":"crossref","unstructured":"Ballester, C., Bugeau, A., Carrillo, H., Cl\u00e9ment, M., Giraud, R., Raad, L., & Vitoria, P. (2022a). Analysis of different losses for deep learning image colorization. Preprint retrieved from http:\/\/arxiv.org\/abs\/2204.02980","DOI":"10.1007\/978-3-030-03009-4_127-1"},{"key":"1872_CR73","doi-asserted-by":"crossref","unstructured":"Kastryulin, S., Zakirov, J., Prokopenko, D., & Dylov, D.\u00a0V. (2022). PyTorch image quality: Metrics for image quality assessment (pp. 1\u201320). Preprint retrieved from http:\/\/arxiv.org\/abs\/2208.14818","DOI":"10.2139\/ssrn.4206741"},{"key":"1872_CR74","doi-asserted-by":"crossref","unstructured":"Sajjadi, M.\u00a0S., Vemulapalli, R., & Brown, M. (2018). Frame-recurrent video super-resolution. In IEEE Conference on Computer Vision and Pattern Recognition (pp. 6626\u20136634).","DOI":"10.1109\/CVPR.2018.00693"},{"key":"1872_CR75","doi-asserted-by":"crossref","unstructured":"Zhang, L., Zhang, L., Mou, X., & Zhang, D. (2012). A comprehensive evaluation of full reference image quality assessment algorithms. In 19th IEEE International Conference on Image Processing (pp. 1477\u20131480). IEEE.","DOI":"10.1109\/ICIP.2012.6467150"},{"key":"1872_CR76","doi-asserted-by":"crossref","unstructured":"Janocha, K., & Czarnecki, W.\u00a0M. (2017). On loss functions for deep neural networks in classification (pp. 1\u201310). Preprint retrieved from http:\/\/arxiv.org\/abs\/1702.05659","DOI":"10.4467\/20838476SI.16.004.6185"},{"key":"1872_CR77","unstructured":"Zhao, H., Gallo, O., Frosio, I., & J.\u00a0Kautz (2015). Loss functions for neural networks for image processing (pp. 1\u201311). Preprint retrieved from http:\/\/arxiv.org\/abs\/1511.08861"},{"key":"1872_CR78","doi-asserted-by":"crossref","unstructured":"Wang, Z., Bovik, A.\u00a0C., & Lu, L. (2002). Why is image quality assessment so difficult? In IEEE International Conference on Acoustics, Speech, and Signal Processing (Vol.\u00a04, pp. IV\u20133313). IEEE.","DOI":"10.1109\/ICASSP.2002.5745362"},{"issue":"4","key":"1872_CR79","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A., Sheikh, H., & Simoncelli, E. (2004). Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4), 600\u2013612.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1872_CR80","unstructured":"Nilsson, J., & Akenine-M\u00f6ller, T. (2020). Understanding SSIM (pp. 1\u20138). Preprint retrieved from http:\/\/arxiv.org\/abs\/2006.13846"},{"key":"1872_CR81","doi-asserted-by":"crossref","unstructured":"Tao, L., Zhu, C., Xiang, G., Li, Y., Jia, H., & Xie, X. (2017). LLCNN: A convolutional neural network for low-light image enhancement. In IEEE Visual Communications and Image Processing (pp. 1\u20134). IEEE.","DOI":"10.1109\/VCIP.2017.8305143"},{"key":"1872_CR82","doi-asserted-by":"crossref","unstructured":"Kornilov, A., Safonov, I., & Yakimchuk, I. (2020). Inpainting of ring artifacts on microtomographic images by 3D CNN. In 26th Conference of Open Innovations Association (pp. 200\u2013206). IEEE.","DOI":"10.23919\/FRUCT48808.2020.9087422"},{"issue":"1","key":"1872_CR83","doi-asserted-by":"crossref","first-page":"1","DOI":"10.3390\/rs13010001","volume":"13","author":"L Hu","year":"2020","unstructured":"Hu, L., Qin, M., Zhang, F., Du, Z., & Liu, R. (2020). RSCNN: A CNN-based method to enhance low-light remote-sensing images. Remote Sensing, 13(1), 1\u201362.","journal-title":"Remote Sensing"},{"key":"1872_CR84","doi-asserted-by":"crossref","unstructured":"Johnson, J., Alahi, A., & Fei-Fei, L. (2016). Perceptual losses for real-time style transfer and super-resolution. In European Conference on Computer Vision\u00a0(pp. 694\u2013711). Springer.","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"1872_CR85","unstructured":"Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition (pp. 1\u201314). Preprint retrieved from http:\/\/arxiv.org\/abs\/1409.1556"},{"issue":"7","key":"1872_CR86","doi-asserted-by":"crossref","first-page":"3312","DOI":"10.1109\/TIP.2019.2895768","volume":"28","author":"A Lucas","year":"2019","unstructured":"Lucas, A., Lopez-Tapia, S., Molina, R., & Katsaggelos, A. K. (2019). Generative adversarial networks and perceptual losses for video super-resolution. IEEE Transactions on Image Processing, 28(7), 3312\u20133327.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1872_CR87","doi-asserted-by":"crossref","first-page":"30536","DOI":"10.1109\/ACCESS.2018.2846546","volume":"6","author":"A Wang","year":"2018","unstructured":"Wang, A., Fang, Z., Gao, Y., Jiang, X., & Ma, S. (2018). Depth estimation of video sequences with perceptual losses. IEEE Access, 6, 30536\u201330546.","journal-title":"IEEE Access"},{"key":"1872_CR88","unstructured":"Dong, H.-W., & Yang, Y.-H. (2019). Towards a deeper understanding of adversarial losses (pp. 1\u201315). Preprint retrieved from http:\/\/arxiv.org\/abs\/1901.08753"},{"key":"1872_CR89","unstructured":"Jolicoeur-Martineau, A. (2018). The relativistic discriminator: A key element missing from standard GAN (pp. 1\u201325). Preprint retrieved from http:\/\/arxiv.org\/abs\/1807.00734"},{"key":"1872_CR90","first-page":"1","volume":"22","author":"X Yang","year":"2009","unstructured":"Yang, X., Kim, S., & Xing, E. (2009). Heterogeneous multitask learning with joint sparsity constraints. Advances in Neural Information Processing Systems, 22, 1\u20139.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"1872_CR91","unstructured":"Zhang, Y., & Yang, Q. (2021). A survey on multi-task learning. IEEE Transactions on Knowledge and Data Engineering, 1\u201324."},{"key":"1872_CR92","doi-asserted-by":"crossref","first-page":"49","DOI":"10.1016\/j.patrec.2022.02.015","volume":"157","author":"T Mondal","year":"2022","unstructured":"Mondal, T., Das, A., & Ming, Z. (2022). Exploring multi-tasking learning in document attribute classification. Pattern Recognition Letters, 157, 49\u201359.","journal-title":"Pattern Recognition Letters"},{"key":"1872_CR93","unstructured":"Kang, Z., Grauman, K., & Sha, F. (2011). Learning with whom to share in multi-task feature learning. In International Conference on Machine Learning (pp. 1\u20138)."},{"key":"1872_CR94","first-page":"1","volume":"30","author":"M Long","year":"2017","unstructured":"Long, M., Cao, Z., Wang, J., & Yu, P. S. (2017). Learning multiple tasks with multilinear relationship networks. Advances in Neural Information Processing Systems, 30, 1\u201310.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"1872_CR95","unstructured":"Liu, R., Zhang, H., Pirsiavash, H., & Liu, X. (2021a). Staf: A spatio-temporal attention fusion network for few-shot video classification (pp. 1\u201310). Preprint retrieved from http:\/\/arxiv.org\/abs\/2112.04585"},{"issue":"1","key":"1872_CR96","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/s41598-021-99269-x","volume":"12","author":"S Sukegawa","year":"2022","unstructured":"Sukegawa, S., Matsuyama, T., Tanaka, F., Hara, T., Yoshii, K., Yamashita, K., Nakano, K., Takabatake, K., Kawai, H., & Nagatsuka, H. (2022). Evaluation of multi-task learning in deep learning-based positioning classification of mandibular third molars. Scientific Reports, 12(1), 1\u201310.","journal-title":"Scientific Reports"},{"key":"1872_CR97","doi-asserted-by":"crossref","unstructured":"Bertalmio,\u00a0M., Sapiro,\u00a0G., Caselles,\u00a0V., & Ballester,\u00a0C. (2000). Image inpainting. In 27th annual Conference on Computer Graphics and Interactive Techniques (pp. 417\u2013424).","DOI":"10.1145\/344779.344972"},{"issue":"11","key":"1872_CR98","doi-asserted-by":"crossref","first-page":"1496","DOI":"10.1109\/83.469931","volume":"4","author":"AC Kokaram","year":"1995","unstructured":"Kokaram, A. C., Morris, R. D., Fitzgerald, W. J., & Rayner, P. J. (1995). Detection of missing data in image sequences. IEEE Transactions on Image Processing, 4(11), 1496\u20131508.","journal-title":"IEEE Transactions on Image Processing"},{"key":"1872_CR99","first-page":"1","volume":"25","author":"J Xie","year":"2012","unstructured":"Xie, J., Xu, L., & Chen, E. (2012). Image denoising and inpainting with deep neural networks. Advances in Neural Information Processing Systems, 25, 1\u20139.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"1872_CR100","doi-asserted-by":"crossref","unstructured":"Chang, Y.-L., Liu, Z.\u00a0Y., Lee, K.-Y., & Hsu, W. (2019). Free-form video inpainting with 3D gated convolution and temporal PatchGAN. In IEEE\/CVF International Conference on Computer Vision\u00a0(pp. 9066\u20139075).","DOI":"10.1109\/ICCV.2019.00916"},{"key":"1872_CR101","doi-asserted-by":"crossref","unstructured":"Kim, D., Woo, S., Lee, J.-Y., & Kweon, I.\u00a0S. (2019, June). Deep video inpainting. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition (pp. 1\u201310).","DOI":"10.1109\/CVPR.2019.00594"},{"issue":"8","key":"1872_CR102","doi-asserted-by":"crossref","first-page":"1280","DOI":"10.1109\/76.809162","volume":"9","author":"A Hanjalic","year":"1999","unstructured":"Hanjalic, A., & Zhang, H. (1999). An integrated scheme for automated video abstraction based on unsupervised cluster-validity analysis. IEEE Transactions on Circuits and Systems for Video Technology, 9(8), 1280\u20131289.","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"key":"1872_CR103","doi-asserted-by":"crossref","unstructured":"Wolf, W. (1996). Key frame selection by motion analysis. In IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings (Vol.\u00a02, pp. 1228\u20131231). IEEE.","DOI":"10.1109\/ICASSP.1996.543588"},{"issue":"1","key":"1872_CR104","doi-asserted-by":"crossref","first-page":"3","DOI":"10.2174\/2213275911666180719111118","volume":"11","author":"MK Asha Paul","year":"2018","unstructured":"Asha Paul, M. K., Kavitha, J., & Jansi Rani, P. A. (2018). Key-frame extraction techniques: A review. Recent Patents on Computer Science, 11(1), 3\u201316.","journal-title":"Recent Patents on Computer Science"},{"key":"1872_CR105","doi-asserted-by":"crossref","unstructured":"Dimitrova, N., McGee, T., & Elenbaas, H. (1997). Video keyframe extraction and filtering: A keyframe is not a keyframe to everyone. In Sixth International Conference on Information and Knowledge Management (pp. 113\u2013120).","DOI":"10.1145\/266714.266876"},{"issue":"1","key":"1872_CR106","doi-asserted-by":"crossref","first-page":"90","DOI":"10.1109\/T-C.1974.223784","volume":"100","author":"N Ahmed","year":"1974","unstructured":"Ahmed, N., Natarajan, T., & Rao, K. R. (1974). Discrete cosine transform. IEEE Transactions on Computers, 100(1), 90\u201393.","journal-title":"IEEE Transactions on Computers"},{"key":"1872_CR107","doi-asserted-by":"crossref","unstructured":"Zong, Z., & Gong, Q. (2017). Key frame extraction based on dynamic color histogram and fast wavelet histogram. In IEEE International Conference on Information and Automation (pp. 183\u2013188). IEEE.","DOI":"10.1109\/ICInfA.2017.8078903"},{"issue":"8","key":"1872_CR108","doi-asserted-by":"crossref","first-page":"532","DOI":"10.1007\/s00371-005-0316-0","volume":"21","author":"K-S Huang","year":"2005","unstructured":"Huang, K.-S., Chang, C.-F., Hsu, Y.-Y., & Yang, S.-N. (2005). Key probe: A technique for animation keyframe extraction. The Visual Computer, 21(8), 532\u2013541.","journal-title":"The Visual Computer"},{"key":"1872_CR109","doi-asserted-by":"crossref","unstructured":"Nixon, M.\u00a0S., & Aguado, A.\u00a0S. (2020). 8 - region-based analysis. In M.\u00a0S. Nixon & A.\u00a0S. Aguado (Eds.), Feature Extraction and Image Processing for Computer Vision (4th ed., pp. 399\u2013432). Academic Press.","DOI":"10.1016\/B978-0-12-814976-8.00008-7"},{"issue":"1","key":"1872_CR110","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1049\/cvi2.12008","volume":"15","author":"A Mukherjee","year":"2021","unstructured":"Mukherjee, A., Sarkar, S., & Saha, S. K. (2021). Segmentation of natural images based on super pixel and graph merging. IET Computer Vision, 15(1), 1\u201311.","journal-title":"IET Computer Vision"},{"key":"1872_CR111","doi-asserted-by":"crossref","unstructured":"Xing, X., Jia, X., & Meng, M.\u00a0Q.-H. (2018). Bleeding detection in wireless capsule endoscopy image video using superpixel-color histogram and a subspace KNN classifier. In 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (pp. 1\u20134). IEEE.","DOI":"10.1109\/EMBC.2018.8513012"},{"issue":"3","key":"1872_CR112","first-page":"273","volume":"20","author":"C Cortes","year":"1995","unstructured":"Cortes, C., & Vapnik, V. (1995). Support vector networks. Machine Learning, 20(3), 273\u2013297.","journal-title":"Machine Learning"},{"issue":"10","key":"1872_CR113","doi-asserted-by":"crossref","first-page":"2931","DOI":"10.1109\/TVCG.2019.2908363","volume":"26","author":"F Fang","year":"2019","unstructured":"Fang, F., Wang, T., Zeng, T., & Zhang, G. (2019). A superpixel-based variational model for image colorization. IEEE Transactions on Visualization and Computer Graphics, 26(10), 2931\u20132943.","journal-title":"IEEE Transactions on Visualization and Computer Graphics"},{"key":"1872_CR114","doi-asserted-by":"crossref","unstructured":"Faridul, H.\u00a0S., Pouli, T., Chamaret, C., Stauder, J., Reinhard, E., Kuzovkin, D., & Tr\u00e9meau, A. (2016). Colour mapping: A review of recent methods, extensions and applications. In Computer Graphics Forum (Vol. 35, pp. 59\u201388). Wiley Online Library.","DOI":"10.1111\/cgf.12671"},{"key":"1872_CR115","doi-asserted-by":"crossref","unstructured":"Liu, Y., Zhang, X., & Xu, X. (2021b). Semantic-aware video style transfer based on temporal consistent sparse patch constraint. In IEEE International Conference on Multimedia and Expo (pp. 1\u20136). IEEE.","DOI":"10.1109\/ICME51207.2021.9428352"},{"issue":"4","key":"1872_CR116","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/2461912.2461939","volume":"32","author":"N Bonneel","year":"2013","unstructured":"Bonneel, N., Sunkavalli, K., Paris, S., & Pfister, H. (2013). Example-based video color grading. ACM Transactions on Graphics, 32(4), 1\u201339.","journal-title":"ACM Transactions on Graphics"},{"key":"1872_CR117","doi-asserted-by":"crossref","unstructured":"Chen, D., Liao,\u00a0J., Yuan,\u00a0L., Yu,\u00a0N., & Hua,\u00a0G. (2017). Coherent online video style transfer. In IEEE International Conference on Computer Vision (pp. 1105\u20131114).","DOI":"10.1109\/ICCV.2017.126"},{"issue":"3","key":"1872_CR118","doi-asserted-by":"crossref","first-page":"660","DOI":"10.1109\/TBC.2008.2000733","volume":"54","author":"S Winkler","year":"2008","unstructured":"Winkler, S., & Mohandas, P. (2008). The evolution of video quality measurement: From psnr to hybrid metrics. IEEE Transactions on Broadcasting, 54(3), 660\u2013668.","journal-title":"IEEE Transactions on Broadcasting"},{"key":"1872_CR119","first-page":"1441412697","volume-title":"Python 3 Reference Manual","author":"G Van Rossum","year":"2009","unstructured":"Van Rossum, G., & Drake, F. L. (2009). Python 3 Reference Manual (p. 1441412697). Scotts Valley, CA: CreateSpace."},{"issue":"2","key":"1872_CR120","doi-asserted-by":"crossref","first-page":"330","DOI":"10.4314\/njt.v41i2.15","volume":"41","author":"U Ukommi","year":"2022","unstructured":"Ukommi, U. (2022). Review of multimedia communication quality assessment techniques. Nigerian Journal of Technology, 41(2), 330\u2013338.","journal-title":"Nigerian Journal of Technology"},{"issue":"1","key":"1872_CR121","first-page":"1","volume":"45","author":"C Lee","year":"2006","unstructured":"Lee, C., Cho, S., Choe, J., Jeong, T., Ahn, W., & Lee, E. (2006). Objective video quality assessment. Optical Engineering, 45(1), 1\u201311.","journal-title":"Optical Engineering"},{"issue":"3","key":"1872_CR122","doi-asserted-by":"crossref","first-page":"81","DOI":"10.1109\/97.995823","volume":"9","author":"Z Wang","year":"2002","unstructured":"Wang, Z., & Bovik, A. C. (2002). A universal image quality index. IEEE Signal Processing Letters, 9(3), 81\u201384.","journal-title":"IEEE Signal Processing Letters"},{"issue":"1","key":"1872_CR123","doi-asserted-by":"crossref","first-page":"79","DOI":"10.3354\/cr030079","volume":"30","author":"CJ Willmott","year":"2005","unstructured":"Willmott, C. J., & Matsuura, K. (2005). Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Climate Research, 30(1), 79\u201382.","journal-title":"Climate Research"},{"issue":"11","key":"1872_CR124","doi-asserted-by":"crossref","first-page":"2273","DOI":"10.1049\/iet-ipr.2019.1438","volume":"14","author":"K Li","year":"2020","unstructured":"Li, K., Yang, S., Dong, R., Wang, X., & Huang, J. (2020). Survey of single image super-resolution reconstruction. IET Image Processing, 14(11), 2273\u20132290.","journal-title":"IET Image Processing"},{"key":"1872_CR125","unstructured":"Xiao, F. et\u00a0al. (2000). DCT-based video quality evaluation. Final Project for EE392J, 769, 1\u201311."},{"issue":"2","key":"1872_CR126","first-page":"121","volume":"19","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Lu, L., & Bovik, A. C. (2004). Video quality assessment based on structural distortion measurement. Signal Processing: Image Communication, 19(2), 121\u2013132.","journal-title":"Signal Processing: Image Communication"},{"key":"1872_CR127","doi-asserted-by":"crossref","unstructured":"Hore, A., & Ziou, D. (2010). Image quality metrics: PSNR vs. SSIM. In 20th International Conference on Pattern Recognition (pp. 2366\u20132369). IEEE.","DOI":"10.1109\/ICPR.2010.579"},{"issue":"4","key":"1872_CR128","doi-asserted-by":"crossref","first-page":"516","DOI":"10.1109\/TCSVT.2011.2168269","volume":"22","author":"S Wang","year":"2011","unstructured":"Wang, S., Rehman, A., Wang, Z., Ma, S., & Gao, W. (2011). SSIM-motivated rate-distortion optimization for video coding. IEEE Transactions on Circuits and Systems for Video Technology, 22(4), 516\u2013529.","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"issue":"4","key":"1872_CR129","doi-asserted-by":"crossref","first-page":"1258","DOI":"10.1007\/s11263-020-01419-7","volume":"129","author":"K Ding","year":"2021","unstructured":"Ding, K., Ma, K., Wang, S., & Simoncelli, E. P. (2021). Comparison of full-reference image quality models for optimization of image processing systems. International Journal of Computer Vision, 129(4), 1258\u20131281.","journal-title":"International Journal of Computer Vision"},{"key":"1872_CR130","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.\u00a0A., Shechtman, E., & Wang, O. (2018). The unreasonable effectiveness of deep features as a perceptual metric. In IEEE Conference on Computer Vision and Pattern Recognition (pp. 586\u2013595).","DOI":"10.1109\/CVPR.2018.00068"},{"key":"1872_CR131","doi-asserted-by":"crossref","unstructured":"Xia, W., Zhang, Y., Yang, Y., Xue, J.-H., Zhou, B., & Yang, M.-H. (2022). GAN inversion: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence.","DOI":"10.1109\/TPAMI.2022.3181070"},{"key":"1872_CR132","doi-asserted-by":"crossref","unstructured":"Vaccaro, F., Bertini, M., Uricchio, T., & Del\u00a0Bimbo, A. (2021). Fast video visual quality and resolution improvement using SR-UNET. In 29th ACM International Conference on Multimedia (pp. 1221\u20131229).","DOI":"10.1145\/3474085.3475683"},{"key":"1872_CR133","first-page":"1","volume":"30","author":"M Heusel","year":"2017","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., & Hochreiter, S. (2017). Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in Neural Information Processing Systems, 30, 1\u201312.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"1872_CR134","unstructured":"Yu, Y., Zhang, W., & Deng, Y. (2021). Fr\u00e9chet Inception Distance (FID) for Evaluating GANs. Research Gate."},{"key":"1872_CR135","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., & Rabinovich, A. (2015). Going deeper with convolutions. In IEEE Conference on Computer Vision and Pattern Recognition (pp. 1\u20139).","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"1872_CR136","unstructured":"Soloveitchik, M., Diskin, T., Morin, E., & Wiesel, A. (2021). Conditional frechet inception distance. Preprint retrieved from http:\/\/arxiv.org\/abs\/2103.11521"},{"issue":"3s","key":"1872_CR137","first-page":"335","volume":"71","author":"M Anjana","year":"2022","unstructured":"Anjana, M., & Dhanya, N. (2022). Anime face generation using generative adversial networks in deep learning. Mathematical Statistician and Engineering Applications, 71(3s), 335\u2013342.","journal-title":"Mathematical Statistician and Engineering Applications"},{"key":"1872_CR138","doi-asserted-by":"crossref","unstructured":"Mozhaeva, A., Streeter, L., Vlasuyk, I., & Potashnikov, A. (2021). Full reference video quality assessment metric on base human visual system consistent with PSNR. In 28th Conference of Open Innovations Association (pp. 309\u2013315). IEEE.","DOI":"10.23919\/FRUCT50888.2021.9347604"}],"container-title":["Journal of Signal Processing Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11265-023-01872-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11265-023-01872-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11265-023-01872-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,13]],"date-time":"2023-12-13T02:45:02Z","timestamp":1702435502000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11265-023-01872-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,5,16]]},"references-count":138,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2023,6]]}},"alternative-id":["1872"],"URL":"https:\/\/doi.org\/10.1007\/s11265-023-01872-w","relation":{},"ISSN":["1939-8018","1939-8115"],"issn-type":[{"type":"print","value":"1939-8018"},{"type":"electronic","value":"1939-8115"}],"subject":[],"published":{"date-parts":[[2023,5,16]]},"assertion":[{"value":"29 March 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 April 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 May 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 May 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}