{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:17:32Z","timestamp":1775578652290,"version":"3.50.1"},"reference-count":113,"publisher":"MDPI AG","issue":"5","license":[{"start":{"date-parts":[[2025,5,16]],"date-time":"2025-05-16T00:00:00Z","timestamp":1747353600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["BDCC"],"abstract":"<jats:p>This study presents a comparative analysis of several multimodal large language models (LLMs) for no-reference image quality assessment, with a particular focus on images containing authentic distortions. We evaluate three models developed by OpenAI and three models from Claude.AI, comparing their performance in estimating image quality without reference images. Our results demonstrate that these LLMs outperform traditional methods based on hand-crafted features. However, more advanced deep learning models, especially those based on deep convolutional networks, surpass LLMs in performance. Notably, we make a unique contribution by publishing the processed outputs of the LLMs, providing a transparent and direct comparison of their quality assessments based solely on the predicted quality scores. This work underscores the potential of multimodal LLMs in image quality evaluation, while also highlighting the continuing advantages of specialized deep learning approaches.<\/jats:p>","DOI":"10.3390\/bdcc9050132","type":"journal-article","created":{"date-parts":[[2025,5,16]],"date-time":"2025-05-16T05:11:57Z","timestamp":1747372317000},"page":"132","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Comparative Evaluation of Multimodal Large Language Models for No-Reference Image Quality Assessment with Authentic Distortions: A Study of OpenAI and Claude.AI Models"],"prefix":"10.3390","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3265-5047","authenticated-orcid":false,"given":"Domonkos","family":"Varga","sequence":"first","affiliation":[{"name":"Nokia Bell Labs, 1082 Budapest, Hungary"}]}],"member":"1968","published-online":{"date-parts":[[2025,5,16]]},"reference":[{"key":"ref_1","unstructured":"Lin, H., Hosu, V., and Saupe, D. (2018). KonIQ-10K: Towards an ecologically valid and large-scale IQA database. arXiv."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"72139","DOI":"10.1109\/ACCESS.2021.3077642","article-title":"KonVid-150k: A dataset for no-reference video quality assessment of videos in-the-wild","volume":"9","author":"Hosu","year":"2021","journal-title":"IEEE Access"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Yang, P., Sturtz, J., and Qingge, L. (2023). Progress in blind image quality assessment: A brief review. Mathematics, 11.","DOI":"10.3390\/math11122766"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Oura, D., Sato, S., Honma, Y., Kuwajima, S., and Sugimori, H. (2023). Quality assurance of chest X-ray images with a combination of deep learning methods. Appl. Sci., 13.","DOI":"10.3390\/app13042067"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Nam, W., Youn, T., and Ha, C. (2025). No-Reference Image Quality Assessment with Moving Spectrum and Laplacian Filter for Autonomous Driving Environment. Vehicles, 7.","DOI":"10.3390\/vehicles7010008"},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Hao, Y., Pei, H., Lyu, Y., Yuan, Z., Rizzo, J.R., Wang, Y., and Fang, Y. (2023, January 1\u20135). Understanding the impact of image quality and distance of objects to object detection performance. Proceedings of the 2023 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), Detroit, MI, USA.","DOI":"10.1109\/IROS55552.2023.10342139"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Dodge, S., and Karam, L. (2016, January 6\u20138). Understanding how image quality affects deep neural networks. Proceedings of the 2016 Eighth International Conference on Quality of Multimedia Experience (QoMEX), Lisbon, Portugal.","DOI":"10.1109\/QoMEX.2016.7498955"},{"key":"ref_8","first-page":"105762N","article-title":"Image quality and segmentation","volume":"10576","author":"Pednekar","year":"2018","journal-title":"Proc. SPIE Int. Soc. Opt. Eng."},{"key":"ref_9","unstructured":"Chiasserini, C.F., and Magli, E. (2002, January 15\u201318). Energy consumption and image quality in wireless video-surveillance networks. Proceedings of the 13th IEEE International Symposium on Personal, Indoor and Mobile Radio Communications, Lisboa, Portugal."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Winkler, S., and Campos, R. (2003, January 20). Video quality evaluation for Internet streaming applications. Proceedings of the Human Vision and Electronic Imaging VIII, Santa Clara, CA, USA.","DOI":"10.1117\/12.477766"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"034502","DOI":"10.1088\/1538-3873\/adb790","article-title":"Astronomical Image Quality Assessment Based on Deep Learning for Resource-constrained Environments","volume":"137","author":"Li","year":"2025","journal-title":"Publ. Astron. Soc. Pac."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"262","DOI":"10.1016\/j.procir.2021.10.042","article-title":"Image based quality inspection in smart manufacturing systems: A literature review","volume":"103","author":"Babic","year":"2021","journal-title":"Procedia CIRP"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Li, S., Yang, Z., and Li, H. (2017). Statistical evaluation of no-reference image quality assessment metrics for remote sensing images. ISPRS Int. J. Geo-Inf., 6.","DOI":"10.3390\/ijgi6050133"},{"key":"ref_14","first-page":"1","article-title":"Palm: Scaling language modeling with pathways","volume":"24","author":"Chowdhery","year":"2023","journal-title":"J. Mach. Learn. Res."},{"key":"ref_15","first-page":"25278","article-title":"Laion-5b: An open large-scale dataset for training next generation image-text models","volume":"35","author":"Schuhmann","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_16","unstructured":"Chen, X., Wang, X., Changpinyo, S., Piergiovanni, A., Padlewski, P., Salz, D., Goodman, S., Grycner, A., Mustafa, B., and Beyer, L. (2022). Pali: A jointly-scaled multilingual language-image model. arXiv."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Sharma, P., Ding, N., Goodman, S., and Soricut, R. (2018, January 15\u201320). Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning. Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia.","DOI":"10.18653\/v1\/P18-1238"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"4695","DOI":"10.1109\/TIP.2012.2214050","article-title":"No-reference image quality assessment in the spatial domain","volume":"21","author":"Mittal","year":"2012","journal-title":"IEEE Trans. Image Process."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Galdran, A., Ara\u00fajo, T., Mendon\u00e7a, A.M., and Campilho, A. (2017, January 18\u201320). Retinal image quality assessment by mean-subtracted contrast-normalized coefficients. Proceedings of the VipIMAGE 2017: Proceedings of the VI ECCOMAS Thematic Conference on Computational Vision and Medical Image Processing, Porto, Portugal.","DOI":"10.1007\/978-3-319-68195-5_92"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"209","DOI":"10.1109\/LSP.2012.2227726","article-title":"Making a \u201ccompletely blind\u201d image quality analyzer","volume":"20","author":"Mittal","year":"2012","journal-title":"IEEE Signal Process. Lett."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"3998","DOI":"10.1109\/TIP.2018.2831899","article-title":"NIMA: Neural image assessment","volume":"27","author":"Talebi","year":"2018","journal-title":"IEEE Trans. Image Process."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Murray, N., Marchesotti, L., and Perronnin, F. (2012, January 16\u201321). AVA: A large-scale database for aesthetic visual analysis. Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA.","DOI":"10.1109\/CVPR.2012.6247954"},{"key":"ref_23","unstructured":"Levina, E., and Bickel, P. (2001, January 7\u201314). The earth mover\u2019s distance is the mallows distance: Some insights from statistics. Proceedings of the Eighth IEEE International Conference on Computer Vision, Vancouver, BC, Canada."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Ke, J., Wang, Q., Wang, Y., Milanfar, P., and Yang, F. (2021, January 11\u201317). Musiq: Multi-scale image quality transformer. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Virtual Event.","DOI":"10.1109\/ICCV48922.2021.00510"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Su, S., Yan, Q., Zhu, Y., Zhang, C., Ge, X., Sun, J., and Zhang, Y. (2020, January 14\u201319). Blindly assess image quality in the wild guided by a self-adaptive hyper network. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00372"},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"3440","DOI":"10.1109\/TIP.2006.881959","article-title":"A statistical evaluation of recent full reference image quality assessment algorithms","volume":"15","author":"Sheikh","year":"2006","journal-title":"IEEE Trans. Image Process."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Ponomarenko, N., Ieremeiev, O., Lukin, V., Jin, L., Egiazarian, K., Astola, J., Vozel, B., Chehdi, K., Carli, M., and Battisti, F. (2013, January 28\u201331). A new color image database TID2013: Innovations and results. Proceedings of the Advanced Concepts for Intelligent Vision Systems: 15th International Conference, ACIVS 2013, Pozna\u0144, Poland.","DOI":"10.1007\/978-3-319-02895-8_36"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Lin, H., Hosu, V., and Saupe, D. (2019, January 5\u20137). KADID-10k: A large-scale artificially distorted IQA database. Proceedings of the 2019 Eleventh International Conference on Quality of Multimedia Experience (QoMEX), Berlin, Germany.","DOI":"10.1109\/QoMEX.2019.8743252"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Men, H., Lin, H., and Saupe, D. (June, January 31). Empirical evaluation of no-reference VQA methods on a natural video quality database. Proceedings of the 2017 Ninth International Conference on Quality of Multimedia Experience (QoMEX), Erfurt, Germany.","DOI":"10.1109\/QoMEX.2017.7965644"},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"138939","DOI":"10.1109\/ACCESS.2021.3118295","article-title":"Subjective image quality assessment with boosted triplet comparisons","volume":"9","author":"Men","year":"2021","journal-title":"IEEE Access"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Lin, H., Men, H., Yan, Y., Ren, J., and Saupe, D. (2022, January 5\u20137). Crowdsourced quality assessment of enhanced underwater images\u2014A pilot study. Proceedings of the 2022 14th International Conference on Quality of Multimedia Experience (QoMEX), Lippstadt, Germany.","DOI":"10.1109\/QoMEX55416.2022.9900904"},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"2671","DOI":"10.1109\/TMM.2023.3301276","article-title":"Going the extra mile in face image quality assessment: A novel database and model","volume":"26","author":"Su","year":"2023","journal-title":"IEEE Trans. Multimed."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"4408","DOI":"10.1109\/TIP.2015.2465145","article-title":"Perceptual quality assessment of screen content images","volume":"24","author":"Yang","year":"2015","journal-title":"IEEE Trans. Image Process."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"372","DOI":"10.1109\/TIP.2015.2500021","article-title":"Massive online crowdsourced study of subjective and objective picture quality","volume":"25","author":"Ghadiyaram","year":"2015","journal-title":"IEEE Trans. Image Process."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"264","DOI":"10.1134\/S1062739121020113","article-title":"Investigation of the Relationship between Speed and Image Quality of Autonomous Vehicles","volume":"57","author":"Xin","year":"2021","journal-title":"J. Min. Sci."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"1332","DOI":"10.1109\/TCSVT.2020.3002662","article-title":"Domain fingerprints for no-reference image quality assessment","volume":"31","author":"Xia","year":"2020","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"111","DOI":"10.1016\/j.neucom.2018.06.042","article-title":"No-reference image quality assessment for photographic images based on robust statistics","volume":"313","author":"Zeng","year":"2018","journal-title":"Neurocomputing"},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"3318","DOI":"10.1049\/ipr2.12328","article-title":"No-reference image quality assessment based on multiscale feature representation","volume":"15","author":"Li","year":"2021","journal-title":"IET Image Process."},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"1025","DOI":"10.1016\/j.neuron.2013.06.034","article-title":"Natural scene statistics account for the representation of scene categories in human visual cortex","volume":"79","author":"Stansbury","year":"2013","journal-title":"Neuron"},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Zhu, K., Asari, V., and Saupe, D. (May, January 29). No-reference quality assessment of H. 264\/AVC encoded video based on natural scene features. Proceedings of the Mobile Multimedia\/Image Processing, Security, and Applications 2013, Baltimore, MD, USA.","DOI":"10.1117\/12.2015594"},{"key":"ref_41","unstructured":"Sheikh, H.R., Bovik, A.C., and Cormack, L. (2003, January 9\u201312). Blind quality assessment of JPEG2000 compressed images using natural scene statistics. Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA."},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"513","DOI":"10.1109\/LSP.2010.2043888","article-title":"A two-step framework for constructing blind image quality indices","volume":"17","author":"Moorthy","year":"2010","journal-title":"IEEE Signal Process. Lett."},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Daubechies, I. (1992). Ten Lectures on Wavelets, Society for Industrial and Applied Mathematics.","DOI":"10.1137\/1.9781611970104"},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Awad, M., Khanna, R., Awad, M., and Khanna, R. (2015). Support vector regression. Efficient Learning Machines: Theories, Concepts, and Applications for Engineers and System Designers, Springer.","DOI":"10.1007\/978-1-4302-5990-9"},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"3350","DOI":"10.1109\/TIP.2011.2147325","article-title":"Blind image quality assessment: From natural scene statistics to perceptual quality","volume":"20","author":"Moorthy","year":"2011","journal-title":"IEEE Trans. Image Process."},{"key":"ref_46","first-page":"855","article-title":"Scale mixtures of Gaussians and the statistics of natural images","volume":"12","author":"Wainwright","year":"1999","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_47","doi-asserted-by":"crossref","unstructured":"Saad, M.A., and Bovik, A.C. (2012, January 4\u20137). Blind quality assessment of videos using a model of natural scene statistics and motion coherency. Proceedings of the 2012 Conference Record of the Forty Sixth Asiloma Pacific, Grove, CA, USA.","DOI":"10.1109\/ACSSC.2012.6489018"},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"583","DOI":"10.1109\/LSP.2010.2045550","article-title":"A DCT statistics-based blind image quality index","volume":"17","author":"Saad","year":"2010","journal-title":"IEEE Signal Process. Lett."},{"key":"ref_49","doi-asserted-by":"crossref","first-page":"B42","DOI":"10.1364\/JOSAA.24.000B42","article-title":"Blind image quality assessment through anisotropy","volume":"24","author":"Gabarda","year":"2007","journal-title":"J. Opt. Soc. Am. A"},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Lasmar, N.E., Stitou, Y., and Berthoumieu, Y. (2009, January 7\u201310). Multiscale skewed heavy tailed model for texture analysis. Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt.","DOI":"10.1109\/ICIP.2009.5414404"},{"key":"ref_51","doi-asserted-by":"crossref","first-page":"814","DOI":"10.1103\/PhysRevLett.73.814","article-title":"Statistics of natural images: Scaling in the woods","volume":"73","author":"Ruderman","year":"1994","journal-title":"Phys. Rev. Lett."},{"key":"ref_52","doi-asserted-by":"crossref","first-page":"8975","DOI":"10.1109\/ACCESS.2018.2890743","article-title":"Texture feature extraction methods: A survey","volume":"7","year":"2019","journal-title":"IEEE Access"},{"key":"ref_53","doi-asserted-by":"crossref","first-page":"4850","DOI":"10.1109\/TIP.2014.2355716","article-title":"Blind image quality assessment using joint statistics of gradient magnitude and Laplacian features","volume":"23","author":"Xue","year":"2014","journal-title":"IEEE Trans. Image Process."},{"key":"ref_54","first-page":"1","article-title":"Face recognition using local binary patterns (LBP)","volume":"13","author":"Rahim","year":"2013","journal-title":"Glob. J. Comput. Sci. Technol."},{"key":"ref_55","doi-asserted-by":"crossref","first-page":"730","DOI":"10.3724\/SP.J.1004.2013.00730","article-title":"Research and perspective on local binary pattern","volume":"39","author":"Song","year":"2013","journal-title":"Acta Autom. Sin."},{"key":"ref_56","doi-asserted-by":"crossref","unstructured":"Garcia Freitas, P., Da Eira, L.P., Santos, S.S., and Farias, M.C.Q.d. (2018). On the application LBP texture descriptors and its variants for no-reference image quality assessment. J. Imaging, 4.","DOI":"10.3390\/jimaging4100114"},{"key":"ref_57","doi-asserted-by":"crossref","first-page":"2457","DOI":"10.1109\/TMM.2016.2601028","article-title":"Blind image quality assessment using statistical structural and luminance features","volume":"18","author":"Li","year":"2016","journal-title":"IEEE Trans. Multimed."},{"key":"ref_58","first-page":"101039","article-title":"An efficient approach for no-reference image quality assessment based on statistical texture and structural features","volume":"30","author":"Rajevenceltha","year":"2022","journal-title":"Eng. Sci. Technol. Int. J."},{"key":"ref_59","doi-asserted-by":"crossref","first-page":"206","DOI":"10.1109\/TIP.2017.2760518","article-title":"Deep neural networks for no-reference and full-reference image quality assessment","volume":"27","author":"Bosse","year":"2017","journal-title":"IEEE Trans. Image Process."},{"key":"ref_60","doi-asserted-by":"crossref","first-page":"146893","DOI":"10.1109\/ACCESS.2019.2942625","article-title":"No-reference image quality assessment based on multi-task generative adversarial network","volume":"7","author":"Ma","year":"2019","journal-title":"IEEE Access"},{"key":"ref_61","doi-asserted-by":"crossref","unstructured":"Kang, L., Ye, P., Li, Y., and Doermann, D. (2014, January 23\u201328). Convolutional neural networks for no-reference image quality assessment. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA.","DOI":"10.1109\/CVPR.2014.224"},{"key":"ref_62","doi-asserted-by":"crossref","unstructured":"Bare, B., Li, K., and Yan, B. (2017, January 10\u201314). An accurate deep convolutional neural networks model for no-reference image quality assessment. Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China.","DOI":"10.1109\/ICME.2017.8019508"},{"key":"ref_63","doi-asserted-by":"crossref","first-page":"2378","DOI":"10.1109\/TIP.2011.2109730","article-title":"FSIM: A feature similarity index for image quality assessment","volume":"20","author":"Zhang","year":"2011","journal-title":"IEEE Trans. Image Process."},{"key":"ref_64","doi-asserted-by":"crossref","unstructured":"Li, Y., Po, L.M., Feng, L., and Yuan, F. (2016, January 16\u201318). No-reference image quality assessment with deep convolutional neural networks. Proceedings of the 2016 IEEE International Conference on Digital Signal Processing (DSP), Beijing, China.","DOI":"10.1109\/ICDSP.2016.7868646"},{"key":"ref_65","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","article-title":"ImageNet classification with deep convolutional neural networks","volume":"60","author":"Krizhevsky","year":"2017","journal-title":"Commun. ACM"},{"key":"ref_66","unstructured":"Lin, M. (2013). Network in network. arXiv."},{"key":"ref_67","doi-asserted-by":"crossref","first-page":"355","DOI":"10.1007\/s11760-017-1166-8","article-title":"On the use of deep learning for blind image quality assessment","volume":"12","author":"Bianco","year":"2018","journal-title":"Signal Image Video Process."},{"key":"ref_68","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., and Wang, O. (2018, January 18\u201322). The unreasonable effectiveness of deep features as a perceptual metric. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00068"},{"key":"ref_69","doi-asserted-by":"crossref","unstructured":"Ryu, J. (2023). Improved image quality assessment by utilizing pre-trained architecture features with unified learning mechanism. Appl. Sci., 13.","DOI":"10.3390\/app13042682"},{"key":"ref_70","doi-asserted-by":"crossref","first-page":"432","DOI":"10.1016\/j.patcog.2018.04.016","article-title":"Blind image quality prediction by exploiting multi-level deep representations","volume":"81","author":"Gao","year":"2018","journal-title":"Pattern Recognit."},{"key":"ref_71","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2020). An image is worth 16 \u00d7 16 words: Transformers for image recognition at scale. arXiv."},{"key":"ref_72","doi-asserted-by":"crossref","unstructured":"Shaw, P., Uszkoreit, J., and Vaswani, A. (2018). Self-attention with relative position representations. arXiv.","DOI":"10.18653\/v1\/N18-2074"},{"key":"ref_73","unstructured":"Keshari, A., and Subudhi, B. (2022). Multi-scale features and parallel transformers based image quality assessment. arXiv."},{"key":"ref_74","doi-asserted-by":"crossref","unstructured":"Yang, S., Wu, T., Shi, S., Lao, S., Gong, Y., Cao, M., Wang, J., and Yang, Y. (2022, January 18\u201324). Maniqa: Multi-dimension attention network for no-reference image quality assessment. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPRW56347.2022.00126"},{"key":"ref_75","unstructured":"Wang, J., Chan, K.C., and Loy, C.C. (2023, January 7\u201314). Exploring clip for assessing the look and feel of images. Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA."},{"key":"ref_76","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., and Clark, J. (2021, January 18\u201324). Learning transferable visual models from natural language supervision. Proceedings of the International Conference on Machine Learning, Virtual."},{"key":"ref_77","doi-asserted-by":"crossref","first-page":"141","DOI":"10.1016\/j.ins.2014.12.055","article-title":"Non-distortion-specific no-reference image quality assessment: A survey","volume":"301","author":"Manap","year":"2015","journal-title":"Inf. Sci."},{"key":"ref_78","doi-asserted-by":"crossref","first-page":"223","DOI":"10.1080\/02564602.2016.1151385","article-title":"No-reference\/blind image quality assessment: A survey","volume":"34","author":"Xu","year":"2017","journal-title":"IETE Tech. Rev."},{"key":"ref_79","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s11432-019-2757-1","article-title":"Perceptual image quality assessment: A survey","volume":"63","author":"Zhai","year":"2020","journal-title":"Sci. China Inf. Sci."},{"key":"ref_80","doi-asserted-by":"crossref","first-page":"123788","DOI":"10.1109\/ACCESS.2019.2938900","article-title":"A survey of DNN methods for blind image quality assessment","volume":"7","author":"Yang","year":"2019","journal-title":"IEEE Access"},{"key":"ref_81","doi-asserted-by":"crossref","unstructured":"Xu, L., Lin, W., and Kuo, C.C.J. (2015). Visual Quality Assessment by Machine Learning, Springer.","DOI":"10.1007\/978-981-287-468-9"},{"key":"ref_82","unstructured":"Jenadeleh, M. (2018). Blind Image and Video Quality Assessment. [Ph.D. Dissertation, Universit\u00e4t Konstanz]."},{"key":"ref_83","unstructured":"Men, H. (2022). Boosting for Visual Quality Assessment with Applications for Frame Interpolation Methods. [Ph.D. Dissertation, Universit\u00e4t Konstanz]."},{"key":"ref_84","doi-asserted-by":"crossref","first-page":"64","DOI":"10.1145\/2812802","article-title":"Yfcc100m: The new data in multimedia research","volume":"59","author":"Thomee","year":"2016","journal-title":"Commun. ACM"},{"key":"ref_85","unstructured":"Saupe, D., Hahn, F., Hosu, V., Zingman, I., Rana, M., and Li, S. (2016, January 6\u20138). Crowd workers proven useful: A comparative study of subjective video quality assessment. Proceedings of the QoMEX 2016: 8th International Conference on Quality of Multimedia Experience, Lisbon, Portugal."},{"key":"ref_86","doi-asserted-by":"crossref","unstructured":"Shahriar, S., Lund, B.D., Mannuru, N.R., Arshad, M.A., Hayawi, K., Bevara, R.V.K., Mannuru, A., and Batool, L. (2024). Putting gpt-4o to the sword: A comprehensive evaluation of language, vision, speech, and multimodal proficiency. Appl. Sci., 14.","DOI":"10.20944\/preprints202406.1635.v1"},{"key":"ref_87","doi-asserted-by":"crossref","unstructured":"Islam, R., and Moushi, O.M. (2024). Gpt-4o: The cutting-edge advancement in multimodal llm. Authorea Prepr.","DOI":"10.36227\/techrxiv.171986596.65533294\/v1"},{"key":"ref_88","unstructured":"Priyanshu, A., Maurya, Y., and Hong, Z. (2024). AI Governance and Accountability: An Analysis of Anthropic\u2019s Claude. arXiv."},{"key":"ref_89","doi-asserted-by":"crossref","first-page":"1132","DOI":"10.1038\/s41433-024-03545-9","article-title":"Benchmarking the performance of large language models in uveitis: A comparative analysis of ChatGPT-3.5, ChatGPT-4.0, Google Gemini, and Anthropic Claude3","volume":"39","author":"Zhao","year":"2024","journal-title":"Eye"},{"key":"ref_90","first-page":"34892","article-title":"Visual instruction tuning","volume":"36","author":"Liu","year":"2023","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_91","doi-asserted-by":"crossref","first-page":"588","DOI":"10.1109\/LSP.2023.3276645","article-title":"MAMIQA: No-Reference Image Quality Assessment Based on Multiscale Attention Mechanism With Natural Scene Statistics","volume":"30","author":"Yu","year":"2023","journal-title":"IEEE Signal Process. Lett."},{"key":"ref_92","doi-asserted-by":"crossref","first-page":"508","DOI":"10.1109\/TBC.2018.2816783","article-title":"Blind image quality estimation via distortion aggravation","volume":"64","author":"Min","year":"2018","journal-title":"IEEE Trans. Broadcast."},{"key":"ref_93","doi-asserted-by":"crossref","first-page":"494","DOI":"10.1016\/j.image.2014.02.004","article-title":"No-reference image quality assessment in curvelet domain","volume":"29","author":"Liu","year":"2014","journal-title":"Signal Process. Image Commun."},{"key":"ref_94","doi-asserted-by":"crossref","first-page":"77","DOI":"10.1186\/s13640-019-0479-7","article-title":"No-reference color image quality assessment: From entropy to perceptual quality","volume":"2019","author":"Chen","year":"2019","journal-title":"EURASIP J. Image Video Process."},{"key":"ref_95","doi-asserted-by":"crossref","first-page":"541","DOI":"10.1109\/LSP.2016.2537321","article-title":"No-reference quality assessment for multiply-distorted images in gradient domain","volume":"23","author":"Li","year":"2016","journal-title":"IEEE Signal Process. Lett."},{"key":"ref_96","doi-asserted-by":"crossref","first-page":"2579","DOI":"10.1109\/TIP.2015.2426416","article-title":"A feature-enriched completely blind image quality evaluator","volume":"24","author":"Zhang","year":"2015","journal-title":"IEEE Trans. Image Process."},{"key":"ref_97","doi-asserted-by":"crossref","unstructured":"Ou, F.Z., Wang, Y.G., and Zhu, G. (2019, January 22\u201325). A novel blind image quality assessment method based on refined natural scene statistics. Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan.","DOI":"10.1109\/ICIP.2019.8803047"},{"key":"ref_98","doi-asserted-by":"crossref","unstructured":"Venkatanath, N., Praneeth, D., Bh, M.C., Channappayya, S.S., and Medasani, S.S. (March, January 27). Blind image quality evaluation using perception based features. Proceedings of the 2015 Twenty First National Conference on Communications (NCC), Mumbai, India.","DOI":"10.1109\/NCC.2015.7084843"},{"key":"ref_99","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.image.2015.10.005","article-title":"Blind image quality assessment by relative gradient statistics and adaboosting neural network","volume":"40","author":"Liu","year":"2016","journal-title":"Signal Process. Image Commun."},{"key":"ref_100","doi-asserted-by":"crossref","unstructured":"Mittal, A., Moorthy, A.K., and Bovik, A.C. (2012, January 4\u20137). Making image quality assessment robust. Proceedings of the 2012 Conference Record of the Forty Sixth Asilomar Conference on Signals, Systems and Computers (ASILOMAR), Pacific Grove, CA, USA.","DOI":"10.1109\/ACSSC.2012.6489326"},{"key":"ref_101","doi-asserted-by":"crossref","first-page":"856","DOI":"10.1016\/j.image.2014.06.006","article-title":"No-reference image quality assessment based on spatial and spectral entropies","volume":"29","author":"Liu","year":"2014","journal-title":"Signal Process. Image Commun."},{"key":"ref_102","doi-asserted-by":"crossref","first-page":"3474","DOI":"10.1109\/TIP.2021.3061932","article-title":"Uncertainty-aware blind image quality assessment in the laboratory and wild","volume":"30","author":"Zhang","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"ref_103","doi-asserted-by":"crossref","unstructured":"Madhusudana, P.C., Birkbeck, N., Wang, Y., Adsumilli, B., and Bovik, A.C. (2021). Image Quality Assessment using Contrastive Learning. arXiv.","DOI":"10.1109\/TIP.2022.3181496"},{"key":"ref_104","doi-asserted-by":"crossref","first-page":"36","DOI":"10.1109\/TCSVT.2018.2886771","article-title":"Blind image quality assessment using a deep bilinear convolutional neural network","volume":"30","author":"Zhang","year":"2018","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"ref_105","unstructured":"Lin, H., Hosu, V., and Saupe, D. (2020). DeepFL-IQA: Weak supervision for deep IQA feature learning. arXiv."},{"key":"ref_106","doi-asserted-by":"crossref","first-page":"4041","DOI":"10.1109\/TIP.2020.2967829","article-title":"KonIQ-10k: An ecologically valid database for deep learning of blind image quality assessment","volume":"29","author":"Hosu","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"ref_107","unstructured":"Su, S., Hosu, V., Lin, H., Zhang, Y., and Saupe, D. (2021, January 22\u201325). KonIQ++: Boosting No-Reference Image Quality Assessment in the Wild by Jointly Predicting Image Quality and Defects. Proceedings of the 32nd British Machine Vision Conference, Virtual."},{"key":"ref_108","doi-asserted-by":"crossref","unstructured":"Hosu, V., Goldlucke, B., and Saupe, D. (2019, January 15\u201320). Effective aesthetics prediction with multi-level spatially pooled features. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00960"},{"key":"ref_109","doi-asserted-by":"crossref","unstructured":"Ying, Z., Niu, H., Gupta, P., Mahajan, D., Ghadiyaram, D., and Bovik, A. (2020, January 14\u201319). From patches to pictures (PaQ-2-PiQ): Mapping the perceptual space of picture quality. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00363"},{"key":"ref_110","unstructured":"Zeng, H., Zhang, L., and Bovik, A.C. (2017). A probabilistic quality representation approach to deep blind image quality prediction. arXiv."},{"key":"ref_111","doi-asserted-by":"crossref","first-page":"1613","DOI":"10.1109\/TIP.2022.3144892","article-title":"VCRNet: Visual compensation restoration network for no-reference image quality assessment","volume":"31","author":"Pan","year":"2022","journal-title":"IEEE Trans. Image Process."},{"key":"ref_112","doi-asserted-by":"crossref","unstructured":"Su, Y., and Korhonen, J. (2020, January 25\u201328). Blind Natural Image Quality Prediction Using Convolutional Neural Networks And Weighted Spatial Pooling. Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates.","DOI":"10.1109\/ICIP40778.2020.9190789"},{"key":"ref_113","doi-asserted-by":"crossref","first-page":"70973","DOI":"10.1109\/ACCESS.2024.3402729","article-title":"ZEN-IQA: Zero-Shot Explainable and No-Reference Image Quality Assessment With Vision Language Model","volume":"12","author":"Miyata","year":"2024","journal-title":"IEEE Access"}],"container-title":["Big Data and Cognitive Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2504-2289\/9\/5\/132\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T17:33:44Z","timestamp":1760031224000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2504-2289\/9\/5\/132"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,16]]},"references-count":113,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2025,5]]}},"alternative-id":["bdcc9050132"],"URL":"https:\/\/doi.org\/10.3390\/bdcc9050132","relation":{},"ISSN":["2504-2289"],"issn-type":[{"value":"2504-2289","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,16]]}}}