{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T05:04:07Z","timestamp":1768280647891,"version":"3.49.0"},"reference-count":76,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Multimedia"],"published-print":{"date-parts":[[2026]]},"DOI":"10.1109\/tmm.2025.3618562","type":"journal-article","created":{"date-parts":[[2025,10,6]],"date-time":"2025-10-06T17:37:36Z","timestamp":1759772256000},"page":"57-70","source":"Crossref","is-referenced-by-count":0,"title":["Quality Evaluation of AI-Generated Images: Subjective Study and Objective Methodology"],"prefix":"10.1109","volume":"28","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-8642-9099","authenticated-orcid":false,"given":"Jianbo","family":"Chen","sequence":"first","affiliation":[{"name":"Faculty of Information Science and Engineering, Ningbo University, Ningbo, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2495-9924","authenticated-orcid":false,"given":"Feng","family":"Shao","sequence":"additional","affiliation":[{"name":"Faculty of Information Science and Engineering, Ningbo University, Ningbo, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3756-2029","authenticated-orcid":false,"given":"Hangwei","family":"Chen","sequence":"additional","affiliation":[{"name":"Faculty of Information Science and Engineering, Ningbo University, Ningbo, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5772-2774","authenticated-orcid":false,"given":"Xuejin","family":"Wang","sequence":"additional","affiliation":[{"name":"Fujian Provincial Key Laboratory of Big Data Mining and Applications, Fujian University of Technology, Fuzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2126-9757","authenticated-orcid":false,"given":"Hui","family":"Guo","sequence":"additional","affiliation":[{"name":"Guangxi Key Laboratory of Machine Vision and Intelligent Control, Wuzhou University, Wuzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6025-9343","authenticated-orcid":false,"given":"Qiuping","family":"Jiang","sequence":"additional","affiliation":[{"name":"Faculty of Information Science and Engineering, Ningbo University, Ningbo, China"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Midjourney","author":"Holz","year":"2023"},{"key":"ref2","article-title":"Sora: A review on background, technology, limitations, and opportunities of large vision models","author":"Liu","year":"2024"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2023.3262180"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00143"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2023.3277758"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.5555\/2969033.2969125"},{"issue":"2","key":"ref7","first-page":"3","article-title":"Hierarchical text-conditional image generation with CLIP latents","volume":"1","author":"Ramesh","year":"2022"},{"key":"ref8","first-page":"19822","article-title":"CogView: Mastering text-to-image generation via transformers","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Ding","year":"2021"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01043"},{"key":"ref12","first-page":"6840","article-title":"Denoising diffusion probabilistic models","volume":"33","author":"Ho","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICMEW59549.2023.00082"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2023.3319020"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-99-9119-8_5"},{"key":"ref16","article-title":"PKU-I2IQA: An image-to-image quality assessment database for AI generated images","author":"Yuan","year":"2023"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ISSI.2018.8538172"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2018.2886771"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3121875"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2023.3284988"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2023.3270621"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2022.3231041"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2024.3481532"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2022.3154588"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2022.3164918"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2025.3539477"},{"key":"ref27","article-title":"PSCR: Patches sampling-based contrastive regression for AIGC image quality assessment","author":"Yuan","year":"2023"},{"key":"ref28","article-title":"TIER: Text-image encoder-based regression for AIGC image quality assessment","author":"Yuan","year":"2024"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TBC.2024.3391060"},{"key":"ref30","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Radford","year":"2021"},{"key":"ref31","article-title":"GLIDE: Towards photorealistic image generation and editing with text-guided diffusion models","author":"Nichol","year":"2021"},{"key":"ref32","article-title":"Text-guided synthesis of artistic images with retrieval-augmented diffusion models","author":"Rombach","year":"2022"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01738"},{"key":"ref34","first-page":"1692","article-title":"One transformer fits all distributions in multi-modal diffusion at scale","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Bao","year":"2023"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"ref36","article-title":"Pixabay","year":"2010"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2012.2227726"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2011.2131660"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2012.2214050"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2023.3327613"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2024.3426529"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2024.3491907"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2024.3521845"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00372"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00510"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW56347.2022.00126"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/WACV51458.2022.00404"},{"key":"ref48","first-page":"2234","article-title":"Improved techniques for training GANs","volume-title":"Proc. 30th Int. Conf. Neural Inf. Process. Syst.","author":"Salimans","year":"2016"},{"key":"ref49","first-page":"6629","article-title":"GANs trained by a two time-scale update rule converge to a local Nash equilibrium","volume-title":"Proc. 31st Int. Conf. Neural Inf. Process. Syst.","author":"Heusel","year":"2017"},{"key":"ref50","article-title":"Demystifying MMD GANs","author":"Bikowski","year":"2018"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/ICME57554.2024.10688254"},{"key":"ref52","article-title":"LLaMA: Open and efficient foundation language models","author":"Touvron","year":"2023"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1145\/3664647.3681471"},{"key":"ref54","article-title":"A-Bench: Are LMMs masters at evaluating AI-generated images?","author":"Zhang","year":"2024"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1991.3.1.79"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2023.3291819"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.279"},{"key":"ref58","article-title":"Microsoft COCO Captions: Data Collection and Evaluation Server","author":"Chen","year":"2015"},{"key":"ref59","article-title":"Q-Align: Teaching LMMs for visual scoring via discrete text-defined levels","author":"Wu","year":"2023"},{"key":"ref60","article-title":"Q-Bench: A Benchmark for general-purpose foundation models on low-level vision","author":"Wu","year":"2023"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72904-1_9"},{"key":"ref62","article-title":"PixArt-$\\alpha$: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis","author":"Chen","year":"2023"},{"key":"ref63","article-title":"DALL.E Mini","author":"Dayma","year":"2021"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19836-6_6"},{"key":"ref65","article-title":"Scaling autoregressive models for content-rich text-to-image generation","author":"Yu","year":"2022"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.51"},{"key":"ref67","volume-title":"Methodology for the Subjective Assessment of the Quality of Television Pictures","year":"2012"},{"key":"ref68","article-title":"AesBench: An Expert Benchmark for Multimodal Large Language Models on Image Aesthetics Perception","author":"Huang","year":"2024"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02408"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2023.3249185"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1145\/1277741.1277808"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1037\/h0070288"},{"key":"ref73","article-title":"Final report from the video quality experts group on the validation of objective models of video quality assessment","author":"Antkowiak","year":"2000"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01352"},{"key":"ref76","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i2.25353"}],"container-title":["IEEE Transactions on Multimedia"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6046\/11342315\/11194189.pdf?arnumber=11194189","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,12]],"date-time":"2026-01-12T22:02:28Z","timestamp":1768255348000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11194189\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"references-count":76,"URL":"https:\/\/doi.org\/10.1109\/tmm.2025.3618562","relation":{},"ISSN":["1520-9210","1941-0077"],"issn-type":[{"value":"1520-9210","type":"print"},{"value":"1941-0077","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]}}}