{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:17:34Z","timestamp":1775578654455,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":34,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819991181","type":"print"},{"value":"9789819991198","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-9119-8_5","type":"book-chapter","created":{"date-parts":[[2024,2,2]],"date-time":"2024-02-02T13:03:04Z","timestamp":1706878984000},"page":"46-57","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":62,"title":["AIGCIQA2023: A Large-Scale Image Quality Assessment Database for\u00a0AI Generated Images: From the\u00a0Perspectives of\u00a0Quality, Authenticity and\u00a0Correspondence"],"prefix":"10.1007","author":[{"given":"Jiarui","family":"Wang","sequence":"first","affiliation":[]},{"given":"Huiyu","family":"Duan","sequence":"additional","affiliation":[]},{"given":"Jing","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Shi","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Xiongkuo","family":"Min","sequence":"additional","affiliation":[]},{"given":"Guangtao","family":"Zhai","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,3]]},"reference":[{"key":"5_CR1","unstructured":"Bao, F., et al.: One transformer fits all distributions in multi-modal diffusion at scale. ArXiv abs\/2303.06555 (2023)"},{"issue":"1","key":"5_CR2","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1109\/TIP.2017.2760518","volume":"27","author":"S Bosse","year":"2017","unstructured":"Bosse, S., Maniry, D., M\u00fcller, K.R., Wiegand, T., Samek, W.: Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans. Image Process. (TIP) 27(1), 206\u2013219 (2017)","journal-title":"IEEE Trans. Image Process. (TIP)"},{"key":"5_CR3","doi-asserted-by":"publisher","first-page":"7206","DOI":"10.1109\/TIP.2022.3220404","volume":"31","author":"H Duan","year":"2022","unstructured":"Duan, H., Min, X., Zhu, Y., Zhai, G., Yang, X., Le Callet, P.: Confusing image quality assessment: toward better augmented reality experience. IEEE Trans. Image Process. (TIP) 31, 7206\u20137221 (2022)","journal-title":"IEEE Trans. Image Process. (TIP)"},{"key":"5_CR4","doi-asserted-by":"crossref","unstructured":"Duan, H., et al.: Develop then rival: A human vision-inspired framework for superimposed image decomposition. IEEE Trans. Multimed. (TMM) (2022)","DOI":"10.1109\/TMM.2022.3172882"},{"key":"5_CR5","doi-asserted-by":"crossref","unstructured":"Duan, H., Shen, W., Min, X., Tu, D., Li, J., Zhai, G.: Saliency in augmented reality. In: Proceedings of the ACM International Conference on Multimedia (ACM MM), pp. 6549\u20136558 (2022)","DOI":"10.1145\/3503161.3547955"},{"key":"5_CR6","unstructured":"Duan, H., et al.: Masked autoencoders as image processors. arXiv preprint arXiv:2303.17316 (2023)"},{"key":"5_CR7","doi-asserted-by":"crossref","unstructured":"Duan, H., Zhai, G., Min, X., Zhu, Y., Fang, Y., Yang, X.: Perceptual quality assessment of omnidirectional images. In: Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS), pp. 1\u20135 (2018)","DOI":"10.1109\/ISCAS.2018.8351786"},{"key":"5_CR8","doi-asserted-by":"crossref","unstructured":"Duan, H., Zhai, G., Yang, X., Li, D., Zhu, W.: Ivqad 2017: An immersive video quality assessment database. In: Proceedings of the IEEE International Conference on Systems, Signals and Image Processing (IWSSIP), pp. 1\u20135. IEEE (2017)","DOI":"10.1109\/IWSSIP.2017.7965610"},{"issue":"11","key":"5_CR9","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., et al.: Generative adversarial networks. Commun. ACM 63(11), 139\u2013144 (2020)","journal-title":"Commun. ACM"},{"key":"5_CR10","unstructured":"Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.C.: Improved training of wasserstein gans. In: Proceedings of the Advances in Neural Information Processing Systems (NeurIPS) 30 (2017)"},{"key":"5_CR11","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"5_CR12","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: Proceedings of the Advances in Neural Information Processing Systems (NeurIPS) 30 (2017)"},{"key":"5_CR13","doi-asserted-by":"crossref","unstructured":"Kang, L., Ye, P., Li, Y., Doermann, D.: Convolutional neural networks for no-reference image quality assessment. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1733\u20131740 (2014)","DOI":"10.1109\/CVPR.2014.224"},{"key":"5_CR14","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013)"},{"key":"5_CR15","unstructured":"Kirstain, Y., Polyak, A., Singer, U., Matiana, S., Penna, J., Levy, O.: Pick-a-pic: an open dataset of user preferences for text-to-image generation. arXiv preprint arXiv:2305.01569 (2023)"},{"issue":"10","key":"5_CR16","doi-asserted-by":"publisher","first-page":"4725","DOI":"10.1109\/TIP.2017.2713945","volume":"26","author":"D Kundu","year":"2017","unstructured":"Kundu, D., Ghadiyaram, D., Bovik, A.C., Evans, B.L.: Large-scale crowdsourced study for tone-mapped hdr pictures. IEEE Trans. Image Process. (TIP) 26(10), 4725\u20134740 (2017)","journal-title":"IEEE Trans. Image Process. (TIP)"},{"key":"5_CR17","unstructured":"Lee, K., et al.: Aligning text-to-image models using human feedback. arXiv preprint arXiv:2302.12192 (2023)"},{"key":"5_CR18","unstructured":"Li, J., Li, D., Xiong, C., Hoi, S.: Blip: bootstrapping language-image pre-training for unified vision-language understanding and generation. In: International Conference on Machine Learning, pp. 12888\u201312900. PMLR (2022)"},{"issue":"8","key":"5_CR19","doi-asserted-by":"publisher","first-page":"2049","DOI":"10.1109\/TMM.2017.2788206","volume":"20","author":"X Min","year":"2017","unstructured":"Min, X., Gu, K., Zhai, G., Liu, J., Yang, X., Chen, C.W.: Blind quality assessment based on pseudo-reference image. IEEE Trans. Multimed. (TMM) 20(8), 2049\u20132062 (2017)","journal-title":"IEEE Trans. Multimed. (TMM)"},{"issue":"11","key":"5_CR20","doi-asserted-by":"publisher","first-page":"5462","DOI":"10.1109\/TIP.2017.2735192","volume":"26","author":"X Min","year":"2017","unstructured":"Min, X., Ma, K., Gu, K., Zhai, G., Wang, Z., Lin, W.: Unified blind quality assessment of compressed natural, graphic, and screen content images. IEEE Trans. Image Process. (TIP) 26(11), 5462\u20135474 (2017)","journal-title":"IEEE Trans. Image Process. (TIP)"},{"issue":"2","key":"5_CR21","doi-asserted-by":"publisher","first-page":"508","DOI":"10.1109\/TBC.2018.2816783","volume":"64","author":"X Min","year":"2018","unstructured":"Min, X., Zhai, G., Gu, K., Liu, Y., Yang, X.: Blind image quality estimation via distortion aggravation. IEEE Trans. Broadcast. 64(2), 508\u2013517 (2018)","journal-title":"IEEE Trans. Broadcast."},{"issue":"12","key":"5_CR22","doi-asserted-by":"publisher","first-page":"4695","DOI":"10.1109\/TIP.2012.2214050","volume":"21","author":"A Mittal","year":"2012","unstructured":"Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. (TIP) 21(12), 4695\u20134708 (2012)","journal-title":"IEEE Trans. Image Process. (TIP)"},{"key":"5_CR23","doi-asserted-by":"crossref","unstructured":"Mittal, A., Soundararajan, R., Bovik, A.C.: Making a \u201ccompletely blind\u201d image quality analyzer. IEEE Signal Process. Lett. 20(3), 209\u2013212 (2012)","DOI":"10.1109\/LSP.2012.2227726"},{"key":"5_CR24","unstructured":"Nichol, A., et al.: Glide: towards photorealistic image generation and editing with text-guided diffusion models, pp. 16784\u201316804 (2021)"},{"key":"5_CR25","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"5_CR26","unstructured":"Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., Chen, M.: Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 (2022)"},{"key":"5_CR27","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10674\u201310685 (2021)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"5_CR28","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"5_CR29","unstructured":"Xu, J., et al.: Imagereward: Learning and evaluating human preferences for text-to-image generation. arXiv preprint arXiv:2304.05977 (2023)"},{"issue":"9","key":"5_CR30","doi-asserted-by":"publisher","first-page":"4444","DOI":"10.1109\/TIP.2016.2585880","volume":"25","author":"J Xu","year":"2016","unstructured":"Xu, J., Ye, P., Li, Q., Du, H., Liu, Y., Doermann, D.: Blind image quality assessment based on high order statistics aggregation. IEEE Trans. Image Process. (TIP) 25(9), 4444\u20134457 (2016)","journal-title":"IEEE Trans. Image Process. (TIP)"},{"key":"5_CR31","doi-asserted-by":"crossref","unstructured":"Xue, W., Zhang, L., Mou, X.: Learning without human scores for blind image quality assessment. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 995\u20131002 (2013)","DOI":"10.1109\/CVPR.2013.133"},{"key":"5_CR32","unstructured":"Yu, J., et al.: Scaling autoregressive models for content-rich text-to-image generation. arXiv preprint arXiv:2206.10789 (2022)"},{"key":"5_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, L., Agrawala, M.: Adding conditional control to text-to-image diffusion models. ArXiv abs\/2302.05543 (2023)","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"5_CR34","doi-asserted-by":"crossref","unstructured":"Zhou, Y., et al.: Towards language-free training for text-to-image generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 17907\u201317917, June 2022","DOI":"10.1109\/CVPR52688.2022.01738"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-9119-8_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,2]],"date-time":"2024-02-02T13:04:35Z","timestamp":1706879075000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-9119-8_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819991181","9789819991198"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-9119-8_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"3 February 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CAAI International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Fuzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 July 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 July 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cicai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/cicai.caai.cn\/#\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"376","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"101","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"16","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.9","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1.9","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}