{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T21:01:54Z","timestamp":1743109314986,"version":"3.40.3"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030893699"},{"type":"electronic","value":"9783030893705"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-89370-5_26","type":"book-chapter","created":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T01:02:59Z","timestamp":1635728579000},"page":"353-365","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["When Distortion Meets Perceptual Quality: A Multi-task Learning Pipeline"],"prefix":"10.1007","author":[{"given":"Jing","family":"Wen","sequence":"first","affiliation":[]},{"given":"Qianyu","family":"Guo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,11,1]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Ahn, N., Kang, B., Sohn, K.A.: Image distortion detection using convolutional neural network (2018)","key":"26_CR1","DOI":"10.1109\/ACPR.2017.95"},{"issue":"18","key":"26_CR2","doi-asserted-by":"publisher","first-page":"25611","DOI":"10.1007\/s11042-019-7668-3","volume":"78","author":"H Al-Bandawi","year":"2019","unstructured":"Al-Bandawi, H., Deng, G.: Classification of image distortion based on the generalized Benford\u2019s law. Multimedia Tools Appl. 78(18), 25611\u201325628 (2019)","journal-title":"Multimedia Tools Appl."},{"key":"26_CR3","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1109\/TIP.2017.2760518","volume":"27","author":"S Bosse","year":"2018","unstructured":"Bosse, S., et al.: Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans. Image Process. 27, 206\u2013219 (2018)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Buczkowski, M., Stasinski, R.: Convolutional neural network-based image distortion classification. In: 2019 International Conference on Systems, Signals and Image Processing (IWSSIP) (2019)","key":"26_CR4","DOI":"10.1109\/IWSSIP.2019.8787212"},{"issue":"6","key":"26_CR5","doi-asserted-by":"publisher","first-page":"679","DOI":"10.1109\/TPAMI.1986.4767851","volume":"8","author":"J Canny","year":"1986","unstructured":"Canny, J.: A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. PAMI 8(6), 679\u2013698 (1986)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell. PAMI"},{"doi-asserted-by":"crossref","unstructured":"Ding, K., et al.: Image quality assessment: unifying structure and texture similarity. IEEE Trans. Pattern Anal. Mach. Intell. (2020)","key":"26_CR6","DOI":"10.1109\/TPAMI.2020.3045810"},{"doi-asserted-by":"crossref","unstructured":"Feng, Y., et al.: Quality assessment of synthetic fluorescence microscopy images for image segmentation. In: 2019 IEEE International Conference on Image Processing (ICIP), pp. 814\u2013818 (2019)","key":"26_CR7","DOI":"10.1109\/ICIP.2019.8802971"},{"key":"26_CR8","doi-asserted-by":"publisher","first-page":"104","DOI":"10.1016\/j.neucom.2017.01.054","volume":"257","author":"F Gao","year":"2017","unstructured":"Gao, F., et al.: DeepSim: deep similarity for image quality assessment. Neurocomputing 257, 104\u2013114 (2017)","journal-title":"Neurocomputing"},{"doi-asserted-by":"crossref","unstructured":"Gao, X., et al.: Image quality assessment - a multiscale geometric analysis-based framework and examples. In: Handbook of Natural Computing (2012)","key":"26_CR9","DOI":"10.1007\/978-3-540-92910-9_11"},{"unstructured":"Girod, B.: What\u2019s wrong with mean-squared error? (1993)","key":"26_CR10"},{"key":"26_CR11","doi-asserted-by":"publisher","first-page":"292","DOI":"10.1016\/j.jvcir.2018.12.045","volume":"59","author":"W Hu","year":"2019","unstructured":"Hu, W., et al.: A new method of thangka image inpainting quality assessment. J. Vis. Commun. Image Represent. 59, 292\u2013299 (2019)","journal-title":"J. Vis. Commun. Image Represent."},{"doi-asserted-by":"crossref","unstructured":"Huang, C., Jiang, T., Jiang, M.: Encoding distortions for multi-task full-reference image quality assessment. In: 2019 IEEE International Conference on Multimedia and Expo (ICME), pp. 1864\u20131869 (2019)","key":"26_CR12","DOI":"10.1109\/ICME.2019.00320"},{"doi-asserted-by":"crossref","unstructured":"Kang, L., et al.: Simultaneous estimation of image quality and distortion via multi-task convolutional neural networks. In: 2015 IEEE International Conference on Image Processing (ICIP), pp. 2791\u20132795 (2015)","key":"26_CR13","DOI":"10.1109\/ICIP.2015.7351311"},{"unstructured":"Kendall, A., Gal, Y., Cipolla, R.: Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7482\u20137491 (2018)","key":"26_CR14"},{"doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, S.: Deep learning of human visual sensitivity in image quality assessment framework. In: IEEE Conference on Computer Vision & Pattern Recognition, pp. 1969\u20131977 (2017)","key":"26_CR15","DOI":"10.1109\/CVPR.2017.213"},{"key":"26_CR16","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1109\/TNNLS.2018.2829819","volume":"30","author":"J Kim","year":"2019","unstructured":"Kim, J., Nguyen, A., Lee, S.: Deep CNN-based blind image quality predictor. IEEE Trans. Neural Netw. Learn. Syst. 30, 11\u201324 (2019)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"26_CR17","doi-asserted-by":"publisher","first-page":"4219","DOI":"10.1109\/TIP.2020.2968283","volume":"29","author":"W Kim","year":"2020","unstructured":"Kim, W., et al.: Dynamic receptive field generation for full-reference image quality assessment. IEEE Trans. Image Process. 29, 4219\u20134231 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"26_CR18","doi-asserted-by":"publisher","first-page":"011006","DOI":"10.1117\/1.3267105","volume":"19","author":"EC Larson","year":"2010","unstructured":"Larson, E.C., Chandler, D.: Most apparent distortion: full-reference image quality assessment and the role of strategy. J. Electron. Imag. 19, 011006 (2010)","journal-title":"J. Electron. Imag."},{"doi-asserted-by":"crossref","unstructured":"Liu, S., Johns, E., Davison, A.: End-to-end multi-task learning with attention. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1871\u20131880 (2019)","key":"26_CR19","DOI":"10.1109\/CVPR.2019.00197"},{"key":"26_CR20","doi-asserted-by":"publisher","first-page":"1202","DOI":"10.1109\/TIP.2017.2774045","volume":"27","author":"K Ma","year":"2018","unstructured":"Ma, K., et al.: End-to-end blind image quality assessment using deep neural networks. IEEE Trans. Image Process. 27, 1202\u20131213 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"6","key":"26_CR21","first-page":"1","volume":"79","author":"X Ma","year":"2020","unstructured":"Ma, X., Jiang, X.: Multimedia image quality assessment based on deep feature extraction. Multimedia Tools Appl. 79(6), 1\u201312 (2020)","journal-title":"Multimedia Tools Appl."},{"unstructured":"Ponomarenko, N., et al.: Color image database TID2013: peculiarities and preliminary results. In: European Workshop on Visual Information Processing (EUVIP), pp. 106\u2013111 (2013)","key":"26_CR22"},{"unstructured":"Ruder, S.: An overview of multi-task learning in deep neural networks. ArXiv abs\/1706.05098 (2017)","key":"26_CR23"},{"key":"26_CR24","doi-asserted-by":"publisher","first-page":"430","DOI":"10.1109\/TIP.2005.859378","volume":"15","author":"H Sheikh","year":"2006","unstructured":"Sheikh, H., Bovik, A.: Image information and visual quality. IEEE Trans. Image Process. 15, 430\u2013444 (2006)","journal-title":"IEEE Trans. Image Process."},{"unstructured":"Wang, Y., et al.: Image quality evaluation based on image weighted separating block peak signal to noise ratio. In: International Conference on Neural Networks & Signal Processing (2003)","key":"26_CR25"},{"issue":"5","key":"26_CR26","doi-asserted-by":"publisher","first-page":"1185","DOI":"10.1109\/TIP.2010.2092435","volume":"20","author":"Z Wang","year":"2011","unstructured":"Wang, Z., Li, Q.: Information content weighting for perceptual image quality assessment. IEEE Trans. Image Process. 20(5), 1185\u20131198 (2011)","journal-title":"IEEE Trans. Image Process."},{"key":"26_CR27","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., et al.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13, 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"26_CR28","doi-asserted-by":"publisher","first-page":"684","DOI":"10.1109\/TIP.2013.2293423","volume":"23","author":"W Xue","year":"2014","unstructured":"Xue, W., et al.: Gradient magnitude similarity deviation: a highly efficient perceptual image quality index. IEEE Trans. Image Process. 23, 684\u2013695 (2014)","journal-title":"IEEE Trans. Image Process."},{"key":"26_CR29","doi-asserted-by":"publisher","first-page":"2378","DOI":"10.1109\/TIP.2011.2109730","volume":"20","author":"L Zhang","year":"2011","unstructured":"Zhang, L., et al.: FSIM: a feature similarity index for image quality assessment. IEEE Trans. Image Process. 20, 2378\u20132386 (2011)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Zhang, W., et al.: RankSRGAN: generative adversarial networks with ranker for image super-resolution. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 3096\u20133105 (2019)","key":"26_CR30","DOI":"10.1109\/ICCV.2019.00319"},{"doi-asserted-by":"crossref","unstructured":"Zhang, Y., Yang, Q.: A survey on multi-task learning. IEEE Trans. Knowl. Data Eng. (2021)","key":"26_CR31","DOI":"10.1109\/TKDE.2021.3070203"}],"container-title":["Lecture Notes in Computer Science","PRICAI 2021: Trends in Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-89370-5_26","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T01:15:22Z","timestamp":1635729322000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-89370-5_26"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030893699","9783030893705"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-89370-5_26","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"1 November 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific Rim International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hanoi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vietnam","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 November 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 November 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"pricai2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.pricai.org\/2021","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"382","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"93","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"24% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}