{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,23]],"date-time":"2025-11-23T06:17:58Z","timestamp":1763878678417,"version":"3.40.3"},"reference-count":22,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T00:00:00Z","timestamp":1743120000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T00:00:00Z","timestamp":1743120000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,5]]},"DOI":"10.1007\/s11760-025-04029-3","type":"journal-article","created":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T05:06:28Z","timestamp":1743397588000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Multi-theme image aesthetic assessment based on incremental learning"],"prefix":"10.1007","volume":"19","author":[{"given":"Wenjing","family":"Cao","sequence":"first","affiliation":[]},{"given":"Yongzhen","family":"Ke","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Shuai","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Fan","family":"Qin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,3,28]]},"reference":[{"key":"4029_CR1","doi-asserted-by":"publisher","unstructured":"He, S., Zhang, Y., Xie, R., Jiang, D., Ming, A.: Rethinking Image Aesthetics Assessment: Models, Datasets and Benchmarks. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, International Joint Conferences on Artificial Intelligence Organization, Vienna, Austria. 942\u2013948 (2022). https:\/\/doi.org\/10.24963\/ijcai.2022\/132","DOI":"10.24963\/ijcai.2022\/132"},{"key":"4029_CR2","doi-asserted-by":"publisher","unstructured":"Jiachen Yang, Y., Zhou, Y., Zhao, W., Lu, X., Gao: MetaMP: Metalearning-Based Multipatch Image Aesthetics Assessment. IEEE Trans. Cybern. 53(9) 5716\u20135728 (2023). https:\/\/doi.org\/10.1109\/TCYB.2022.3169017","DOI":"10.1109\/TCYB.2022.3169017"},{"key":"4029_CR3","unstructured":"Bruce Barnbaum: The Art of Photography: A Personal Approach To Artistic Expression, 2nd edn. Rocky Nook (2017)"},{"key":"4029_CR4","doi-asserted-by":"publisher","unstructured":"Hossein Talebi and Peyman Milanfar: NIMA: Neural Image Assessment. IEEE Trans. on Image Process. 27(8), 3998\u20134011 (2018). https:\/\/doi.org\/10.1109\/TIP.2018.2831899","DOI":"10.1109\/TIP.2018.2831899"},{"key":"4029_CR5","doi-asserted-by":"publisher","unstructured":"Shuang Ma, J., Liu, C. A-Lamp: Adaptive Layout-Aware Multi-patch Deep Convolutional Neural Network for Photo Aesthetic Assessment. In 2017 IEEE Conference on Computer, Vision, P., Recognition, H.I.: 722\u2013731 (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.84","DOI":"10.1109\/CVPR.2017.84"},{"key":"4029_CR6","doi-asserted-by":"publisher","unstructured":"Chaoran Cui, H., Liu, T., Lian, L., Nie, L., Zhu, Y.Y.: Distribution-Oriented Aesthetics Assessment With Semantic-Aware Hybrid Network. IEEE Trans. Multimedia 21(5) 1209\u20131220. (2019). https:\/\/doi.org\/10.1109\/TMM.2018.2875357","DOI":"10.1109\/TMM.2018.2875357"},{"key":"4029_CR7","unstructured":"Nishi Doshi, G., Shikhenawis, S.K., Mitra: Image Aesthetics Assessment using Multi Channel Convolutional Neural Networks. Retrieved August 23, 2023 from (2019). http:\/\/arxiv.org\/abs\/1911.09301"},{"key":"4029_CR8","unstructured":"Zhizhong Li and Derek Hoiem: Learning without Forgetting. Retrieved August 23, 2023 from (2017). http:\/\/arxiv.org\/abs\/1606.09282"},{"key":"4029_CR9","unstructured":"Rebuffi, S.-A., Kolesnikov, A., Sperl, G., Christoph, H., Lampert: iCaRL: Incremental Classifier and Representation Learning. Retrieved August 23, 2023 from (2017). http:\/\/arxiv.org\/abs\/1611.07725"},{"key":"4029_CR10","unstructured":"Shipeng, Y., Xie, J., He, X.: DER: Dynamically Expandable Representation for Class Incremental Learning. Retrieved August 23, 2023 from (2021). http:\/\/arxiv.org\/abs\/2103.16788"},{"key":"4029_CR11","unstructured":"Jie, H., Albanie, L.S.S.: Gang Sun, and Enhua Wu. Squeeze-and-Excitation Networks. Retrieved August 23, 2023 from (2019). http:\/\/arxiv.org\/abs\/1709.01507"},{"key":"4029_CR12","unstructured":"Max Jaderberg, K., Simonyan, A., Zisserman, K.K.: Spatial Transformer Networks. Retrieved August 23, 2023 from (2016). http:\/\/arxiv.org\/abs\/1506.02025"},{"key":"4029_CR13","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep Residual Learning for Image Recognition. Retrieved August 23, 2023 from (2015). http:\/\/arxiv.org\/abs\/1512.03385"},{"key":"4029_CR14","unstructured":"Fu, J., Liu, J., Tian, H., Li, Y., Bao, Y., Hanqing Lu: Zhiwei Fang, and. Dual Attention Network for Scene Segmentation. Retrieved August 23, 2023 from (2019). http:\/\/arxiv.org\/abs\/1809.02983"},{"key":"4029_CR15","unstructured":"Joan Serr\u00e0, D., Sur\u00eds, M., Miron, A., Karatzoglou: Overcoming catastrophic forgetting with hard attention to the task. Retrieved August 23, 2023 from (2018). http:\/\/arxiv.org\/abs\/1801.01423"},{"key":"4029_CR16","doi-asserted-by":"publisher","unstructured":"Murray, N., Marchesotti, L., Perronnin, F. AVA: A large-scale database for aesthetic visual analysis. In 2012 IEEE Conference on Computer Vision and, Recognition, P., IEEE, Providence, R.I.: 2408\u20132415 (2012). https:\/\/doi.org\/10.1109\/CVPR.2012.6247954","DOI":"10.1109\/CVPR.2012.6247954"},{"key":"4029_CR17","doi-asserted-by":"publisher","unstructured":"Levina, E., Bickel, P.: The Earth Mover\u2019s distance is the Mallows distance: some insights from statistics. In Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001, IEEE Comput. Soc, Vancouver, BC, Canada, 251\u2013256. (2001). https:\/\/doi.org\/10.1109\/ICCV.2001.937632","DOI":"10.1109\/ICCV.2001.937632"},{"key":"4029_CR18","doi-asserted-by":"publisher","unstructured":"Dongyu She, Y.-K., Lai, G.Y., Xu, K.: Hierarchical Layout-Aware Graph Convolutional Network for Unified Aesthetics Assessment. In 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, Nashville, TN, USA, 8471\u20138480. (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00837","DOI":"10.1109\/CVPR46437.2021.00837"},{"key":"4029_CR19","doi-asserted-by":"publisher","unstructured":"Li, L., Zhu, H., Zhao, S., Ding, G.: and Weisi Lin. Personality-Assisted Multi-Task Learning for Generic and Personalized Image Aesthetics Assessment. IEEE Trans. on Image Process. 29 3898\u20133910. (2020). https:\/\/doi.org\/10.1109\/TIP.2020.2968285","DOI":"10.1109\/TIP.2020.2968285"},{"key":"4029_CR20","doi-asserted-by":"publisher","unstructured":"Zhu, H., Li, L., Wu, J., Zhao, S., Ding, G.: and Guangming Shi. Personalized Image Aesthetics Assessment via Meta-Learning With Bilevel Gradient Optimization. IEEE Trans. Cybern. 52(3) 1798\u20131811. (2022). https:\/\/doi.org\/10.1109\/TCYB.2020.2984670","DOI":"10.1109\/TCYB.2020.2984670"},{"key":"4029_CR21","unstructured":"Chen, Q., Zhang, W., Zhou, N., Lei, P., Xu, Y.: Yu Zheng, and Jianping Fan. Adaptive Fractional Dilated Convolution Network for Image Aesthetics Assessment. Retrieved August 23, 2023 from (2020). http:\/\/arxiv.org\/abs\/2004.03015"},{"key":"4029_CR22","doi-asserted-by":"publisher","unstructured":"Zhang, X., Zhang, X., Xiao, Y.: and Gang Liu. Theme-Aware Semi-Supervised Image Aesthetic Quality Assessment. Mathematics 10(15) 2609. (2022). https:\/\/doi.org\/10.3390\/math10152609","DOI":"10.3390\/math10152609"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04029-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04029-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04029-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,8]],"date-time":"2025-04-08T20:11:58Z","timestamp":1744143118000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04029-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,28]]},"references-count":22,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,5]]}},"alternative-id":["4029"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04029-3","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2025,3,28]]},"assertion":[{"value":"29 August 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 February 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 March 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 March 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"This article operates a type 3 research data policy, which permits use, sharing, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third-party material in this article are included in the article\u2019s Creative Commons license unless indicated otherwise in a credit line to the material. If material is not included in the article\u2019s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you must obtain permission directly from the copyright holder.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Open access"}}],"article-number":"421"}}