{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T15:25:06Z","timestamp":1773933906176,"version":"3.50.1"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T00:00:00Z","timestamp":1763078400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T00:00:00Z","timestamp":1763078400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100004761","name":"Natural Science Foundation of Henan Province of China","doi-asserted-by":"crossref","award":["252300421760"],"award-info":[{"award-number":["252300421760"]}],"id":[{"id":"10.13039\/501100004761","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cluster Comput"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s10586-025-05858-0","type":"journal-article","created":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T19:24:26Z","timestamp":1763148266000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Emotion-style dual prediction: a multi-task deep learning approach for artistic images"],"prefix":"10.1007","volume":"29","author":[{"given":"Yao","family":"Lu","sequence":"first","affiliation":[]},{"given":"Dan","family":"Shao","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Ni","sequence":"additional","affiliation":[]},{"given":"Hanrui","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Manxing","family":"Shi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,14]]},"reference":[{"issue":"10","key":"5858_CR1","doi-asserted-by":"publisher","first-page":"1236","DOI":"10.1109\/JPROC.2023.3273517","volume":"111","author":"JZ Wang","year":"2023","unstructured":"Wang, J.Z., Zhao, S., Wu, C., Adams, R.B., Newman, M.G., Shafir, T., Tsachor, R.: Unlocking the emotional world of visual media: an overview of the science, research, and impact of understanding emotion. Proc. IEEE 111(10), 1236\u20131286 (2023)","journal-title":"Proc. IEEE"},{"key":"5858_CR2","doi-asserted-by":"crossref","unstructured":"Zheng, J., Li, W., Hong, J., Petersson, L., Barnes, N.: Towards open-set object detection and discovery. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3961\u20133970 (2022)","DOI":"10.1109\/CVPRW56347.2022.00441"},{"issue":"7","key":"5858_CR3","doi-asserted-by":"publisher","first-page":"4225","DOI":"10.1109\/TVCG.2023.3255820","volume":"30","author":"H Zhang","year":"2023","unstructured":"Zhang, H., Cheng, S., El Amm, C., Kim, J.: Efficient pooling operator for 3d morphable models. IEEE Trans. Visual Comput. Graphics 30(7), 4225\u20134233 (2023)","journal-title":"IEEE Trans. Visual Comput. Graphics"},{"issue":"2","key":"5858_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3475799","volume":"18","author":"E Cetinic","year":"2022","unstructured":"Cetinic, E., She, J.: Understanding and creating art with ai: review and outlook. ACM Trans. Multimed. Comput. Commun. Appl. 18(2), 1\u201322 (2022)","journal-title":"ACM Trans. Multimed. Comput. Commun. Appl."},{"issue":"S18","key":"5858_CR5","doi-asserted-by":"publisher","first-page":"50","DOI":"10.14733\/cadaps.2024.S18.50-65","volume":"21","author":"J Huang","year":"2024","unstructured":"Huang, J.: Automatic extraction and reconstruction of drawing design elements based on computer vision and neural networks. Comput. Aided Design Appl. 21(S18), 50\u201365 (2024)","journal-title":"Comput. Aided Design Appl."},{"key":"5858_CR6","doi-asserted-by":"crossref","unstructured":"Jin, K., Wang, Y., Santos, L., Fang, T., Yang, X., Im, S.K., Oliveira, H.G.: Reasoning or not? A comprehensive evaluation of reasoning llms for dialogue summarization. Expert Systems with Applications, 129831 (2025)","DOI":"10.1016\/j.eswa.2025.129831"},{"issue":"7","key":"5858_CR7","doi-asserted-by":"publisher","DOI":"10.3390\/plants12071431","volume":"12","author":"J Yuan","year":"2023","unstructured":"Yuan, J., Zhang, L., Kim, C.-S.: Multimodal interaction of mu plant landscape design in marine urban based on computer vision technology. Plants 12(7), 1431 (2023)","journal-title":"Plants"},{"issue":"13","key":"5858_CR8","doi-asserted-by":"publisher","DOI":"10.3390\/math13132086","volume":"13","author":"Y Tian","year":"2025","unstructured":"Tian, Y., Xu, S., Cao, Y., Wang, Z., Wei, Z.: An empirical comparison of machine learning and deep learning models for automated fake news detection. Mathematics 13(13), 2086 (2025)","journal-title":"Mathematics"},{"key":"5858_CR9","doi-asserted-by":"crossref","unstructured":"Martinez Pandiani, D.S., Lazzari, N., Presutti, V.: Stitching gaps: fusing situated perceptual knowledge with vision transformers for high-level image classification. In: Knowledge Graphs in the Age of Language Models and Neuro-Symbolic AI. IOS Press, pp. 68\u201387 (2024)","DOI":"10.3233\/SSW240008"},{"key":"5858_CR10","unstructured":"Kleinlein, R., Luna-Jim\u00e9nez, C., Mart\u00edn-Fern\u00e1ndez, I., Fern\u00e1ndez-Mart\u00ednez, F.: Leveraging in-domain clip for emotion prediction elicited by artworks. Available at SSRN 4453302"},{"issue":"1","key":"5858_CR11","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-023-39865-1","volume":"13","author":"BT Spee","year":"2023","unstructured":"Spee, B.T., Mikuni, J., Leder, H., Scharnowski, F., Pelowski, M., Steyrl, D.: Machine learning revealed symbolism, emotionality, and imaginativeness as primary predictors of creativity evaluations of western art paintings. Sci. Rep. 13(1), 12966 (2023)","journal-title":"Sci. Rep."},{"issue":"2","key":"5858_CR12","doi-asserted-by":"publisher","DOI":"10.3390\/electronics13020302","volume":"13","author":"J Wang","year":"2024","unstructured":"Wang, J., Maeda, A., Kawagoe, K.: Multartrec: a multimodal neural topic modeling for integrating image and text features in artwork recommendation. Electronics 13(2), 302 (2024)","journal-title":"Electronics"},{"key":"5858_CR13","doi-asserted-by":"crossref","unstructured":"Zhang, H., Li, X., Kim, J., Cheng, S., El Amm, C.: Neural qslim for mesh autoencoders. In: International Conference on Artificial Neural Networks. Springer, pp. 51\u201365 (2025)","DOI":"10.1007\/978-3-032-04555-3_5"},{"key":"5858_CR14","doi-asserted-by":"publisher","first-page":"75 742","DOI":"10.1109\/ACCESS.2022.3192026","volume":"10","author":"M Zhang","year":"2022","unstructured":"Zhang, M., Xie, K., Zhang, Y.-H., Wen, C., He, J.-B.: Fine segmentation on faces with masks based on a multistep iterative segmentation algorithm. IEEE Access 10, 75 742-75 753 (2022)","journal-title":"IEEE Access"},{"key":"5858_CR15","doi-asserted-by":"publisher","first-page":"66 357","DOI":"10.1109\/ACCESS.2023.3289713","volume":"11","author":"A W\u0119do\u0142owska","year":"2023","unstructured":"W\u0119do\u0142owska, A., Weber, D., Kostek, B.: Predicting emotion from color present in images and video excerpts by machine learning. IEEE Access 11, 66 357-66 373 (2023)","journal-title":"IEEE Access"},{"issue":"1","key":"5858_CR16","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1386\/tear_00097_1","volume":"21","author":"S Vlachou","year":"2023","unstructured":"Vlachou, S., Panagopoulos, M.: Applying machine learning methods to quantify emotional experience in installation art. Technoetic Arts 21(1), 53\u201372 (2023)","journal-title":"Technoetic Arts"},{"key":"5858_CR17","doi-asserted-by":"crossref","unstructured":"Dai, W., Jiang, Y., Liu, Y., Chen, J., Sun, X., Tao, J.: Cab-kws: contrastive augmentation: an unsupervised learning approach for keyword spotting in speech technology. In: International Conference on Pattern Recognition. Springer, pp. 98\u2013112 (2025)","DOI":"10.1007\/978-3-031-78122-3_7"},{"issue":"11","key":"5858_CR18","doi-asserted-by":"publisher","first-page":"2590","DOI":"10.1007\/s11263-022-01664-y","volume":"130","author":"G Castellano","year":"2022","unstructured":"Castellano, G., Vessio, G.: A deep learning approach to clustering visual arts. International Journal of Computer Vision 130(11), 2590\u20132605 (2022)","journal-title":"International Journal of Computer Vision"},{"key":"5858_CR19","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1007\/978-981-10-7221-5_11","volume-title":"Heritage Preservation: A Computational Approach","author":"A Mohanty","year":"2018","unstructured":"Mohanty, A., Roy, K., Sahay, R.R.: Nrityamanthan: unravelling the intent of the dancer using deep learning. In: Heritage Preservation: A Computational Approach, pp. 209\u2013239. Springer, Berlin (2018)"},{"key":"5858_CR20","doi-asserted-by":"crossref","unstructured":"Suma, K., Balasm, Z., Sasirekha, N., Manikandan, V., Sandhya, K.: Emotion recognition in visual art on color and texture analysis using improved vision transformer. In: 2025 International Conference on Intelligent Systems and Computational Networks (ICISCN). IEEE, pp. 1\u20136 (2025)","DOI":"10.1109\/ICISCN64258.2025.10934620"},{"issue":"17","key":"5858_CR21","doi-asserted-by":"publisher","DOI":"10.3390\/math13172740","volume":"13","author":"D Qu","year":"2025","unstructured":"Qu, D., Ma, Y.: Magnet-bn: markov-guided Bayesian neural networks for calibrated long-horizon sequence forecasting and community tracking. Mathematics 13(17), 2740 (2025)","journal-title":"Mathematics"},{"key":"5858_CR22","doi-asserted-by":"crossref","unstructured":"Kleinlein, R., Luna-Jim\u00e9nez, C., Mart\u00edn-Fern\u00e1ndez, I., Fern\u00e1ndez-Mart\u00ednez, F.: Leveraging in-domain clip for emotion prediction elicited by artworks, Available at SSRN 4453302 (2023)","DOI":"10.2139\/ssrn.4453302"},{"key":"5858_CR23","doi-asserted-by":"publisher","first-page":"14 343","DOI":"10.1007\/s11042-020-10203-1","volume":"80","author":"J Wei","year":"2021","unstructured":"Wei, J., Yang, X., Dong, Y.: User-generated video emotion recognition based on key frames. Multimed. Tools Appl. 80, 14 343-14 361 (2021)","journal-title":"Multimed. Tools Appl."},{"key":"5858_CR24","doi-asserted-by":"crossref","unstructured":"Cakmak, M.C., Shaik, M., Agarwal, N.: Emotion assessment of youtube videos using color theory. In: Proceedings of the 2024 9th International Conference on Multimedia and Image Processing, pp. 6\u201314 (2024)","DOI":"10.1145\/3665026.3665028"},{"key":"5858_CR25","doi-asserted-by":"publisher","DOI":"10.1007\/s13369-023-08614-x","author":"Z Malik","year":"2024","unstructured":"Malik, Z., Shapiai, M.I.B., Zaidi, S.Z.Y.: A novel key flow frame selection method for video classification. Arab. J. Sci. Eng. (2024). https:\/\/doi.org\/10.1007\/s13369-023-08614-x","journal-title":"Arab. J. Sci. Eng."},{"key":"5858_CR26","doi-asserted-by":"crossref","unstructured":"Behrad, F., Tuytelaars, T., Wagemans, J.: Charm: the missing piece in vit fine-tuning for image aesthetic assessment. In: Proceedings of the Computer Vision and Pattern Recognition Conference, pp. 7815\u20137824 (2025)","DOI":"10.1109\/CVPR52734.2025.00732"},{"issue":"10","key":"5858_CR27","doi-asserted-by":"publisher","first-page":"2259","DOI":"10.5829\/ije.2025.38.10a.05","volume":"38","author":"M Rohani","year":"2025","unstructured":"Rohani, M., Farsi, H., Mohamadzadeh, S.: Advanced multi-task learning with lightweight networks and multi-head attention for efficient facial attribute estimation. International Journal of Engineering 38(10), 2259\u20132272 (2025)","journal-title":"International Journal of Engineering"},{"key":"5858_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2025.130810","volume":"650","author":"A Li","year":"2025","unstructured":"Li, A., Xu, L., Ling, C., Zhang, J., Wang, P.: Emoverse: enhancing multimodal large language models for affective computing via multitask learning. Neurocomputing 650, 130810 (2025)","journal-title":"Neurocomputing"},{"issue":"22","key":"5858_CR29","doi-asserted-by":"publisher","first-page":"23 660","DOI":"10.1609\/aaai.v39i22.34536","volume":"39","author":"Y Chen","year":"2025","unstructured":"Chen, Y., Li, Z., You, S., Chen, Z., Chang, J., Zhang, Y., Dai, W., Guo, Q., Xiao, Y.: Attributive reasoning for hallucination diagnosis of large language models. Proceedings of the AAAI Conference on Artificial Intelligence 39(22), 23 660-23 668 (2025)","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"issue":"11\u201312","key":"5858_CR30","doi-asserted-by":"publisher","first-page":"843","DOI":"10.1037\/bul0000381","volume":"148","author":"A Koutsoumpis","year":"2022","unstructured":"Koutsoumpis, A., Oostrom, J.K., Holtrop, D., Van Breda, W., Ghassemi, S., de Vries, R.E.: The kernel of truth in text-based personality assessment: a meta-analysis of the relations between the big five and the linguistic inquiry and word count (liwc). Psychol. Bull. 148(11\u201312), 843 (2022)","journal-title":"Psychol. Bull."},{"issue":"5","key":"5858_CR31","doi-asserted-by":"publisher","first-page":"5516","DOI":"10.1007\/s11227-022-04881-x","volume":"79","author":"A Mewada","year":"2023","unstructured":"Mewada, A., Dewang, R.K.: Sa-asba: a hybrid model for aspect-based sentiment analysis using synthetic attention in pre-trained language bert model with extreme gradient boosting. J. Supercomput. 79(5), 5516\u20135551 (2023)","journal-title":"J. Supercomput."},{"key":"5858_CR32","unstructured":"Dou, Z.-Y., Xu, Y., Gan, Z., Wang, J., Wang, S., Wang, L., Zhu, C., Zhang, P., Yuan, L., Peng, N., et al.: An empirical study of training end-to-end vision-and-language transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 18\u00a0166\u201318\u00a0176 (2022)"},{"issue":"5","key":"5858_CR33","doi-asserted-by":"publisher","first-page":"903","DOI":"10.1007\/s12530-022-09464-y","volume":"14","author":"S Porkodi","year":"2023","unstructured":"Porkodi, S., Sarada, V., Maik, V., Gurushankar, K.: Generic image application using gans (generative adversarial networks): a review. Evolving Systems 14(5), 903\u2013917 (2023)","journal-title":"Evolving Systems"},{"key":"5858_CR34","first-page":"529","volume":"47","author":"A Mohanty","year":"2016","unstructured":"Mohanty, A., Vaishnavi, P., Jana, P., Majumdar, A., Ahmed, A., Goswami, T., Sahay, R.R.: Nrityabodha: towards understanding Indian classical dance using a deep learning approach. Signal Processing: Image Communication 47, 529\u2013548 (2016)","journal-title":"Signal Processing: Image Communication"},{"key":"5858_CR35","first-page":"400","volume-title":"International Conference on Pattern Recognition","author":"K Roy","year":"2024","unstructured":"Roy, K., Mohanty, A., Ranjan Sahay, R.: Hand over face gesture classification with feature driven vision transformer and supervised contrastive learning. In: International Conference on Pattern Recognition, pp. 400\u2013417. Springer, Berlin (2024)"},{"issue":"9","key":"5858_CR36","doi-asserted-by":"publisher","first-page":"6507","DOI":"10.1007\/s00371-023-03179-0","volume":"40","author":"A Mohanty","year":"2024","unstructured":"Mohanty, A., Roy, K., Sahay, R.R.: Robust static hand gesture recognition: harnessing sparsity of deeply learned features. Vis. Comput. 40(9), 6507\u20136531 (2024)","journal-title":"Vis. Comput."},{"key":"5858_CR37","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2020.114185","volume":"168","author":"X Zhang","year":"2021","unstructured":"Zhang, X., Wu, F., Li, Z.: Application of convolutional neural network to traditional data. Expert Systems with Applications 168, 114185 (2021)","journal-title":"Expert Systems with Applications"},{"key":"5858_CR38","doi-asserted-by":"publisher","first-page":"33 679","DOI":"10.1109\/ACCESS.2022.3161428","volume":"10","author":"R Castro","year":"2022","unstructured":"Castro, R., Pineda, I., Lim, W., Morocho-Cayamcela, M.E.: Deep learning approaches based on transformer architectures for image captioning tasks. IEEE Access 10, 33 679-33 694 (2022)","journal-title":"IEEE Access"},{"key":"5858_CR39","doi-asserted-by":"publisher","first-page":"262","DOI":"10.1016\/j.neucom.2022.10.076","volume":"520","author":"H Zhang","year":"2023","unstructured":"Zhang, H., Luo, Y., Zhang, L., Wu, Y., Wang, M., Shen, Z.: Considering three elements of aesthetics: multi-task self-supervised feature learning for image style classification. Neurocomputing 520, 262\u2013273 (2023)","journal-title":"Neurocomputing"},{"key":"5858_CR40","doi-asserted-by":"crossref","unstructured":"Wei, Z., Zhang, J., Lin, Z., Lee, J.-Y., Balasubramanian, N., Hoai, M., Samaras, D.: Learning visual emotion representations from web data. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13106\u201313115 (2020)","DOI":"10.1109\/CVPR42600.2020.01312"},{"issue":"5","key":"5858_CR41","doi-asserted-by":"publisher","first-page":"2177","DOI":"10.1007\/s00371-022-02472-8","volume":"39","author":"H Yang","year":"2023","unstructured":"Yang, H., Fan, Y., Lv, G., Liu, S., Guo, Z.: Exploiting emotional concepts for image emotion recognition. Vis. Comput. 39(5), 2177\u20132190 (2023)","journal-title":"Vis. Comput."},{"issue":"6","key":"5858_CR42","doi-asserted-by":"publisher","DOI":"10.3390\/jimaging10060136","volume":"10","author":"T Chen","year":"2024","unstructured":"Chen, T., Garcia, N., Li, L., Nakashima, Y.: Exploring emotional stimuli detection in artworks: a benchmark dataset and baselines evaluation. J. Imaging 10(6), 136 (2024)","journal-title":"J. Imaging"},{"key":"5858_CR43","doi-asserted-by":"publisher","first-page":"1640","DOI":"10.1016\/j.procs.2023.01.142","volume":"218","author":"G Meena","year":"2023","unstructured":"Meena, G., Mohbey, K.K.: Sentiment analysis on images using different transfer learning models. Procedia Computer Science 218, 1640\u20131649 (2023)","journal-title":"Procedia Computer Science"},{"key":"5858_CR44","first-page":"129","volume-title":"International Conference on Image Analysis and Processing","author":"S Aslan","year":"2022","unstructured":"Aslan, S., Castellano, G., Digeno, V., Migailo, G., Scaringi, R., Vessio, G.: Recognizing the emotions evoked by artworks through visual features and knowledge graph-embeddings. In: International Conference on Image Analysis and Processing, pp. 129\u2013140. Springer, Berlin (2022)"},{"issue":"9","key":"5858_CR45","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0291647","volume":"18","author":"P Meng","year":"2023","unstructured":"Meng, P., Meng, X., Hu, R., Zhang, L.: Predicting the aesthetics of dynamic generative artwork based on statistical image features: a time-dependent model. PLoS ONE 18(9), e0291647 (2023)","journal-title":"PLoS ONE"},{"key":"5858_CR46","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1016\/j.patcog.2018.01.035","volume":"79","author":"A Mohanty","year":"2018","unstructured":"Mohanty, A., Sahay, R.R.: Rasabodha: understanding Indian classical dance by recognizing emotions using deep learning. Pattern Recognition 79, 97\u2013113 (2018)","journal-title":"Pattern Recognition"}],"container-title":["Cluster Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-025-05858-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10586-025-05858-0","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-025-05858-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T13:07:25Z","timestamp":1773925645000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10586-025-05858-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,14]]},"references-count":46,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["5858"],"URL":"https:\/\/doi.org\/10.1007\/s10586-025-05858-0","relation":{},"ISSN":["1386-7857","1573-7543"],"issn-type":[{"value":"1386-7857","type":"print"},{"value":"1573-7543","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,14]]},"assertion":[{"value":"31 March 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 October 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 November 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 November 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}],"article-number":"31"}}