{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T14:40:21Z","timestamp":1758724821164,"version":"3.44.0"},"reference-count":71,"publisher":"Springer Science and Business Media LLC","issue":"13","license":[{"start":{"date-parts":[[2025,7,12]],"date-time":"2025-07-12T00:00:00Z","timestamp":1752278400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,7,12]],"date-time":"2025-07-12T00:00:00Z","timestamp":1752278400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100021171","name":"Basic and Applied Basic Research Foundation of Guangdong Province","doi-asserted-by":"publisher","award":["2023A1515010939"],"award-info":[{"award-number":["2023A1515010939"]}],"id":[{"id":"10.13039\/501100021171","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Project of Education Department of Guangdong Province","award":["2022KTSCX068"],"award-info":[{"award-number":["2022KTSCX068"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62006049","62172113"],"award-info":[{"award-number":["62006049","62172113"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Guangdong Province Key Area R and D Plan Project","award":["2020B1111120001"],"award-info":[{"award-number":["2020B1111120001"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s00371-025-04086-2","type":"journal-article","created":{"date-parts":[[2025,7,12]],"date-time":"2025-07-12T08:32:29Z","timestamp":1752309149000},"page":"11069-11082","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["ViDMNet: vision transformer-based dual-polarity memory network for image emotion recognition"],"prefix":"10.1007","volume":"41","author":[{"given":"Rui","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Zhongcheng","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Huihui","family":"Li","sequence":"additional","affiliation":[]},{"given":"Xiaoyong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Guihua","family":"Wen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,12]]},"reference":[{"unstructured":"de\u00a0Charms, R.: Personal causation: The internal affective determinants of behavior (1983)","key":"4086_CR1"},{"doi-asserted-by":"crossref","unstructured":"Zhang, Z., Luo, P., Loy, C.-C., Tang, X.: Learning social relation traits from face images. In 2015 IEEE International Conference on Computer Vision (ICCV), pp. 3631\u20133639 (2015)","key":"4086_CR2","DOI":"10.1109\/ICCV.2015.414"},{"doi-asserted-by":"crossref","unstructured":"Fathi, A., Hodgins, J.K., Rehg, J.M.: Social interactions: a first-person perspective. In 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1226\u20131233 (2012)","key":"4086_CR3","DOI":"10.1109\/CVPR.2012.6247805"},{"key":"4086_CR4","doi-asserted-by":"publisher","first-page":"1548","DOI":"10.1109\/TIP.2019.2941778","volume":"29","author":"H Zeng","year":"2019","unstructured":"Zeng, H., Cao, Z., Zhang, L., Bovik, A.C.: A unified probabilistic formulation of image aesthetic assessment. IEEE Trans. Image Process. 29, 1548\u20131561 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"4086_CR5","doi-asserted-by":"publisher","first-page":"3898","DOI":"10.1109\/TIP.2020.2968285","volume":"29","author":"L Li","year":"2020","unstructured":"Li, L., Zhu, H., Zhao, S., Ding, G., Lin, W.: Personality-assisted multi-task learning for generic and personalized image aesthetics assessment. IEEE Trans. Image Process. 29, 3898\u20133910 (2020)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"4086_CR6","doi-asserted-by":"publisher","first-page":"12","DOI":"10.1086\/209044","volume":"13","author":"AA Mitchell","year":"1986","unstructured":"Mitchell, A.A.: The effect of verbal and visual components of advertisements on brand attitudes and attitude toward the advertisement. J. Consum. Res. 13(1), 12\u201324 (1986)","journal-title":"J. Consum. Res."},{"issue":"2","key":"4086_CR7","doi-asserted-by":"publisher","first-page":"45","DOI":"10.1002\/mar.4220010206","volume":"1","author":"MB Holbrook","year":"1984","unstructured":"Holbrook, M.B., O\u2019Shaughnessy, J.: The role of emotion in advertising. Psychol. Mark. 1(2), 45\u201364 (1984)","journal-title":"Psychol. Mark."},{"issue":"9","key":"4086_CR8","doi-asserted-by":"publisher","first-page":"1207","DOI":"10.1016\/j.cortex.2011.06.006","volume":"48","author":"MJ Wieser","year":"2012","unstructured":"Wieser, M.J., Klupp, E., Weyers, P., Pauli, P., Weise, D., Zeller, D., Classen, J., M\u00fchlberger, A.: Reduced early visual emotion discrimination as an index of diminished emotion processing in parkinson\u2019s disease?-evidence from event-related brain potentials. Cortex 48(9), 1207\u20131217 (2012)","journal-title":"Cortex"},{"doi-asserted-by":"crossref","unstructured":"Jiang, M., Zhao, Q.: Learning visual attention to identify people with autism spectrum disorder. In 2017 IEEE International Conference on Computer Vision (ICCV), pp. 3287\u20133296 (2017)","key":"4086_CR9","DOI":"10.1109\/ICCV.2017.354"},{"issue":"4","key":"4086_CR10","doi-asserted-by":"publisher","first-page":"470","DOI":"10.1016\/j.giq.2012.06.005","volume":"29","author":"P Sobkowicz","year":"2012","unstructured":"Sobkowicz, P., Kaschesky, M., Bouchard, G.: Opinion mining in social media: modeling, simulating, and forecasting political opinions in the web. Gov. Inf. Q. 29(4), 470\u2013479 (2012)","journal-title":"Gov. Inf. Q."},{"key":"4086_CR11","doi-asserted-by":"publisher","first-page":"6939","DOI":"10.1007\/s11042-018-6445-z","volume":"78","author":"Z Li","year":"2019","unstructured":"Li, Z., Fan, Y., Jiang, B., Lei, T., Liu, W.: A survey on sentiment analysis and opinion mining for social multimedia. Multimed. Tools Appl. 78, 6939\u20136967 (2019)","journal-title":"Multimed. Tools Appl."},{"key":"4086_CR12","volume-title":"Tat Seng Chua, and Kurt Keutzer","author":"S Zhao","year":"2018","unstructured":"Zhao, S., Ding, G., Huang, Q.: Tat Seng Chua, and Kurt Keutzer. A comprehensive survey, Affective image content analysis (2018)"},{"issue":"10","key":"4086_CR13","doi-asserted-by":"publisher","first-page":"6729","DOI":"10.1109\/TPAMI.2021.3094362","volume":"44","author":"S Zhao","year":"2022","unstructured":"Zhao, S., Yao, X., Yang, J., Jia, G., Ding, G., Chua, T.-S., Schuller, B.W., Keutzer, K.: Affective image content analysis: two decades review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 44(10), 6729\u20136751 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"5","key":"4086_CR14","doi-asserted-by":"publisher","first-page":"1358","DOI":"10.1109\/TMM.2019.2939744","volume":"22","author":"D She","year":"2020","unstructured":"She, D., Yang, J., Cheng, M.-M., Lai, Y.-K., Rosin, P.L., Wang, L.: Wscnet: weakly supervised coupled networks for visual sentiment classification and detection. IEEE Trans. Multimed. 22(5), 1358\u20131371 (2020)","journal-title":"IEEE Trans. Multimed."},{"key":"4086_CR15","doi-asserted-by":"publisher","first-page":"7432","DOI":"10.1109\/TIP.2021.3106813","volume":"30","author":"J Yang","year":"2021","unstructured":"Yang, J., Li, J., Wang, X., Ding, Y., Gao, X.: Stimuli-aware visual emotion analysis. IEEE Trans. Image Process. 30, 7432\u20137445 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"4086_CR16","doi-asserted-by":"publisher","first-page":"7139","DOI":"10.1109\/TMM.2022.3217414","volume":"25","author":"S Lee","year":"2023","unstructured":"Lee, S., Ryu, C., Park, E.: Osanet: object semantic attention network for visual sentiment analysis. IEEE Trans. Multimed. 25, 7139\u20137148 (2023)","journal-title":"IEEE Trans. Multimed."},{"doi-asserted-by":"crossref","unstructured":"Yao, J., Chen, J., Niu, L., Sheng, B.: Scene-aware human pose generation using transformer. Proceedings of the 31st ACM International Conference on Multimedia (2023)","key":"4086_CR17","DOI":"10.1145\/3581783.3612439"},{"unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)","key":"4086_CR18"},{"doi-asserted-by":"crossref","unstructured":"You, Q., Luo, J., Jin, H., Yang, J.: Building a large scale dataset for image emotion recognition: The fine print and the benchmark. In Proceedings of the AAAI conference on artificial intelligence, vol. 30 (2016)","key":"4086_CR19","DOI":"10.1609\/aaai.v30i1.9987"},{"key":"4086_CR20","doi-asserted-by":"publisher","first-page":"1894","DOI":"10.1109\/TMM.2023.3289762","volume":"26","author":"Y Luo","year":"2024","unstructured":"Luo, Y., Zhong, X., Zeng, M., Xie, J., Wang, S., Liu, G.: Cglf-net: image emotion recognition network by combining global self-attention features and local multiscale features. IEEE Trans. Multimed. 26, 1894\u20131908 (2024)","journal-title":"IEEE Trans. Multimed."},{"unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S. et\u00a0al.: An image is worth 16x16 words: Transformers for image recognition at scale. arxiv 2020. arXiv preprint arXiv:2010.11929 (2010)","key":"4086_CR21"},{"key":"4086_CR22","doi-asserted-by":"publisher","first-page":"5189","DOI":"10.1109\/TIP.2022.3193749","volume":"31","author":"J Yang","year":"2022","unstructured":"Yang, J., Li, J., Li, L., Wang, X., Ding, Y., Gao, X.: Seeking subjectivity in visual emotion distribution learning. IEEE Trans. Image Process. 31, 5189\u20135202 (2022)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Mikels, J.A., Fredrickson, B.L., Larkin, G.R., Lindberg, C.M., Maglio, S.J., Reuter-Lorenz, P.A.: Emotional category data on images from the international affective picture system (2005)","key":"4086_CR23","DOI":"10.3758\/BF03192732"},{"issue":"3\u20134","key":"4086_CR24","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1080\/02699939208411068","volume":"6","author":"Ekman and Paul","year":"1992","unstructured":"Ekman and Paul: An argument for basic emotions. Cogn. Emot. 6(3\u20134), 169\u2013200 (1992)","journal-title":"Cogn. Emot."},{"issue":"2","key":"4086_CR25","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1037\/h0054570","volume":"61","author":"H Schlosberg","year":"1954","unstructured":"Schlosberg, H.: Three dimensions of emotion. Psychol. Rev. 61(2), 81\u20138 (1954)","journal-title":"Psychol. Rev."},{"issue":"5","key":"4086_CR26","doi-asserted-by":"publisher","first-page":"1031","DOI":"10.1109\/TMM.2011.2158530","volume":"13","author":"J Lee","year":"2011","unstructured":"Lee, J., Park, E.: Fuzzy similarity-based emotional classification of color images. IEEE Trans. Multimed. 13(5), 1031\u20131039 (2011)","journal-title":"IEEE Trans. Multimed."},{"doi-asserted-by":"crossref","unstructured":"Machajdik, J., Hanbury, A.: Affective image classification using features inspired by psychology and art theory. In Proceedings of the 18th ACM international conference on Multimedia, pp. 83\u201392 (2010)","key":"4086_CR27","DOI":"10.1145\/1873951.1873965"},{"key":"4086_CR28","doi-asserted-by":"publisher","first-page":"3534","DOI":"10.1109\/ICSMC.2006.384667","volume":"4","author":"W Wei-ning","year":"2006","unstructured":"Wei-ning, W., Ying-lin, Y., Sheng-ming, J.: Image retrieval by emotional semantics: a study of emotional space and feature extraction. 2006 IEEE Int. Conf. Syst. Man Cybernet. 4, 3534\u20133539 (2006)","journal-title":"2006 IEEE Int. Conf. Syst. Man Cybernet."},{"doi-asserted-by":"crossref","unstructured":"Yanulevskaya, V., van Gemert, J.C., Roth, K., Herbold, A.K., Sebe, N., Geusebroek, J.M.: Emotional valence categorization using holistic image features. In 2008 15th IEEE International Conference on Image Processing, pp. 101\u2013104 (2008)","key":"4086_CR29","DOI":"10.1109\/ICIP.2008.4711701"},{"key":"4086_CR30","doi-asserted-by":"publisher","first-page":"584","DOI":"10.1038\/s41591-023-02702-z","volume":"30","author":"L Dai","year":"2024","unstructured":"Dai, L., Sheng, B., Chen, T., Qiang, W., Liu, R., Cai, C., Liang, W., Yang, D., Hamzah, H., Liu, Y., Wang, X., Guan, Z., Shujie, Y., Li, T., Tang, Z., Ran, A.R., Che, H., Chen, H., Zheng, Y., Shu, J., Huang, S., Chan, W., Lin, S., Liu, D., Li, J., Wang, Z., Meng, Z., Shen, J., Hou, X., Deng, C., Ruan, L., Feng, L., Chee, M.L., Quek, T.C., Srinivasan, R., Raman, R., Sun, X., Wang, Y.X., Jiarui, W., Jin, H., Dai, R., Shen, D., Yang, X., Guo, M., Zhang, C., Cheung, C.Y., Tan, G., Tham, Y.-C., Cheng, C.-Y., Li, H., Wong, T.Y., Jia, W.: A deep learning system for predicting time to progression of diabetic retinopathy. Nat. Med. 30, 584\u2013594 (2024)","journal-title":"Nat. Med."},{"key":"4086_CR31","doi-asserted-by":"publisher","first-page":"7192","DOI":"10.1109\/TIP.2020.2999854","volume":"29","author":"A Nazir","year":"2020","unstructured":"Nazir, A., Cheema, M.N., Sheng, B., Li, H., Li, P., Yang, P., Jung, Y., Qin, J., Kim, J., Feng, D.D.: Off-enet: an optimally fused fully end-to-end network for automatic dense volumetric 3d intracranial blood vessels segmentation. IEEE Trans. Image Process. 29, 7192\u20137202 (2020)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Dai, L., Wu, L., Li, H., Cai, C., Wu, Q., Kong, H., Liu, R., Wang, X., Hou, X., Liu, Y., Long, X., Wen, Y., Lu, L., Shen, Y., Chen, Y., Shen, D., Yang, X., Zou, H., Sheng, B., Jia, W.: A deep learning system for detecting diabetic retinopathy across the disease spectrum. Nat. Commun. 12 (2021)","key":"4086_CR32","DOI":"10.1038\/s41467-021-23458-5"},{"doi-asserted-by":"crossref","unstructured":"Qin, Y., Zhao, N., Yang, J., Pan, S., Sheng, B., Lau, R.W.H.: Urbanevolver: function-aware urban layout regeneration. Int. J. Comput. Vis. (2024)","key":"4086_CR33","DOI":"10.1007\/s11263-024-02030-w"},{"issue":"1","key":"4086_CR34","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1109\/TII.2021.3085669","volume":"18","author":"J Li","year":"2022","unstructured":"Li, J., Chen, J., Sheng, B., Li, P., Yang, P., Feng, D.D., Qi, J.: Automatic detection and classification system of domestic waste via multimodel cascaded convolutional neural network. IEEE Trans. Ind. Inform. 18(1), 163\u2013173 (2022)","journal-title":"IEEE Trans. Ind. Inform."},{"key":"4086_CR35","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2023.3244837","volume":"72","author":"AS Tomar","year":"2023","unstructured":"Tomar, A.S., Arya, K.V., Rajput, S.S.: Deep hyfeat based attention in attention model for face super-resolution. IEEE Trans. Instrum. Meas. 72, 1\u201311 (2023)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"4086_CR36","doi-asserted-by":"publisher","first-page":"2226","DOI":"10.1109\/TMM.2022.3144890","volume":"25","author":"N Jiang","year":"2023","unstructured":"Jiang, N., Sheng, B., Li, P., Lee, T.-Y.: Photohelper: portrait photographing guidance via deep feature retrieval and fusion. IEEE Trans. Multimed. 25, 2226\u20132238 (2023)","journal-title":"IEEE Trans. Multimed."},{"issue":"5","key":"4086_CR37","doi-asserted-by":"publisher","first-page":"e2141","DOI":"10.1002\/cav.2141","volume":"34","author":"Y Shuqing","year":"2023","unstructured":"Shuqing, Y., Wang, Z., Zhou, S., Yang, X., Chao, W., Wang, Z.: Perimetrynet: a multiscale fine grained deep network for three-dimensional eye gaze estimation using visual field analysis. Comput. Anim. Virtual Worlds 34(5), e2141 (2023)","journal-title":"Comput. Anim. Virtual Worlds"},{"issue":"C","key":"4086_CR38","doi-asserted-by":"publisher","first-page":"58","DOI":"10.1016\/j.patrec.2023.03.025","volume":"169","author":"AS Tomar","year":"2023","unstructured":"Tomar, A.S., Arya, K.V., Rajput, S.S.: Attentive exfeat based deep generative adversarial network for noise robust face super-resolution. Pattern Recogn. Lett. 169(C), 58\u201366 (2023)","journal-title":"Pattern Recogn. Lett."},{"issue":"16","key":"4086_CR39","doi-asserted-by":"publisher","first-page":"25449","DOI":"10.1007\/s11042-023-14472-4","volume":"82","author":"AS Tomar","year":"2023","unstructured":"Tomar, A.S., Arya, K.V., Rajput, S.S.: Noise robust face super-resolution via learning of spatial attentive features. Multimed. Tools Appl. 82(16), 25449\u201325465 (2023)","journal-title":"Multimed. Tools Appl."},{"doi-asserted-by":"crossref","unstructured":"Li, H., Yuan, X., Xu, C., Zhang, R., Liu, X., Liu, L.: Complexity aware center loss for facial expression recognition. The Visual Computer (2024)","key":"4086_CR40","DOI":"10.1007\/s00371-023-03221-1"},{"key":"4086_CR41","doi-asserted-by":"publisher","first-page":"46","DOI":"10.1007\/s00530-023-01227-2","volume":"30","author":"W Yiqing","year":"2024","unstructured":"Yiqing, W., Li, D., Chen, X., Tang, Y., Huang, S.: An ensemble pruning method considering classifiers\u2019 interaction based on information theory for facial expression recognition. Multim. Syst. 30, 46 (2024)","journal-title":"Multim. Syst."},{"key":"4086_CR42","doi-asserted-by":"publisher","first-page":"105245","DOI":"10.1016\/j.knosys.2019.105245","volume":"191","author":"J Zhang","year":"2020","unstructured":"Zhang, J., Chen, M., Sun, H., Li, D., Wang, Z.: Object semantics sentiment correlation analysis enhanced image sentiment classification. Knowl.-Based Syst. 191, 105245 (2020)","journal-title":"Knowl.-Based Syst."},{"key":"4086_CR43","doi-asserted-by":"publisher","first-page":"8686","DOI":"10.1109\/TIP.2021.3118983","volume":"30","author":"J Yang","year":"2021","unstructured":"Yang, J., Gao, X., Li, L., Wang, X., Ding, J.: Solver: scene-object interrelated visual emotion reasoning network. IEEE Trans. Image Process. 30, 8686\u20138701 (2021)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Rao, T., Li, X., Zhang, H., Xu, M.: Multi-level region-based convolutional neural network for image emotion classification. Neurocomputing (2019)","key":"4086_CR44","DOI":"10.1016\/j.neucom.2018.12.053"},{"issue":"5","key":"4086_CR45","doi-asserted-by":"publisher","first-page":"2177","DOI":"10.1007\/s00371-022-02472-8","volume":"39","author":"H Yang","year":"2023","unstructured":"Yang, H., Fan, Y., Lv, G., Liu, S., Guo, Z.: Exploiting emotional concepts for image emotion recognition. Vis. Comput. 39(5), 2177\u20132190 (2023)","journal-title":"Vis. Comput."},{"issue":"3\u20134","key":"4086_CR46","doi-asserted-by":"publisher","first-page":"e2160","DOI":"10.1002\/cav.2160","volume":"34","author":"A Che","year":"2023","unstructured":"Che, A., Yang, J.-H., Guo, C., Dai, H.-N., Xie, H., Li, P.: Aegan: generating imperceptible face synthesis via autoencoder-based generative adversarial network. Comput. Anim. Vir. Worlds 34(3\u20134), e2160 (2023)","journal-title":"Comput. Anim. Vir. Worlds"},{"issue":"3","key":"4086_CR47","doi-asserted-by":"publisher","first-page":"e2275","DOI":"10.1002\/cav.2275","volume":"35","author":"S Feng","year":"2024","unstructured":"Feng, S., Hou, F., Chen, J., Wang, W.: Extracting roads from satellite images via enhancing road feature investigation in learning. Comput. Anim. Vir. Worlds 35(3), e2275 (2024)","journal-title":"Comput. Anim. Vir. Worlds"},{"key":"4086_CR48","doi-asserted-by":"publisher","first-page":"06","DOI":"10.1155\/2024\/7321175","volume":"2024","author":"H Li","year":"2024","unstructured":"Li, H., Xiao, X., Liu, X., Wen, G., Liu, L.: Learning cognitive features as complementary for facial expression recognition. Int. J. Intell. Syst. 2024, 06 (2024)","journal-title":"Int. J. Intell. Syst."},{"key":"4086_CR49","doi-asserted-by":"publisher","first-page":"125625","DOI":"10.1016\/j.eswa.2024.125625","volume":"262","author":"AS Tomar","year":"2025","unstructured":"Tomar, A.S., Arya, K.V., Rajput, S.S.: Learning face super-resolution through identity features and distilling facial prior knowledge. Expert Syst. Appl. 262, 125625 (2025)","journal-title":"Expert Syst. Appl."},{"issue":"2","key":"4086_CR50","doi-asserted-by":"publisher","first-page":"1763","DOI":"10.1007\/s10462-022-10212-6","volume":"56","author":"Z Li","year":"2023","unstructured":"Li, Z., Huibin, L., Zhao, C., Feng, L., Guanghua, G., Chen, W.: Weakly supervised discriminate enhancement network for visual sentiment analysis. Artif. Intell. Rev. 56(2), 1763\u20131785 (2023)","journal-title":"Artif. Intell. Rev."},{"unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S. et\u00a0al.: An image is worth 16x16 words: Transformers for image recognition at scale. arxiv 2020. arXiv preprint arXiv:2010.11929 (2010)","key":"4086_CR51"},{"unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS\u201917, page 6000\u20136010, Red Hook, NY, USA, (2017). Curran Associates Inc","key":"4086_CR52"},{"unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u2019egou, H.: Training data-efficient image transformers & distillation through attention. In International Conference on Machine Learning (2020)","key":"4086_CR53"},{"doi-asserted-by":"crossref","unstructured":"Wang, W., Xie, E., Li, X., Fan, D.-P., Song, K., Liang, D., Lu, T., Luo, P., Shao, L.: Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. In 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 548\u2013558 (2021)","key":"4086_CR54","DOI":"10.1109\/ICCV48922.2021.00061"},{"doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: hierarchical vision transformer using shifted windows. 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 9992\u201310002 (2021)","key":"4086_CR55","DOI":"10.1109\/ICCV48922.2021.00986"},{"doi-asserted-by":"crossref","unstructured":"Liu, C., Zhao, S., Luo, Y., Liu, G.: Transiea: transformer-baseartd image emotion analysis. In 2022 7th International Conference on Computer and Communication Systems (ICCCS), pp. 310\u2013313 (2022)","key":"4086_CR56","DOI":"10.1109\/ICCCS55155.2022.9846146"},{"key":"4086_CR57","doi-asserted-by":"publisher","first-page":"103968","DOI":"10.1016\/j.jvcir.2023.103968","volume":"97","author":"X Wang","year":"2023","unstructured":"Wang, X., Yang, J., Min, H., Ren, F.: Eerca-vit: enhanced effective region and context-aware vision transformers for image sentiment analysis. J. Vis. Commun. Image Represent. 97, 103968 (2023)","journal-title":"J. Vis. Commun. Image Represent."},{"unstructured":"Weston, J., Chopra, S., Bordes, A.: Memory networks. arXiv preprint arXiv:1410.3916 (2014)","key":"4086_CR58"},{"unstructured":"Graves, A., Wayne, G., Danihelka, I.: Neural turing machines. arXiv preprint arXiv:1410.5401 (2014)","key":"4086_CR59"},{"unstructured":"Sukhbaatar, S., Weston, J., Fergus, R. et\u00a0al.: End-to-end memory networks. Adv. Neural Inf. Process. Syst. 28 (2015)","key":"4086_CR60"},{"issue":"5","key":"4086_CR61","doi-asserted-by":"publisher","first-page":"2530","DOI":"10.1109\/TIP.2018.2887017","volume":"28","author":"Z Wang","year":"2019","unstructured":"Wang, Z., Yi, P., Jiang, K., Jiang, J., Han, Z., Tao, L., Ma, J.: Multi-memory convolutional neural network for video super-resolution. IEEE Trans. Image Process. 28(5), 2530\u20132544 (2019)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Shi, P., Song, Y., Ma, C., Zhang, H., Yang, M.-H.: Learning recurrent memory activation networks for visual tracking. IEEE Trans. Image Process. 30, 725\u2013738 (2021)","key":"4086_CR62","DOI":"10.1109\/TIP.2020.3038356"},{"doi-asserted-by":"crossref","unstructured":"Xie, Z., Zhang, W., Sheng, B., Li, P., Philip Chen, C.L.: Bagfn: broad attentive graph fusion network for high-order feature interactions. IEEE Trans. Neural Netw. Learn. Syst. 34(8), 4499\u20134513 (2023)","key":"4086_CR63","DOI":"10.1109\/TNNLS.2021.3116209"},{"doi-asserted-by":"crossref","unstructured":"Yang, J., She, D., Lai, Y.-K., Rosin, P.L., Yang, M.-H.: Weakly supervised coupled networks for visual sentiment analysis. In 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7584\u20137592 (2018)","key":"4086_CR64","DOI":"10.1109\/CVPR.2018.00791"},{"doi-asserted-by":"crossref","unstructured":"Peng, K.-C., Sadovnik, A., Gallagher, A., Chen, T.: Where do emotions come from? predicting the emotion stimuli map. In 2016 IEEE International Conference on Image Processing (ICIP), pp. 614\u2013618 (2016)","key":"4086_CR65","DOI":"10.1109\/ICIP.2016.7532430"},{"doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2009)","key":"4086_CR66","DOI":"10.1109\/CVPR.2009.5206848"},{"doi-asserted-by":"crossref","unstructured":"He, T., Zhang, Z., Zhang, H., Zhang, Z., Xie, J., Li, M.: Bag of tricks for image classification with convolutional neural networks. In 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 558\u2013567 (2019)","key":"4086_CR67","DOI":"10.1109\/CVPR.2019.00065"},{"key":"4086_CR68","doi-asserted-by":"publisher","first-page":"2043","DOI":"10.1007\/s11063-019-10033-9","volume":"51","author":"T Rao","year":"2020","unstructured":"Rao, T., Li, X., Min, X.: Learning multi-level deep representations for image emotion classification. Neural Process. Lett. 51, 2043\u20132061 (2020)","journal-title":"Neural Process. Lett."},{"doi-asserted-by":"crossref","unstructured":"Xu, L., Wang, Z., Wu, B., Lui, S.: Mdan: multi-level dependent attention network for visual emotion analysis. In 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9469\u20139478 (2022)","key":"4086_CR69","DOI":"10.1109\/CVPR52688.2022.00926"},{"key":"4086_CR70","doi-asserted-by":"publisher","first-page":"429","DOI":"10.1016\/j.neucom.2018.12.053","volume":"333","author":"T Rao","year":"2019","unstructured":"Rao, T., Li, X., Zhang, H., Min, X.: Multi-level region-based convolutional neural network for image emotion classification. Neurocomputing 333, 429\u2013439 (2019)","journal-title":"Neurocomputing"},{"doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: visual explanations from deep networks via gradient-based localization. In 2017 IEEE International Conference on Computer Vision (ICCV), pp. 618\u2013626 (2017)","key":"4086_CR71","DOI":"10.1109\/ICCV.2017.74"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04086-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04086-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04086-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T14:03:15Z","timestamp":1758722595000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04086-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,12]]},"references-count":71,"journal-issue":{"issue":"13","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["4086"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04086-2","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2025,7,12]]},"assertion":[{"value":"19 June 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 July 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}