{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T20:56:31Z","timestamp":1761598591932,"version":"3.35.0"},"reference-count":26,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2025,2,3]],"date-time":"2025-02-03T00:00:00Z","timestamp":1738540800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,2,3]],"date-time":"2025-02-03T00:00:00Z","timestamp":1738540800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SN COMPUT. SCI."],"DOI":"10.1007\/s42979-025-03670-6","type":"journal-article","created":{"date-parts":[[2025,2,3]],"date-time":"2025-02-03T06:59:48Z","timestamp":1738565988000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Optimization for Image Sentiment Analysis Using Novel Dual Moth Flame Algorithm"],"prefix":"10.1007","volume":"6","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4735-6218","authenticated-orcid":false,"given":"Siddhi","family":"Kadu","sequence":"first","affiliation":[]},{"given":"Bharti","family":"Joshi","sequence":"additional","affiliation":[]},{"given":"Pratik","family":"Agrawal","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,3]]},"reference":[{"key":"3670_CR1","doi-asserted-by":"crossref","unstructured":"Ortis A, Farinella GM, Battiato S. Survey on visual sentiment analysis. IET Image Process. 2020;14(8):1440\u201356.","DOI":"10.1049\/iet-ipr.2019.1270"},{"key":"3670_CR2","doi-asserted-by":"crossref","unstructured":"She D, Yang J, Cheng MM, Lai YK, Rosin PL, Wang L. WSCNet: Weakly supervised coupled networks for visual sentiment classification and detection. IEEE Trans Multimed. 2019;22(5):1358\u201371.","DOI":"10.1109\/TMM.2019.2939744"},{"key":"3670_CR3","doi-asserted-by":"crossref","unstructured":"Xu J, Li Z, Huang F, Li C, Philip SY. Visual sentiment analysis with social relations-guided multiattention networks. IEEE Trans Cybern. 2020;52(6):4472\u201384.","DOI":"10.1109\/TCYB.2020.3027766"},{"key":"3670_CR4","doi-asserted-by":"crossref","unstructured":"Akhand MA, Roy S, Siddique N, Kamal MA, Shimamura T. Facial emotion recognition using transfer learning in the deep CNN. Electronics. 2021;10(9):1036.","DOI":"10.3390\/electronics10091036"},{"key":"3670_CR5","doi-asserted-by":"crossref","unstructured":"Urolagin S, Nayak J, Acharya UR. Gabor CNN based intelligent system for visual sentiment analysis of social media data on cloud environment. IEEE Access. 2022;10:132455\u201371.","DOI":"10.1109\/ACCESS.2022.3228263"},{"issue":"3","key":"3670_CR6","doi-asserted-by":"publisher","first-page":"2403","DOI":"10.1007\/s11063-020-10201-2","volume":"51","author":"W Zhuanghui","year":"2020","unstructured":"Zhuanghui W, Meng M, Jigang W. Visual sentiment prediction with attribute augmentation and multi-attention mechanism. Neural Process Lett. 2020;51(3):2403\u201316.","journal-title":"Neural Process Lett"},{"issue":"3","key":"3670_CR7","doi-asserted-by":"publisher","first-page":"4303","DOI":"10.1109\/JSYST.2020.3026879","volume":"15","author":"T Zhou","year":"2020","unstructured":"Zhou T, Cao J, Zhu X, Liu B, Li S. Visual-textual sentiment analysis enhanced by hierarchical cross-modality interaction. IEEE Syst J. 2020;15(3):4303\u201314.","journal-title":"IEEE Syst J"},{"key":"3670_CR8","doi-asserted-by":"crossref","unstructured":"Lee S, Han DK, Ko H. Multimodal emotion recognition fusion analysis adapting BERT with heterogeneous feature unification. IEEE Access. 2021;9:94557\u201372.","DOI":"10.1109\/ACCESS.2021.3092735"},{"key":"3670_CR9","doi-asserted-by":"crossref","unstructured":"Meena G, Mohbey KK, Kumar S. Sentiment analysis on images using convolutional neural networks based Inception-V3 transfer learning approach. Int J Inform Manag Data Insights. 2023;3(1):100174.","DOI":"10.1016\/j.jjimei.2023.100174"},{"key":"3670_CR10","doi-asserted-by":"crossref","unstructured":"Boughida A, Kouahla MN, Lafifi Y. A novel approach for facial expression recognition based on Gabor filters and genetic algorithm. Evol Syst. 2022;13(2):331\u201345.","DOI":"10.1007\/s12530-021-09393-2"},{"key":"3670_CR11","unstructured":"Fallahzadeh MR, Farokhi F, Harimi A, Sabbaghi-Nadooshan R. Facial expression recognition based on image gradient and deep convolutional neural network. J AI Data Min. 2021;9(2):259\u201368."},{"key":"3670_CR12","doi-asserted-by":"publisher","first-page":"24321","DOI":"10.1109\/ACCESS.2019.2900231","volume":"7","author":"W Hua","year":"2019","unstructured":"Hua W, Dai F, Huang L, Xiong J, Gui G. Hero: Human emotions recognition for realizing intelligent internet of things. IEEE Access. 2019;7:24321\u201332.","journal-title":"IEEE Access"},{"issue":"2","key":"3670_CR13","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/s00371-019-01630-9","volume":"36","author":"A Agrawal","year":"2020","unstructured":"Agrawal A, Mittal N. Using cnn for facial expression recognition: a study of the effects of kernel size and number of filters on accuracy. Vis Comput. 2020;36(2):405\u201312.","journal-title":"Vis Comput"},{"key":"3670_CR14","doi-asserted-by":"publisher","first-page":"47070","DOI":"10.1109\/ACCESS.2023.3274744","volume":"11","author":"F Alzamzami","year":"2023","unstructured":"Alzamzami F, El Saddik A. Transformer-based feature fusion approach for multimodal visual sentiment recognition using tweets in the wild. IEEE Access. 2023;11:47070\u20139.","journal-title":"IEEE Access"},{"key":"3670_CR15","first-page":"1243","volume":"25","author":"S Yuting","year":"2022","unstructured":"Yuting S, Zhao W, Jing P, Nie L. Exploiting low-rank latent gaussian graphical model estimation for visual sentiment distributions. IEEE Trans Multimedia. 2022;25:1243\u201355.","journal-title":"IEEE Trans Multimedia"},{"issue":"3","key":"3670_CR16","doi-asserted-by":"publisher","first-page":"1022","DOI":"10.1109\/TPAMI.2019.2944808","volume":"43","author":"J Kossaifi","year":"2019","unstructured":"Kossaifi J, Walecki R, Panagakis Y, Shen J, Schmitt M, Ringeval F, Han J, Pandit V, Toisoul A, Schuller B, et al. Sewa db: A rich database for audio-visual emotion and sentiment research in the wild. IEEE Trans Pattern Anal Mach Intell. 2019;43(3):1022\u201340.","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"3670_CR17","doi-asserted-by":"publisher","first-page":"1334","DOI":"10.1109\/TAFFC.2021.3097002","volume":"14","author":"L Stappen","year":"2021","unstructured":"Stappen L, Baird A, Schumann L, Schuller B. The multimodal sentiment analysis in car reviews (muse-car) dataset: Collection, insights and improvements. IEEE Trans Affect Comput. 2021;14(2):1334\u201350.","journal-title":"IEEE Trans Affect Comput"},{"key":"3670_CR18","doi-asserted-by":"publisher","first-page":"992","DOI":"10.1109\/LSP.2021.3078074","volume":"28","author":"J He","year":"2021","unstructured":"He J, Mai S, Haifeng H. A unimodal reinforced transformer with time squeeze fusion for multimodal sentiment analysis. IEEE Signal Process Lett. 2021;28:992\u20136.","journal-title":"IEEE Signal Process Lett"},{"key":"3670_CR19","doi-asserted-by":"crossref","unstructured":"Qayyum H, Majid M, Anwar SM, Khan B. Facial expression recognition using stationary wavelet transform features. Math Probl Eng. 2017;2017(1):9854050.","DOI":"10.1155\/2017\/9854050"},{"issue":"4","key":"3670_CR20","doi-asserted-by":"publisher","first-page":"10951","DOI":"10.1007\/s11042-023-16074-6","volume":"83","author":"D Avc\u0131","year":"2024","unstructured":"Avc\u0131 D, Sert E, \u00d6zyurt F, Avc\u0131 E. Mfif-dwt-cnn: Multi-focus \u0131mage fusion based on discrete wavelet transform with deep convolutional neural network. Multimed Tools Appl. 2024;83(4):10951\u201368.","journal-title":"Multimed Tools Appl"},{"issue":"3","key":"3670_CR21","doi-asserted-by":"publisher","first-page":"1837","DOI":"10.3390\/app13031837","volume":"13","author":"N You","year":"2023","unstructured":"You N, Han L, Zhu D, Song W. Research on image denoising in edge detection based on wavelet transform. Appl Sci. 2023;13(3):1837.","journal-title":"Appl Sci"},{"key":"3670_CR22","doi-asserted-by":"crossref","unstructured":"Truong QT, Lauw HW. Concept-oriented transformers for visual sentiment analysis. In: Proceedings of the sixteenth ACM international conference on web search and data mining. 2023;1111\u20139.","DOI":"10.1145\/3539597.3570437"},{"issue":"4","key":"3670_CR23","doi-asserted-by":"publisher","first-page":"11299","DOI":"10.1007\/s11042-023-15861-5","volume":"83","author":"S Chatterjee","year":"2024","unstructured":"Chatterjee S, Saha D, Sen S, Oliva D, Sarkar R. Moth-flame optimization based deep feature selection for facial expression recognition using thermal images. Multimed Tools Appl. 2024;83(4):11299\u2013322.","journal-title":"Multimed Tools Appl"},{"key":"3670_CR24","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1016\/j.knosys.2015.07.006","volume":"89","author":"S Mirjalili","year":"2015","unstructured":"Mirjalili S. Moth-flame optimization algorithm: a novel nature-inspired heuristic paradigm. Knowl-Based Syst. 2015;89:228\u201349.","journal-title":"Knowl-Based Syst"},{"key":"3670_CR25","unstructured":"Lyons MJ, Shigeru A, Miyuki K, Jiro G, Julien B. The Japanese female facial expression (jaffe) database. In: Proceedings of third international conference on automatic face and gesture recognition. 1998;14\u20136."},{"key":"3670_CR26","doi-asserted-by":"crossref","unstructured":"Goodfellow IJ, Erhan D, Carrier PL, Courville A, Mirza M, Hamner B, Cukierski W, Tang Y, Thaler D, Lee DH, Zhou Y. Challenges in representation learning: A report on three machine learning contests. In: Neural information processing: 20th international conference, ICONIP 2013, Daegu, Korea, November 3\u20137, 2013. Proceedings, Part III 20. 2013;117\u201324. Springer, Berlin, Heidelberg.","DOI":"10.1007\/978-3-642-42051-1_16"}],"container-title":["SN Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-025-03670-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s42979-025-03670-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-025-03670-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,3]],"date-time":"2025-02-03T06:59:59Z","timestamp":1738565999000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s42979-025-03670-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,3]]},"references-count":26,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2025,2]]}},"alternative-id":["3670"],"URL":"https:\/\/doi.org\/10.1007\/s42979-025-03670-6","relation":{},"ISSN":["2661-8907"],"issn-type":[{"value":"2661-8907","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,3]]},"assertion":[{"value":"21 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 January 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 February 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Siddhi Kadu declares that she has no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Informed consent was obtained from all individual participants included in the study.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}},{"value":"This article does not contain any studies with human participants or animals performed by any of the authors.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Research involving human and\/or animals"}}],"article-number":"134"}}