{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T23:30:42Z","timestamp":1773963042631,"version":"3.50.1"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"18","license":[{"start":{"date-parts":[[2023,12,1]],"date-time":"2023-12-01T00:00:00Z","timestamp":1701388800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,1]],"date-time":"2023-12-01T00:00:00Z","timestamp":1701388800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/fund","name":"National Research Foundation (NRF) of Korea","doi-asserted-by":"publisher","award":["2020R1A2C1012196"],"award-info":[{"award-number":["2020R1A2C1012196"]}],"id":[{"id":"10.13039\/fund","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-17601-1","type":"journal-article","created":{"date-parts":[[2023,12,1]],"date-time":"2023-12-01T10:02:47Z","timestamp":1701424967000},"page":"54249-54278","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":20,"title":["A context-sensitive multi-tier deep learning framework for multimodal sentiment analysis"],"prefix":"10.1007","volume":"83","author":[{"given":"Ganesh Kumar","family":"P","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5543-7547","authenticated-orcid":false,"given":"Arul Antran Vijay","family":"S","sequence":"additional","affiliation":[]},{"given":"Jothi Prakash","family":"V","sequence":"additional","affiliation":[]},{"given":"Anand","family":"Paul","sequence":"additional","affiliation":[]},{"given":"Anand","family":"Nayyar","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,1]]},"reference":[{"key":"17601_CR1","doi-asserted-by":"publisher","unstructured":"S\u00e1nchez-Rada JF, Iglesias, CA (2019) Social context in sentiment analysis: Formal definition, overview of current trends and framework for comparison. Inf Fus 52:344\u2013356. https:\/\/doi.org\/10.1016\/j.inffus.2019.05.003","DOI":"10.1016\/j.inffus.2019.05.003"},{"key":"17601_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2022\/3297316","volume":"2022","author":"HD Praveena","year":"2022","unstructured":"Praveena HD, Guptha NS, Kazemzadeh A, Parameshachari BD, Hemalatha KL (2022) Effective cbmir system using hybrid features-based independent condensed nearest neighbor model. J Health Eng 2022:1\u20139. https:\/\/doi.org\/10.1155\/2022\/3297316","journal-title":"J Health Eng"},{"key":"17601_CR3","doi-asserted-by":"publisher","unstructured":"Thouheed\u00a0Ahmed SS, Thanuja K,Guptha NS, Narasimha S (2016) Telemedicine approach for remote patient monitoring system using smart phones with an economical hardware kit. In: 2016 international conference on computing technologies and intelligent data engineering (ICCTIDE\u201916), pp 1\u20134. https:\/\/doi.org\/10.1109\/ICCTIDE.2016.7725324","DOI":"10.1109\/ICCTIDE.2016.7725324"},{"key":"17601_CR4","doi-asserted-by":"publisher","first-page":"256","DOI":"10.1109\/ICCTIDE.2016.7725324","volume":"11","author":"N Guptha","year":"2018","unstructured":"Guptha N, Patil K (2018) Detection of macro and micro nodule using online region based-active contour model in histopathological liver cirrhosis. Int J Intell Eng Syst 11:256\u2013265. https:\/\/doi.org\/10.1109\/ICCTIDE.2016.7725324","journal-title":"Int J Intell Eng Syst"},{"key":"17601_CR5","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1504\/IJSISE.2017.084568","volume":"10","author":"NS Guptha","year":"2017","unstructured":"Guptha NS, Patil KK (2017) Earth mover\u2019s distance-based cbir using adaptive regularised kernel fuzzy c-means method of liver cirrhosis histopathological segmentation. Int J Signal Imaging Syst Eng 10:39. https:\/\/doi.org\/10.1504\/IJSISE.2017.084568","journal-title":"Int J Signal Imaging Syst Eng"},{"key":"17601_CR6","doi-asserted-by":"publisher","unstructured":"Abd\u00a0El-Jawad MH, Hodhod R, Omar YMK (2018) Sentiment analysis of social media networks using machine learning. In: 2018 14th international computer engineering conference (ICENCO), pp 174\u2013176. https:\/\/doi.org\/10.1109\/ICENCO.2018.8636124","DOI":"10.1109\/ICENCO.2018.8636124"},{"key":"17601_CR7","doi-asserted-by":"crossref","unstructured":"Zadeh A, Chen M, Poria S, Cambria E, Morency L-P (2017) Tensor fusion network for multimodal sentiment analysis 1","DOI":"10.18653\/v1\/D17-1115"},{"key":"17601_CR8","doi-asserted-by":"publisher","unstructured":"Poria S, Cambria E, Gelbukh A (2015) Deep convolutional neural network textual features and multiple kernel learning for utterance-level multimodal sentiment analysis. In: Proceedings of the 2015 conference on empirical methods in natural language processing, pp 2539\u20132544. Assoc Comput Linguist, Lisbon, Portugal. https:\/\/doi.org\/10.18653\/v1\/D15-1303","DOI":"10.18653\/v1\/D15-1303"},{"key":"17601_CR9","doi-asserted-by":"publisher","first-page":"48","DOI":"10.1109\/MCI.2014.2307227","volume":"9","author":"E Cambria","year":"2014","unstructured":"Cambria E, White B (2014) Jumping nlp curves: A review of natural language processing research [review article]. IEEE computational intelligence magazine 9:48\u201357. https:\/\/doi.org\/10.1109\/MCI.2014.2307227","journal-title":"IEEE computational intelligence magazine"},{"key":"17601_CR10","doi-asserted-by":"publisher","unstructured":"Ahmed ST, Guptha NS, Lavanya NL, Basha SM, Fathima AS (2022) Improving medical image pixel quality using micq unsupervised machine learning technique. Malays J Comput Sci 53\u201364. https:\/\/doi.org\/10.22452\/mjcs.sp2022no2.5","DOI":"10.22452\/mjcs.sp2022no2.5"},{"key":"17601_CR11","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1016\/j.patrec.2022.04.038","volume":"159","author":"NS Guptha","year":"2011","unstructured":"Guptha NS, Balamurugan V, Megharaj G, Sattar KNA, Rose JD (2011) Cross lingual handwritten character recognition using long short term memory network with aid of elephant herding optimization algorithm. Pattern Recogn Lett 159:16\u201322. https:\/\/doi.org\/10.1016\/j.patrec.2022.04.038","journal-title":"Pattern Recogn Lett"},{"key":"17601_CR12","doi-asserted-by":"publisher","unstructured":"Liu Z, Shen Y, Lakshminarasimhan VB, Liang PP, Bagher\u00a0Zadeh A, Morency L-P (2018) Efficient low-rank multimodal fusion with modality-specific factors. In: Proceedings of the 56th annual meeting of the association for computational linguistics (Volume 1: Long Papers), pp 2247\u20132256. Assoc Comput Linguist, Melbourne, Australia. https:\/\/doi.org\/10.18653\/v1\/P18-1209","DOI":"10.18653\/v1\/P18-1209"},{"key":"17601_CR13","doi-asserted-by":"publisher","unstructured":"Hu G, Hua Y, Yuan Y, Zhang Z, Lu Z, Mukherjee SS, Hospedales TM, Robertson NM, Yang Y (2017) Attribute-enhanced face recognition with neural tensor fusion networks. In: 2017 IEEE international conference on computer vision (ICCV), pp 3764\u20133773. https:\/\/doi.org\/10.1109\/ICCV.2017.404","DOI":"10.1109\/ICCV.2017.404"},{"key":"17601_CR14","doi-asserted-by":"publisher","unstructured":"Chen M, Wang S, Liang PP, Baltru\u0161aitis T, Zadeh A, Morency L-P (2017) Multimodal sentiment analysis with word-level fusion and reinforcement learning. In: Proceedings of the 19th ACM international conference on multimodal interaction. ICMI \u201917, pp 163\u2013171. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/3136755.3136801","DOI":"10.1145\/3136755.3136801"},{"key":"17601_CR15","doi-asserted-by":"publisher","first-page":"164","DOI":"10.1609\/aaai.v34i01.5347","volume":"34","author":"S Mai","year":"2020","unstructured":"Mai S, Hu H, Xing S (2020) Modality to modality translation: An adversarial representation learning and graph fusion network for multimodal fusion. Proceedings of the AAAI conference on artificial intelligence 34:164\u2013172. https:\/\/doi.org\/10.1609\/aaai.v34i01.5347","journal-title":"Proceedings of the AAAI conference on artificial intelligence"},{"key":"17601_CR16","doi-asserted-by":"publisher","first-page":"2513","DOI":"10.1109\/TMM.2018.2803520","volume":"20","author":"J Yang","year":"2018","unstructured":"Yang J, She D, Sun M, Cheng M-M, Rosin PL, Wang L (2018) Visual sentiment prediction based on automatic discovery of affective regions. IEEE transactions on multimedia 20:2513\u20132525. https:\/\/doi.org\/10.1109\/TMM.2018.2803520","journal-title":"IEEE transactions on multimedia"},{"key":"17601_CR17","doi-asserted-by":"publisher","unstructured":"Poria S, Cambria E, Hazarika D, Majumder N, Zadeh A, Morency L-P (2017) Context-dependent sentiment analysis in user-generated videos. In: Proceedings of the 55th annual meeting of the association for computational linguistics (Volume 1: Long Papers), pp 873\u2013883. Assoc Comput Linguist, Vancouver, Canada. https:\/\/doi.org\/10.18653\/v1\/P17-1081","DOI":"10.18653\/v1\/P17-1081"},{"key":"17601_CR18","doi-asserted-by":"publisher","unstructured":"Gu Y, Li X, Huang K, Fu S, Yang K, Chen S, Zhou M, Marsic I (2018) Human conversation analysis using attentive multimodal networks with hierarchical encoder-decoder. In: Proceedings of the 26th ACM international conference on multimedia. MM \u201918, pp 537\u2013545. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/3240508.3240714","DOI":"10.1145\/3240508.3240714"},{"key":"17601_CR19","doi-asserted-by":"crossref","unstructured":"Akhtar MS, Chauhan DS, Ghosal D, Poria S, Ekbal A, Bhattacharyya P (2019) Multi-task learning for multi-modal emotion recognition and sentiment analysis 1","DOI":"10.18653\/v1\/N19-1034"},{"key":"17601_CR20","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1016\/j.knosys.2019.01.019","volume":"167","author":"F Huang","year":"2019","unstructured":"Huang F, Zhang X, Zhao Z, Xu J, Li Z (2019) Image-text sentiment analysis via deep multimodal attentive fusion. Knowledge-Based Systems 167:26\u201337. https:\/\/doi.org\/10.1016\/j.knosys.2019.01.019","journal-title":"Knowledge-Based Systems"},{"key":"17601_CR21","doi-asserted-by":"publisher","unstructured":"Pham H, Manzini T, Liang PP, Pocz\u00f3s B (2018) Seq2Seq2Sentiment: Multimodal sequence to sequence models for sentiment analysis. In: Proceedings of grand challenge and workshop on human multimodal language (Challenge-HML), pp 53\u201363. Assoc Comput Linguist, Melbourne, Australia. https:\/\/doi.org\/10.18653\/v1\/W18-3308","DOI":"10.18653\/v1\/W18-3308"},{"key":"17601_CR22","doi-asserted-by":"publisher","first-page":"6892","DOI":"10.1609\/aaai.v33i01.33016892","volume":"33","author":"H Pham","year":"2019","unstructured":"Pham H, Liang PP, Manzini T, Morency L-P, P\u00f3czos B (2019) Found in translation: Learning robust joint representations by cyclic translations between modalities. Proceedings of the AAAI conference on artificial intelligence 33:6892\u20136899. https:\/\/doi.org\/10.1609\/aaai.v33i01.33016892","journal-title":"Proceedings of the AAAI conference on artificial intelligence"},{"key":"17601_CR23","doi-asserted-by":"publisher","first-page":"8992","DOI":"10.1609\/aaai.v34i05.6431","volume":"34","author":"Z Sun","year":"2020","unstructured":"Sun Z, Sarma P, Sethares W, Liang Y (2020) Learning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis. Proceedings of the AAAI conference on artificial intelligence 34:8992\u20138999. https:\/\/doi.org\/10.1609\/aaai.v34i05.6431","journal-title":"Proceedings of the AAAI conference on artificial intelligence"},{"key":"17601_CR24","unstructured":"Andrew G, Arora R, Bilmes J, Livescu K (2013) Deep canonical correlation analysis. In: Dasgupta S, McAllester D (eds) Proceedings of the 30th international conference on machine learning. Proceedings of machine learning research, vol 28, pp 1247\u20131255. PMLR, Atlanta, Georgia, USA. https:\/\/proceedings.mlr.press\/v28\/andrew13.html"},{"key":"17601_CR25","unstructured":"Park G, Im W (2016) Image-text multi-modal representation learning by adversarial backpropagation 1"},{"key":"17601_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3284750","volume":"15","author":"Y Peng","year":"2019","unstructured":"Peng Y, Qi J (2019) Cm-gans. ACM Trans Multimed Comput Commun Appl 15:1\u201324. https:\/\/doi.org\/10.1145\/3284750","journal-title":"ACM Trans Multimed Comput Commun Appl"},{"key":"17601_CR27","doi-asserted-by":"publisher","first-page":"149077","DOI":"10.1109\/ACCESS.2021.3118537","volume":"9","author":"Q Zhu","year":"2021","unstructured":"Zhu Q, Jiang X, Ye R (2021) Sentiment analysis of review text based on bigru-attention and hybrid cnn. IEEE Access 9:149077\u2013149088. https:\/\/doi.org\/10.1109\/ACCESS.2021.3118537","journal-title":"IEEE Access"},{"key":"17601_CR28","doi-asserted-by":"publisher","first-page":"7819","DOI":"10.3934\/mbe.2020398","volume":"17","author":"Y Liu","year":"2020","unstructured":"Liu Y, Lu J, Yang J, Mao F (2020) Sentiment analysis for e-commerce product reviews by deep learning model of bert-bigru-softmax. Math Biosci Eng 17:7819\u20137837","journal-title":"Math Biosci Eng"},{"key":"17601_CR29","doi-asserted-by":"publisher","unstructured":"Stateczny A, Narahari SC, Vurubindi P, Guptha NS, Srinivas K (2023) Underground water level prediction in remote sensing images using improved hydro index value with ensemble classifier. Remote Sensing 15. https:\/\/doi.org\/10.3390\/rs15082015","DOI":"10.3390\/rs15082015"},{"key":"17601_CR30","doi-asserted-by":"publisher","unstructured":"Kim T, Lee B (2020) Multi-attention multimodal sentiment analysis. In: Proceedings of the 2020 international conference on multimedia retrieval. ICMR \u201920, pp 436\u2013441. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/3372278.3390698","DOI":"10.1145\/3372278.3390698"},{"key":"17601_CR31","doi-asserted-by":"publisher","unstructured":"Han W, Chen H, Gelbukh A, Zadeh A, Morency L-p, Poria S (2021) Bi-bimodal modality fusion for correlation-controlled multimodal sentiment analysis. In: Proceedings of the 2021 international conference on multimodal interaction. ICMI \u201921, pp 6\u201315. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/3462244.3479919","DOI":"10.1145\/3462244.3479919"},{"key":"17601_CR32","unstructured":"Mikolov T, Chen K, Corrado G, Dean J (2013) Efficient estimation of word representations in vector space 1"},{"key":"17601_CR33","doi-asserted-by":"publisher","unstructured":"Eyben F, W\u00f6llmer M, Schuller B (2010) Opensmile: The munich versatile and fast open-source audio feature extractor. In: Proceedings of the 18th ACM international conference on multimedia. MM \u201910, pp 1459\u20131462. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/1873951.1874246","DOI":"10.1145\/1873951.1874246"},{"key":"17601_CR34","doi-asserted-by":"crossref","unstructured":"Tran D, Bourdev L, Fergus R, Torresani L, Paluri M (2015) Learning spatiotemporal features with 3d convolutional networks, vol 1","DOI":"10.1109\/ICCV.2015.510"},{"key":"17601_CR35","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9:1735\u20131780. https:\/\/doi.org\/10.1162\/neco.1997.9.8.1735","journal-title":"Neural Comput"},{"key":"17601_CR36","doi-asserted-by":"crossref","unstructured":"Cho K, Merrienboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using rnn encoder-decoder for statistical machine translation 1","DOI":"10.3115\/v1\/D14-1179"},{"key":"17601_CR37","doi-asserted-by":"publisher","unstructured":"Ma Y, Peng H, Cambria E (2018) Targeted aspect-based sentiment analysis via embedding commonsense knowledge into an attentive lstm. Proceedings of the AAAI conference on artificial intelligence 32. https:\/\/doi.org\/10.1609\/aaai.v32i1.12048","DOI":"10.1609\/aaai.v32i1.12048"},{"key":"17601_CR38","doi-asserted-by":"crossref","unstructured":"Wang Y, Huang M, Zhu X, Zhao L (2016) Attention-based lstm for aspect-level sentiment classification 1:606\u2013615","DOI":"10.18653\/v1\/D16-1058"},{"key":"17601_CR39","doi-asserted-by":"publisher","unstructured":"Wang Z, Yang B (2020) Attention-based bidirectional long short-term memory networks for relation classification using knowledge distillation from bert. In: 2020 IEEE Intl Conf on Dependable, Autonomic and Secure Computing, pp 562\u2013568. https:\/\/doi.org\/10.1109\/DASC-PICom-CBDCom-CyberSciTech49142.2020.00100","DOI":"10.1109\/DASC-PICom-CBDCom-CyberSciTech49142.2020.00100"},{"key":"17601_CR40","unstructured":"Chung J, Gulcehre C, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling 1"},{"key":"17601_CR41","doi-asserted-by":"publisher","unstructured":"Borth D, Ji R, Chen T, Breuel T, Chang S-F (2013) Large-scale visual sentiment ontology and detectors using adjective noun pairs. In: Proceedings of the 21st ACM international conference on multimedia. MM \u201913, pp 223\u2013232. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/2502081.2502282","DOI":"10.1145\/2502081.2502282"},{"key":"17601_CR42","doi-asserted-by":"publisher","unstructured":"Niu T, Zhu S, Pang L, Saddik AE (2016) Sentiment Analysis on Multi-View Social Data. https:\/\/doi.org\/10.1007\/978-3-319-27674-8_2","DOI":"10.1007\/978-3-319-27674-8_2"},{"key":"17601_CR43","doi-asserted-by":"publisher","unstructured":"You Q, Luo J, Jin H, Yang J (2016) Cross-modality consistent regression for joint visual-textual sentiment analysis of social multimedia. In: Proceedings of the ninth ACM international conference on web search and data mining. WSDM \u201916, pp 13\u201322. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/2835776.2835779","DOI":"10.1145\/2835776.2835779"},{"key":"17601_CR44","doi-asserted-by":"crossref","unstructured":"Vadicamo L, Carrara F, Cimino A, Cresci S, Dell\u2019Orletta F, Falchi F, Tesconi M (2017) Cross-media learning for image sentiment analysis in the wild, vol 1","DOI":"10.1109\/ICCVW.2017.45"},{"key":"17601_CR45","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1109\/MIS.2016.94","volume":"31","author":"A Zadeh","year":"2016","unstructured":"Zadeh A, Zellers R, Pincus E, Morency L-P (2016) Multimodal sentiment intensity analysis in videos: Facial gestures and verbal messages. IEEE intelligent systems 31:82\u201388. https:\/\/doi.org\/10.1109\/MIS.2016.94","journal-title":"IEEE intelligent systems"},{"key":"17601_CR46","doi-asserted-by":"publisher","unstructured":"Yang K, Xu H, Gao K (2020) Cm-bert: Cross-modal bert for text-audio sentiment analysis. In: Proceedings of the 28th ACM international conference on multimedia. MM \u201920, pp 521\u2013528. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/3394171.3413690","DOI":"10.1145\/3394171.3413690"},{"key":"17601_CR47","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3388861","volume":"16","author":"F Huang","year":"2020","unstructured":"Huang F, Wei K, Weng J, Li Z (2020) Attention-based modality-gated networks for image-text sentiment analysis. ACM Transactions on Multimedia Computing, Communications, and Applications 16:1\u201319. https:\/\/doi.org\/10.1145\/3388861","journal-title":"ACM Transactions on Multimedia Computing, Communications, and Applications"},{"key":"17601_CR48","doi-asserted-by":"publisher","unstructured":"Xu N, Mao W, Chen G (2018) A co-memory network for multimodal sentiment analysis. In: The 41st International ACM SIGIR conference on research & development in information retrieval. SIGIR \u201918, pp 929\u2013932. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/3209978.3210093","DOI":"10.1145\/3209978.3210093"},{"key":"17601_CR49","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-47436-2_59","author":"T Jiang","year":"2020","unstructured":"Jiang T, Wang J, Liu Z, Ling Y (2020) Fusion-Extraction Network for Multimodal Sentiment. Analysis. https:\/\/doi.org\/10.1007\/978-3-030-47436-2_59","journal-title":"Analysis."},{"key":"17601_CR50","doi-asserted-by":"publisher","unstructured":"Xu N, Mao W (2017) Multisentinet: A deep semantic network for multimodal sentiment analysis. In: Proceedings of the 2017 ACM on conference on information and knowledge management. CIKM \u201917, pp 2399\u20132402. Assoc Comput Mach, New York, NY, USA. https:\/\/doi.org\/10.1145\/3132847.3133142","DOI":"10.1145\/3132847.3133142"},{"key":"17601_CR51","doi-asserted-by":"publisher","unstructured":"Rahman W, Hasan MK, Lee S, Bagher\u00a0Zadeh A, Mao C, Morency L-P, Hoque E (2020) Integrating multimodal information in large pretrained transformers. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 2359\u20132369. Assoc Comput Linguist, Online. https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.214, https:\/\/aclanthology.org\/2020.acl-main.214","DOI":"10.18653\/v1\/2020.acl-main.214"},{"key":"17601_CR52","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1016\/j.knosys.2019.04.018","volume":"178","author":"J Xu","year":"2019","unstructured":"Xu J, Huang F, Zhang X, Wang S, Li C, Li Z, He Y (2019) Visual-textual sentiment classification with bi-directional multi-level attention networks. Knowledge-Based Systems 178:61\u201373. https:\/\/doi.org\/10.1016\/j.knosys.2019.04.018","journal-title":"Knowledge-Based Systems"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-17601-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-17601-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-17601-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,15]],"date-time":"2024-05-15T10:32:29Z","timestamp":1715769149000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-17601-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,1]]},"references-count":52,"journal-issue":{"issue":"18","published-online":{"date-parts":[[2024,5]]}},"alternative-id":["17601"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-17601-1","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,1]]},"assertion":[{"value":"26 May 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 October 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 October 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 December 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare there is no conflicts of interest regarding the publication of this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}