{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,6]],"date-time":"2026-01-06T02:20:52Z","timestamp":1767666052369,"version":"3.37.3"},"reference-count":40,"publisher":"Wiley","license":[{"start":{"date-parts":[[2020,11,3]],"date-time":"2020-11-03T00:00:00Z","timestamp":1604361600000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Sichuan Science and Technology Program","award":["2018JY0273","2019YJ0532","18ZA0409","KYTZ201708","201908510026"],"award-info":[{"award-number":["2018JY0273","2019YJ0532","18ZA0409","KYTZ201708","201908510026"]}]},{"DOI":"10.13039\/501100004884","name":"Education Department of Sichuan Province","doi-asserted-by":"publisher","award":["2018JY0273","2019YJ0532","18ZA0409","KYTZ201708","201908510026"],"award-info":[{"award-number":["2018JY0273","2019YJ0532","18ZA0409","KYTZ201708","201908510026"]}],"id":[{"id":"10.13039\/501100004884","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Scientific Research Foundation of CUIT","award":["2018JY0273","2019YJ0532","18ZA0409","KYTZ201708","201908510026"],"award-info":[{"award-number":["2018JY0273","2019YJ0532","18ZA0409","KYTZ201708","201908510026"]}]},{"DOI":"10.13039\/501100004543","name":"China Scholarship Council","doi-asserted-by":"publisher","award":["2018JY0273","2019YJ0532","18ZA0409","KYTZ201708","201908510026"],"award-info":[{"award-number":["2018JY0273","2019YJ0532","18ZA0409","KYTZ201708","201908510026"]}],"id":[{"id":"10.13039\/501100004543","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Computational Intelligence and Neuroscience"],"published-print":{"date-parts":[[2020,11,3]]},"abstract":"<jats:p>Text sentiment classification is an essential research field of natural language processing. Recently, numerous deep learning-based methods for sentiment classification have been proposed and achieved better performances compared with conventional machine learning methods. However, most of the proposed methods ignore the interactive relationship between contextual semantics and sentimental tendency while modeling their text representation. In this paper, we propose a novel Interactive Dual Attention Network (IDAN) model that aims to interactively learn the representation between contextual semantics and sentimental tendency information. Firstly, we design an algorithm that utilizes linguistic resources to obtain sentimental tendency information from text and then extract word embeddings from the BERT (Bidirectional Encoder Representations from Transformers) pretraining model as the embedding layer of IDAN. Next, we use two Bidirectional LSTM (BiLSTM) networks to learn the long-range dependencies of contextual semantics and sentimental tendency information, respectively. Finally, two types of attention mechanisms are implemented in IDAN. One is multihead attention, which is the next layer of BiLSTM and is used to learn the interactive relationship between contextual semantics and sentimental tendency information. The other is global attention that aims to make the model focus on the important parts of the sequence and generate the final representation for classification. These two attention mechanisms enable IDAN to interactively learn the relationship between semantics and sentimental tendency information and improve the classification performance. A large number of experiments on four benchmark datasets show that our IDAN model is superior to competitive methods. Moreover, both the result analysis and the attention weight visualization further demonstrate the effectiveness of our proposed method.<\/jats:p>","DOI":"10.1155\/2020\/8858717","type":"journal-article","created":{"date-parts":[[2020,11,4]],"date-time":"2020-11-04T17:35:12Z","timestamp":1604511312000},"page":"1-11","source":"Crossref","is-referenced-by-count":20,"title":["Interactive Dual Attention Network for Text Sentiment Classification"],"prefix":"10.1155","volume":"2020","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9633-3077","authenticated-orcid":true,"given":"Yinglin","family":"Zhu","sequence":"first","affiliation":[{"name":"College of Software Engineering, Chengdu University of Information Technology, Chengdu 610225, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9183-796X","authenticated-orcid":true,"given":"Wenbin","family":"Zheng","sequence":"additional","affiliation":[{"name":"College of Software Engineering, Chengdu University of Information Technology, Chengdu 610225, China"},{"name":"Software Automatic Generation and Intelligent Service Key Laboratory of Sichuan Province, Chengdu 610225, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9691-9369","authenticated-orcid":true,"given":"Hong","family":"Tang","sequence":"additional","affiliation":[{"name":"College of Engineering, Sichuan Normal University, Chengdu 610068, China"}]}],"member":"311","reference":[{"key":"1","doi-asserted-by":"publisher","DOI":"10.2200\/S00416ED1V01Y201204HLT016"},{"key":"2","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2014.09.002"},{"key":"3","doi-asserted-by":"publisher","DOI":"10.1007\/s10586-017-1077-z"},{"key":"4","doi-asserted-by":"publisher","DOI":"10.1561\/1500000011"},{"first-page":"79","article-title":"Thumbs up?: sentiment classification using machine learning techniques","author":"B. Pang","key":"5"},{"first-page":"1724","article-title":"Learning phrase representations using RNN encoder\u2013decoder for statistical machine translation","author":"K. Cho","key":"6"},{"first-page":"6077","article-title":"Hierarchical attention flow for multiple-choice reading comprehension","author":"H. Zhu","key":"7"},{"key":"8","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2016.10.065"},{"issue":"3","key":"9","first-page":"459","article-title":"Chinese text sentiment classification by H-RNN-CNN","volume":"54","author":"F. Luo","year":"2018","journal-title":"Beijing Da Xue Xue Bao"},{"key":"10","doi-asserted-by":"publisher","DOI":"10.1155\/2019\/3831809"},{"first-page":"1188","article-title":"Distributed representations of sentences and documents","author":"Q. Le","key":"11"},{"key":"12","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"first-page":"4171","article-title":"BERT: pre-training of deep bidirectional transformers for language understanding","author":"J. Devlin","key":"13"},{"author":"Y. Cui","key":"14","article-title":"Pre-training with whole word masking for Chinese BERT"},{"author":"X. Li","key":"15","article-title":"Exploiting BERT for end-to-end aspect-based sentiment analysis"},{"first-page":"1679","article-title":"Linguistically regularized LSTM for sentiment classification","author":"Q. Qian","key":"16"},{"key":"17","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2018.8489612"},{"first-page":"8105","article-title":"Sentiment lexicon enhanced attention-based LSTM for sentiment classification","author":"Z. Lei","key":"18"},{"key":"19","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"author":"D. Bahdanau","key":"20","article-title":"Neural machine translation by jointly learning to align and translate"},{"first-page":"1412","article-title":"Effective approaches to attention-based neural machine translation","author":"M.-T. Luong","key":"21"},{"first-page":"5998","article-title":"Attention is all you need","author":"A. Vaswani","key":"22"},{"key":"23","doi-asserted-by":"publisher","DOI":"10.3115\/1073083.1073153"},{"key":"24","doi-asserted-by":"publisher","DOI":"10.1162\/coli_a_00049"},{"first-page":"94","article-title":"Incorporating lexicon knowledge into svm learning to improve sentiment classification","author":"J. Fang","key":"25"},{"key":"26","doi-asserted-by":"publisher","DOI":"10.1109\/access.2019.2920075"},{"key":"27","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"first-page":"1746","article-title":"Convolutional neural networks for sentence classification","author":"Y. Kim","key":"28"},{"first-page":"2428","article-title":"Combination of convolutional and recurrent neural network for sentiment analysis of short texts","author":"X. Wang","key":"29"},{"first-page":"1422","article-title":"Document modeling with gated recurrent neural network for sentiment classification","author":"D. Tang","key":"30"},{"key":"31","doi-asserted-by":"publisher","DOI":"10.1155\/2019\/8320316"},{"first-page":"1480","article-title":"Hierarchical attention networks for document classification","author":"Z. Yang","key":"32"},{"first-page":"1555","article-title":"Learning sentiment-specific word embedding for twitter sentiment classification","author":"D. Tang","key":"33"},{"key":"34","doi-asserted-by":"publisher","DOI":"10.1155\/2015\/715730"},{"first-page":"376","article-title":"BERT based hierarchical sequence classification for context-aware microblog sentiment analysis","author":"J. Lei","key":"35"},{"key":"36","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2005.06.042"},{"first-page":"115","article-title":"Seeing stars: exploiting class relationships for sentiment categorization with respect to rating scales","author":"B. Pang","key":"37"},{"author":"D. P. Kingma","key":"38","article-title":"Adam: a method for stochastic optimization"},{"first-page":"427","article-title":"Bag of tricks for efficient text classification","author":"A. Joulin","key":"39"},{"first-page":"138","article-title":"Analogical reasoning on Chinese morphological and semantic relations","author":"S. Li","key":"40"}],"container-title":["Computational Intelligence and Neuroscience"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/downloads.hindawi.com\/journals\/cin\/2020\/8858717.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/cin\/2020\/8858717.xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/cin\/2020\/8858717.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2020,11,4]],"date-time":"2020-11-04T17:35:19Z","timestamp":1604511319000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.hindawi.com\/journals\/cin\/2020\/8858717\/"}},"subtitle":[],"editor":[{"given":"Jos\u00e9 Alfredo","family":"Hern\u00e1ndez-P\u00e9rez","sequence":"additional","affiliation":[]}],"short-title":[],"issued":{"date-parts":[[2020,11,3]]},"references-count":40,"alternative-id":["8858717","8858717"],"URL":"https:\/\/doi.org\/10.1155\/2020\/8858717","relation":{},"ISSN":["1687-5273","1687-5265"],"issn-type":[{"type":"electronic","value":"1687-5273"},{"type":"print","value":"1687-5265"}],"subject":[],"published":{"date-parts":[[2020,11,3]]}}}