{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T05:44:13Z","timestamp":1769924653600,"version":"3.49.0"},"reference-count":30,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,12,8]],"date-time":"2023-12-08T00:00:00Z","timestamp":1701993600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,12,8]],"date-time":"2023-12-08T00:00:00Z","timestamp":1701993600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62277040"],"award-info":[{"award-number":["62277040"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2024,4]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Numerous graph neural network (GNN) models have been used for sentiment analysis in recent years. Nevertheless, addressing the issue of over-smoothing in GNNs for node representation and finding more effective ways to learn both global and local information within the graph structure, while improving model efficiency for scalability to large text sentiment corpora, remains a challenge. To tackle these issues, we propose a novel Bert-based unlinked graph embedding (BUGE) model for sentiment analysis. Initially, the model constructs a comprehensive text sentiment heterogeneous graph that more effectively captures global co-occurrence information between words. Next, by using specific sampling strategies, it efficiently preserves both global and local information within the graph structure, enabling nodes to receive more feature information. During the representation learning process, BUGE relies solely on attention mechanisms, without using graph convolutions or aggregation operators, thus avoiding the over-smoothing problem associated with node aggregation. This enhances model training efficiency and reduces memory storage requirements. Extensive experimental results and evaluations demonstrate that the adopted Bert-based unlinked graph embedding method is highly effective for sentiment analysis, especially when applied to large text sentiment corpora.<\/jats:p>","DOI":"10.1007\/s40747-023-01289-9","type":"journal-article","created":{"date-parts":[[2023,12,8]],"date-time":"2023-12-08T03:02:28Z","timestamp":1702004548000},"page":"2627-2638","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Bert-based graph unlinked embedding for sentiment analysis"],"prefix":"10.1007","volume":"10","author":[{"given":"Youkai","family":"Jin","sequence":"first","affiliation":[]},{"given":"Anping","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,8]]},"reference":[{"key":"1289_CR1","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1007\/978-981-16-3398-0_12","volume-title":"Principles of social networking: the new horizon and emerging challenges","author":"A Saxena","year":"2022","unstructured":"Saxena A, Reddy H, Saxena P (2022) Introduction to sentiment analysis covering basics, tools, evaluation metrics, challenges, and applications. In: Biswas A et al (eds) Principles of social networking: the new horizon and emerging challenges. Springer, Singapore, pp 249\u2013277"},{"key":"1289_CR2","doi-asserted-by":"publisher","first-page":"1846","DOI":"10.4018\/978-1-6684-6303-1.ch098","volume-title":"Research anthology on implementing sentiment analysis across multiple disciplines","author":"R Kaur","year":"2022","unstructured":"Kaur R, Kautish S (2022) Multimodal sentiment analysis: a survey and comparison. Research anthology on implementing sentiment analysis across multiple disciplines. IGI Global, USA, pp 1846\u20131870"},{"key":"1289_CR3","doi-asserted-by":"crossref","unstructured":"Yao L, Mao C, Luo Y. (2019) Graph convolutional networks for text classification. In: Proceedings of the AAAI conference on artificial intelligence, Honolulu, HI, USA, vol 33, pp 7370\u20137377","DOI":"10.1609\/aaai.v33i01.33017370"},{"key":"1289_CR4","doi-asserted-by":"crossref","unstructured":"Niu L, Zheng Q, Zhang L (2021) Enhance gated graph neural network with syntactic for sentiment analysis. In: 2021 IEEE international conference on advances in electrical engineering and computer applications (AEECA), Dalian, China, pp 1055\u20131060","DOI":"10.1109\/AEECA52519.2021.9574275"},{"key":"1289_CR5","unstructured":"Niepert M, Ahmed M, Kutzkov K (2016) Learning convolutional neural networks for graphs. USA, PMLR, In: International conference on machine learning. New York, pp 2014\u20132023"},{"key":"1289_CR6","doi-asserted-by":"crossref","unstructured":"Li Q, Han Z, Wu XM (2018) Deeper insights into graph convolutional networks for semi-supervised learning. In: Proceedings of the AAAI conference on artificial intelligence, vol 32(1), New Orleans, Louisiana, USA","DOI":"10.1609\/aaai.v32i1.11604"},{"key":"1289_CR7","first-page":"5998","volume":"30","author":"A Vaswani","year":"2017","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN et al (2017) Attention is all you need. Adv Neural Inf Process Syst 30:5998\u20136008","journal-title":"Adv Neural Inf Process Syst"},{"key":"1289_CR8","unstructured":"Kipf TN, Welling M (2016) Semi-supervised classification with graph convolutional networks. arXiv arXiv:1609.02907"},{"key":"1289_CR9","doi-asserted-by":"crossref","unstructured":"Huang L, Ma D, Li S, Zhang X, Wang H (2019) Text level graph neural network for text classification. arXiv arXiv:1910.02356","DOI":"10.18653\/v1\/D19-1345"},{"key":"1289_CR10","doi-asserted-by":"crossref","unstructured":"Ding K, Wang J, Li J, Li D, Liu H (2020) Be more with less: hypergraph attention networks for inductive text classification. arXiv arXiv:2011.00387","DOI":"10.18653\/v1\/2020.emnlp-main.399"},{"key":"1289_CR11","unstructured":"Zhu H, Koniusz P (2021) Simple spectral graph convolution. In: International conference on learning representations, Vienna, Austria"},{"key":"1289_CR12","doi-asserted-by":"crossref","unstructured":"Zhang Y, Yu X, Cui Z, Wu S, Wen Z, Wang L (2020) Every document owns its structure: inductive text classification via graph neural networks. arXiv:2004.13826","DOI":"10.18653\/v1\/2020.acl-main.31"},{"key":"1289_CR13","doi-asserted-by":"publisher","first-page":"15712","DOI":"10.1016\/j.eswa.2021.115712","volume":"186","author":"X Zhu","year":"2021","unstructured":"Zhu X, Zhu L, Guo J, Liang S, Dietze S (2021) GL-GCN: global and local dependency guided graph convolutional networks for aspect-based sentiment classification. Expert Syst Appl 186:15712","journal-title":"Expert Syst Appl"},{"issue":"4","key":"1289_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.102946","volume":"59","author":"Y Yang","year":"2022","unstructured":"Yang Y, Miao R, Wang Y, Wang X (2022) Contrastive graph convolutional networks with adaptive augmentation for text classification. Inf Process Manage 59(4):102946","journal-title":"Inf Process Manage"},{"key":"1289_CR15","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv arXiv:1810.04805"},{"key":"1289_CR16","doi-asserted-by":"crossref","unstructured":"Lu Z, Du P, Nie JY (2020) VGCN-BERT: augmenting BERT with graph embedding for text classification. Advances in information retrieval: 42nd European conference on IR research, ECIR 2020, Proceedings, Part I 42, Lisbon, Portugal, April 14\u201317","DOI":"10.1007\/978-3-030-45439-5_25"},{"issue":"11","key":"1289_CR17","doi-asserted-by":"publisher","first-page":"1536","DOI":"10.3390\/e23111536","volume":"23","author":"Y Yang","year":"2021","unstructured":"Yang Y, Cui X (2021) Bert-enhanced text graph neural network for classification. Entropy 23(11):1536","journal-title":"Entropy"},{"key":"1289_CR18","doi-asserted-by":"crossref","unstructured":"Lin Y, Meng Y, Sun X, Han Q, Kuang K, Li J, Wu F (2021) Bertgcn: transductive text classification by combining gcn and bert. arXiv arXiv:2105.05727","DOI":"10.18653\/v1\/2021.findings-acl.126"},{"key":"1289_CR19","doi-asserted-by":"crossref","unstructured":"Shen H, Ju X, Chen X, Yang G (2023) EDP-BGCNN: effective defect prediction via BERT-based graph convolutional neural network. In: 2023 IEEE 47th annual computers, software, and applications conference (COMPSAC), Turin, Italy pp 850\u2013859","DOI":"10.1109\/COMPSAC57700.2023.00114"},{"key":"1289_CR20","unstructured":"Zhang J, Zhang H, Xia C, Sun L (2020) Graph-bert: only attention is needed for learning graph representations. arXiv arXiv:2001.05140"},{"key":"1289_CR21","doi-asserted-by":"crossref","unstructured":"Zhang L, Meng Q (2019) Probabilistic ship domain with applications to ship collision risk assessment. Ocean Eng 186:106130","DOI":"10.1016\/j.oceaneng.2019.106130"},{"key":"1289_CR22","doi-asserted-by":"crossref","unstructured":"Pang B, Lee L (2005) Seeing stars: exploiting class relationships for sentiment categorization with respect to rating scales. arXiv arXiv:cs\/0506075","DOI":"10.3115\/1219840.1219855"},{"key":"1289_CR23","doi-asserted-by":"crossref","unstructured":"Tang J, Qu M, Wang M, Zhang M, Yan J, Mei Q (2015) Line: large-scale information network embedding. In: Proceedings of the 24th international conference on world wide web, Florence, Italy, pp 1067\u20131077","DOI":"10.1145\/2736277.2741093"},{"key":"1289_CR24","unstructured":"Maas A, Daly RE, Pham PT, Huang D, Ng AY, Potts C (2011) Learning word vectors for sentiment analysis. In: Proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies, Portland, OR, USA, pp 142\u2013150"},{"key":"1289_CR25","unstructured":"Socher R, Perelygin A, Wu J, Chuang J, Manning CD, Ng AY, Potts C (2013) Recursive deep models for semantic compositionality over a sentiment treebank. In: Proceedings of the 2013 conference on empirical methods in natural language processing, Seattle, WA, USA, pp 1631\u20131642"},{"key":"1289_CR26","doi-asserted-by":"crossref","unstructured":"Pennington J, Socher R, Manning CD (2014) Glove: global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), Boston, MA, USA","DOI":"10.3115\/v1\/D14-1162"},{"key":"1289_CR27","unstructured":"Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv:1412.6980"},{"key":"1289_CR28","doi-asserted-by":"crossref","unstructured":"Piao Y, Lee S, Lee D, Kim S (2022) Sparse structure learning via graph neural networks for inductive document classification. In: Proceedings of the AAAI conference on artificial intelligence, vol 36(10), Washington, USA, 7\u201314 February, pp 11165\u201311173","DOI":"10.1609\/aaai.v36i10.21366"},{"key":"1289_CR29","doi-asserted-by":"crossref","unstructured":"Liu X, You X, Zhang X, Wu J, Lv P (2020) Tensor graph convolutional networks for text classification. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, New York, NY, USA, pp 8409\u20138416","DOI":"10.1609\/aaai.v34i05.6359"},{"key":"1289_CR30","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.107659","volume":"236","author":"Y Dai","year":"2022","unstructured":"Dai Y, Shou L, Gong M, Xia X, Kang Z, Xu Z, Jiang D (2022) Graph fusion network for text classification. Knowl-Based Syst 236:107659","journal-title":"Knowl-Based Syst"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-023-01289-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-023-01289-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-023-01289-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,30]],"date-time":"2024-03-30T15:31:38Z","timestamp":1711812698000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-023-01289-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,8]]},"references-count":30,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["1289"],"URL":"https:\/\/doi.org\/10.1007\/s40747-023-01289-9","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,8]]},"assertion":[{"value":"16 July 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 November 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 December 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there is no conflict of interests regarding the publication of this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}