{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,17]],"date-time":"2025-11-17T02:30:16Z","timestamp":1763346616262,"version":"3.45.0"},"reference-count":45,"publisher":"Tech Science Press","issue":"3","license":[{"start":{"date-parts":[[2025,11,2]],"date-time":"2025-11-02T00:00:00Z","timestamp":1762041600000},"content-version":"vor","delay-in-days":305,"URL":"https:\/\/doi.org\/10.32604\/TSP-CROSSMARKPOLICY"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["CMC"],"published-print":{"date-parts":[[2025]]},"DOI":"10.32604\/cmc.2025.071656","type":"journal-article","created":{"date-parts":[[2025,10,13]],"date-time":"2025-10-13T08:42:35Z","timestamp":1760344955000},"page":"5823-5845","update-policy":"https:\/\/doi.org\/10.32604\/tsp-crossmarkpolicy","source":"Crossref","is-referenced-by-count":0,"title":["GLAMSNet: A Gated-Linear Aspect-Aware Multimodal Sentiment Network with Alignment Supervision and External Knowledge Guidance"],"prefix":"10.32604","volume":"85","author":[{"given":"Dan","family":"Wang","sequence":"first","affiliation":[]},{"given":"Zhoubin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yuze","family":"Xia","sequence":"additional","affiliation":[]},{"given":"Zhenhua","family":"Yu","sequence":"additional","affiliation":[]}],"member":"17807","published-online":{"date-parts":[[2025]]},"reference":[{"key":"ref1","doi-asserted-by":"crossref","first-page":"845","DOI":"10.1109\/TAFFC.2020.2970399","article-title":"Issues and challenges of aspect-based sentiment analysis: a comprehensive survey","volume":"13","author":"Nazir","year":"2022","journal-title":"IEEE Trans Affect Comput"},{"key":"ref2","series-title":"Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing; 2016 Nov 1\u20135","first-page":"606","article-title":"Attention-based LSTM for aspect-level sentiment classification","author":"Wang"},{"key":"ref3","series-title":"Proceedings of the 26th International Joint Conference on Artificial Intelligence; 2017 Aug 19\u201326","first-page":"4068","article-title":"Interactive attention networks for aspect-level sentiment classification","author":"Ma"},{"key":"ref4","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805. 2018. doi:10.48550\/arXiv.1810.04805."},{"key":"ref5","doi-asserted-by":"crossref","unstructured":"Lewis M, Liu Y, Goyal N, Ghazvininejad M, Mohamed A, Levy O, et al. BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv:1910.13461. 2019. doi:10.48550\/arXiv.1910.13461.","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"ref6","first-page":"1863","article-title":"Low-rank adapter layers and bidirectional gated feature fusion for multimodal hateful memes classification","volume":"84","author":"Huang","year":"2025","journal-title":"Comput Mater Contin"},{"key":"ref7","first-page":"2659","article-title":"Fake news detection based on cross-modal ambiguity computation and multi-scale feature fusion","volume":"83","author":"Cao","year":"2025","journal-title":"Comput Mater Contin"},{"key":"ref8","doi-asserted-by":"crossref","first-page":"102552","DOI":"10.1016\/j.inffus.2024.102552","article-title":"Multimodal aspect-based sentiment analysis: a survey of tasks, methods, challenges and future directions","volume":"112","author":"Zhao","year":"2024","journal-title":"Inf Fusion"},{"key":"ref9","first-page":"4105","article-title":"Exploring the effectiveness of machine learning and deep learning algorithms for sentiment analysis: a systematic literature review","volume":"84","author":"Shin","year":"2025","journal-title":"Comput Mater Contin"},{"key":"ref10","first-page":"4977","article-title":"X-OODM: leveraging explainable object-oriented design methodology for multi-domain sentiment analysis","volume":"82","author":"Javed","year":"2025","journal-title":"Comput Mater Contin"},{"key":"ref11","first-page":"371","article-title":"Multi-interactive memory network for aspect based multimodal sentiment analysis","volume":"33","author":"Xu","year":"2019","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"ref12","series-title":"Proceedings of the 28th International Joint Conference on Artificial Intelligence; 2019 Aug 10\u201316","first-page":"5408","article-title":"Adapting BERT for target-oriented multimodal sentiment classification","author":"Yu"},{"key":"ref13","doi-asserted-by":"crossref","first-page":"429","DOI":"10.1109\/TASLP.2019.2957872","article-title":"Entity-sensitive attention and fusion network for entity-level multimodal sentiment classification","volume":"28","author":"Yu","year":"2020","journal-title":"IEEE\/ACM Trans Audio Speech Lang Process"},{"key":"ref14","doi-asserted-by":"crossref","first-page":"1966","DOI":"10.1109\/TAFFC.2022.3171091","article-title":"Hierarchical interactive multimodal transformer for aspect-based multimodal sentiment analysis","volume":"14","author":"Yu","year":"2023","journal-title":"IEEE Trans Affect Comput"},{"key":"ref15","series-title":"Proceedings of the 29th International Conference on Computational Linguistics; 2022 Oct 12\u201317","first-page":"6784","article-title":"Learning from adjective-noun pairs: a knowledge-enhanced framework for target-oriented multimodal sentiment classification","author":"Zhao"},{"key":"ref16","doi-asserted-by":"crossref","first-page":"903","DOI":"10.1007\/s41060-023-00497-3","article-title":"Aspect-level multimodal sentiment analysis based on co-attention fusion","volume":"20","author":"Wang","year":"2025","journal-title":"Int J Data Sci Anal"},{"key":"ref17","series-title":"Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing; 2017 Sep 7\u201311","first-page":"452","article-title":"Recurrent attention network on memory for aspect sentiment analysis","author":"Chen"},{"key":"ref18","series-title":"Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2018 Jul 15\u201320","first-page":"946","article-title":"Transformation networks for target-oriented sentiment classification","author":"Li"},{"key":"ref19","series-title":"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing; 2018 Oct 31\u2013Nov 4","first-page":"3433","article-title":"Multi-grained attention network for aspect-level sentiment classification","author":"Fan"},{"key":"ref20","series-title":"IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2016 Jun 27\u201330","first-page":"770","article-title":"Deep residual learning for image recognition","author":"He"},{"key":"ref21","doi-asserted-by":"crossref","first-page":"78","DOI":"10.1007\/s10462-023-10685-z","article-title":"Multi-level textual-visual alignment and fusion network for multimodal aspect-based sentiment analysis","volume":"57","author":"Li","year":"2024","journal-title":"Artif Intell Rev"},{"key":"ref22","first-page":"13","article-title":"ViLBERT: pretraining task-agnostic visiolinguistic representations for vision-and-language tasks","volume":"32","author":"Lu","year":"2019","journal-title":"Adv Neural Inf Process Syst"},{"key":"ref23","series-title":"Proceedings of the 29th ACM International Conference on Multimedia; 2021 Oct 20\u201324","first-page":"3034","article-title":"Exploiting BERT for multimodal target sentiment classification through input space translation","author":"Khan"},{"key":"ref24","series-title":"Chinese Conference on Pattern Recognition and Computer Vision (PRCV); 2021 Oct 29\u2013Nov 1","first-page":"3","article-title":"SaliencyBERT: recurrent attention network for target-oriented multimodal sentiment classification","author":"Wang"},{"key":"ref25","series-title":"Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing; 2022 Dec 7\u201311","first-page":"3324","article-title":"Face-sensitive image-to-emotional-text cross-modal translation for multimodal aspect-based sentiment analysis","author":"Yang"},{"key":"ref26","doi-asserted-by":"crossref","first-page":"688","DOI":"10.1109\/TETC.2022.3231746","article-title":"Affective region recognition and fusion network for target-level multimodal sentiment classification","volume":"12","author":"Jia","year":"2024","journal-title":"IEEE Trans Emerg Top Comput"},{"key":"ref27","doi-asserted-by":"crossref","first-page":"9298","DOI":"10.1109\/TNNLS.2024.3415028","article-title":"Dual causes generation assisted model for multimodal aspect-based sentiment classification","volume":"36","author":"Fan","year":"2025","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"ref28","first-page":"18869","article-title":"A novel energy based model mechanism for multi-modal aspect-based sentiment analysis","volume":"38","author":"Peng","year":"2024","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"ref29","doi-asserted-by":"crossref","first-page":"309","DOI":"10.1109\/TAFFC.2023.3274829","article-title":"Efficient multimodal transformer with dual-level feature restoration for robust multimodal sentiment analysis","volume":"15","author":"Sun","year":"2024","journal-title":"IEEE Trans Affect Comput"},{"key":"ref30","doi-asserted-by":"crossref","first-page":"1856","DOI":"10.1109\/TAFFC.2024.3378570","article-title":"Contrastive learning based modality-invariant feature acquisition for robust multimodal emotion recognition with missing modalities","volume":"15","author":"Liu","year":"2024","journal-title":"IEEE Trans Affect Comput"},{"key":"ref31","doi-asserted-by":"crossref","first-page":"125970","DOI":"10.1016\/j.eswa.2024.125970","article-title":"Exploit the visual sentiment of the item images to fuse with textual sentiment in context aware collaborative filtering","volume":"265","author":"Wu","year":"2025","journal-title":"Expert Syst Appl"},{"key":"ref32","unstructured":"Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, et al. RoBERTa: a robustly optimized BERT pretraining approach. arXiv:1907.11692. 2019. doi:10.48550\/arXiv.1907.11692."},{"key":"ref33","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","article-title":"Faster R-CNN: towards real-time object detection with region proposal networks","volume":"39","author":"Ren","year":"2017","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"ref34","unstructured":"He P, Liu X, Gao J, Chen W. DeBERTa: decoding-enhanced BERT with disentangled attention. arXiv:2006.03654. 2020. doi:10.48550\/arxiv.2006.03654."},{"key":"ref35","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, et al. An image is worth 16 \u00d7 16 words: transformers for image recognition at scale. arXiv:2010.11929. 2020. doi:10.48550\/arXiv.2010.11929."},{"key":"ref36","unstructured":"Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, et al. Learning transferable visual models from natural language supervision. arXiv:2103.00020. 2021. doi:10.48550\/arxiv.2103.00020."},{"key":"ref37","series-title":"International Conference on Machine learning; 2022 Jul 17\u201323","first-page":"12888","article-title":"Blip: bootstrapping language-image pre-training for unified vision-language understanding and generation","author":"Li"},{"key":"ref38","series-title":"Proceedings of the 31st International Joint Conference on Artificial Intelligence; 2022 Jul 23\u201329","first-page":"4482","article-title":"Targeted multimodal sentiment classification based on coarse-to-fine grained image-target matching","author":"Yu"},{"key":"ref39","series-title":"Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics; 2019 Jul 28\u2013Aug 2","first-page":"6559","article-title":"Multimodal transformer for unaligned multimodal language sequences","author":"Tsai"},{"key":"ref40","series-title":"Proceedings of the 27th International Joint Conference on Artificial Intelligence; 2018 Jul 13\u201319","first-page":"1114","article-title":"Rethinking diversified and discriminative proposal generation for visual grounding","author":"Yu"},{"key":"ref41","series-title":"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; 2020 Jul 5\u201310","first-page":"8211","article-title":"TVQA+: spatio-temporal grounding for video question answering","author":"Lei"},{"key":"ref42","series-title":"2023 IEEE\/CVF International Conference on Computer Vision (ICCV); 2023 Oct 1\u20136","first-page":"17256","article-title":"EfficientViT: lightweight multi-scale attention for high-resolution dense prediction","author":"Cai"},{"key":"ref43","series-title":"Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2018 Jul 15\u201320","first-page":"1990","article-title":"Visual attention model for name tagging in multimodal social media","author":"Lu"},{"key":"ref44","first-page":"5674","article-title":"Adaptive co-attention network for named entity recognition in tweets","volume":"32","author":"Zhang","year":"2018","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"ref45","first-page":"2956","article-title":"Attention guided CAM: visual explanations of vision transformer guided by self-attention","volume":"38","author":"Leem","year":"2024","journal-title":"Proc AAAI Conf Artif Intell"}],"container-title":["Computers, Materials &amp; Continua"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/cdn.techscience.cn\/files\/cmc\/2025\/TSP_CMC-85-3\/TSP_CMC_71656\/TSP_CMC_71656.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,17]],"date-time":"2025-11-17T02:26:12Z","timestamp":1763346372000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.techscience.com\/cmc\/v85n3\/64207"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":45,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2025]]},"published-print":{"date-parts":[[2025]]}},"URL":"https:\/\/doi.org\/10.32604\/cmc.2025.071656","relation":{},"ISSN":["1546-2226"],"issn-type":[{"type":"electronic","value":"1546-2226"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"2025-08-09","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-09-19","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-10-23","order":2,"name":"published","label":"Published Online","group":{"name":"publication_history","label":"Publication History"}}]}}