{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T05:41:20Z","timestamp":1769924480559,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":36,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,27]],"date-time":"2023-10-27T00:00:00Z","timestamp":1698364800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"the Natural Science Foundation of Guangdong Province, China","award":["No. 2021A1515011598"],"award-info":[{"award-number":["No. 2021A1515011598"]}]},{"name":"the Fundamental Research Funds for the Central Universities","award":["No. 2022ZYGXZR023"],"award-info":[{"award-number":["No. 2022ZYGXZR023"]}]},{"name":"the National Natural Science Foundation of China","award":["Nos. 62003141 and U20A20224"],"award-info":[{"award-number":["Nos. 62003141 and U20A20224"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,27]]},"DOI":"10.1145\/3633637.3633642","type":"proceedings-article","created":{"date-parts":[[2024,2,28]],"date-time":"2024-02-28T08:08:05Z","timestamp":1709107685000},"page":"34-39","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["A Deep Learning Sentiment Analysis Method based on ERNIE and Modified DPCNN"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-1041-3984","authenticated-orcid":false,"given":"Yang","family":"Yang","sequence":"first","affiliation":[{"name":"School of Automation Science and Engineering, South China University of Technology, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8688-2662","authenticated-orcid":false,"given":"Xunde","family":"Dong","sequence":"additional","affiliation":[{"name":"School of Automation Science and Engineering, South China University of Technology, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-3908-4328","authenticated-orcid":false,"given":"Yupeng","family":"Qiang","sequence":"additional","affiliation":[{"name":"School of Automation Science and Engineering, South China University of Technology, China"}]}],"member":"320","published-online":{"date-parts":[[2024,2,28]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"A Statistical Approach To Machine Translation. Computational Linguistics 16, 2","author":"Brown F.","year":"2002","unstructured":"P.\u00a0F. Brown, J. Cocke, Sad Pietra, Vjd Pietra, and P.\u00a0S. Roossin. 2002. A Statistical Approach To Machine Translation. Computational Linguistics 16, 2 (2002)."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Peng Chen Zhongqian Sun Lidong Bing and Wei Yang. 2017. Recurrent attention network on memory for aspect sentiment analysis. 452\u2013461.","DOI":"10.18653\/v1\/D17-1047"},{"key":"e_1_3_2_1_3_1","volume-title":"Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555","author":"Clark Kevin","year":"2020","unstructured":"Kevin Clark, Minh-Thang Luong, Quoc\u00a0V. Le, and Christopher\u00a0D. Manning. 2020. Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555 (2020)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.12720\/jait.10.1.14-19"},{"key":"e_1_3_2_1_5_1","article-title":"Quantifying the natural sentiment strength of polar term senses using semantic gloss information and degree adverbs","volume":"11","author":"Darwich Mohammad","year":"2020","unstructured":"Mohammad Darwich, Shahrul Azman\u00a0Mohd Noah, Nazlia Omar, Nurul\u00a0Aida Osman, and Ibrahim Said. 2020. Quantifying the natural sentiment strength of polar term senses using semantic gloss information and degree adverbs. Journal of Advances in Information Technology Vol 11, 3 (2020).","journal-title":"Journal of Advances in Information Technology"},{"key":"e_1_3_2_1_6_1","volume-title":"GoEmotions: A dataset of fine-grained emotions. arXiv preprint arXiv:2005.00547","author":"Demszky Dorottya","year":"2020","unstructured":"Dorottya Demszky, Dana Movshovitz-Attias, Jeongwoo Ko, Alan Cowen, Gaurav Nemade, and Sujith Ravi. 2020. GoEmotions: A dataset of fine-grained emotions. arXiv preprint arXiv:2005.00547 (2020)."},{"key":"e_1_3_2_1_7_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_8_1","unstructured":"Cicero Dos\u00a0Santos and Maira Gatti. 2014. Deep convolutional neural networks for sentiment analysis of short texts. 69\u201378."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.12720\/jait.14.2.342-349"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2946594"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.4304\/jait.2.4.199-203"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/s41060-018-0096-z"},{"key":"e_1_3_2_1_13_1","volume-title":"Tenth Acm Sigkdd International Conference on Knowledge Discovery and Data Mining(Tenth Acm Sigkdd International Conference on Knowledge Discovery and Data Mining).","author":"Hu M.","unstructured":"M. Hu and L. Bing. 2004. Mining and summarizing customer reviews. In Tenth Acm Sigkdd International Conference on Knowledge Discovery and Data Mining(Tenth Acm Sigkdd International Conference on Knowledge Discovery and Data Mining)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"R. Johnson and Z. Tong. 2017. Deep Pyramid Convolutional Neural Networks for Text Categorization. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)(Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)).","DOI":"10.18653\/v1\/P17-1052"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"crossref","unstructured":"A. Joulin E. Grave P. Bojanowski and T. Mikolov. 2017. Bag of Tricks for Efficient Text Classification. (2017).","DOI":"10.18653\/v1\/E17-2068"},{"key":"e_1_3_2_1_16_1","volume-title":"Convolutional Neural Networks for Sentence Classification. Eprint Arxiv","author":"Kim Y.","year":"2014","unstructured":"Y. Kim. 2014. Convolutional Neural Networks for Sentence Classification. Eprint Arxiv (2014)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10489-020-01964-1"},{"key":"e_1_3_2_1_18_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_1_19_1","unstructured":"Ilya Loshchilov and Frank Hutter. 2017. Fixing weight decay regularization in adam. (2017)."},{"key":"e_1_3_2_1_20_1","volume-title":"Prompt-Based Generative Multi-label Emotion Prediction with Label Contrastive Learning. Vol.\u00a013551","author":"Lu Wei","unstructured":"Wei Lu, Shujian Huang, Yu Hong, and Xiabing Zhou. 2022. Prompt-Based Generative Multi-label Emotion Prediction with Label Contrastive Learning. Vol.\u00a013551. Springer International Publishing AG, Switzerland, 551\u2013563. Q334-342 Book, Section ER -."},{"key":"e_1_3_2_1_21_1","volume-title":"Efficient Estimation of Word Representations in Vector Space. Computer Science","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient Estimation of Word Representations in Vector Space. Computer Science (2013)."},{"key":"e_1_3_2_1_22_1","volume-title":"The emotion machine: Commonsense thinking, artificial intelligence, and the future of the human mind","author":"Minsky Marvin","unstructured":"Marvin Minsky. 2007. The emotion machine: Commonsense thinking, artificial intelligence, and the future of the human mind. Simon and Schuster."},{"key":"e_1_3_2_1_23_1","volume-title":"Mohammad and Felipe Bravo-Marquez","author":"M.","year":"2017","unstructured":"Saif\u00a0M. Mohammad and Felipe Bravo-Marquez. 2017. WASSA-2017 shared task on emotion intensity. arXiv preprint arXiv:1708.03700 (2017)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"Soujanya Poria Erik Cambria and Alexander Gelbukh. 2015. Deep convolutional neural network textual features and multiple kernel learning for utterance-level multimodal sentiment analysis. 2539\u20132544.","DOI":"10.18653\/v1\/D15-1303"},{"key":"e_1_3_2_1_25_1","volume-title":"Linguistically regularized lstms for sentiment classification. arXiv preprint arXiv:1611.03949","author":"Qian Qiao","year":"2016","unstructured":"Qiao Qian, Minlie Huang, Jinhao Lei, and Xiaoyan Zhu. 2016. Linguistically regularized lstms for sentiment classification. arXiv preprint arXiv:1611.03949 (2016)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0245909"},{"key":"e_1_3_2_1_27_1","volume-title":"Fine-Grained Emotion Prediction by Modeling Emotion Definitions","author":"Singh Gargi","unstructured":"Gargi Singh, Dhanajit Brahma, Piyush Rai, and Ashutosh Modi. 2021. Fine-Grained Emotion Prediction by Modeling Emotion Definitions. IEEE, 1\u20138."},{"key":"e_1_3_2_1_28_1","volume-title":"Utilizing BERT for aspect-based sentiment analysis via constructing auxiliary sentence. arXiv preprint arXiv:1903.09588","author":"Sun Chi","year":"2019","unstructured":"Chi Sun, Luyao Huang, and Xipeng Qiu. 2019. Utilizing BERT for aspect-based sentiment analysis via constructing auxiliary sentence. arXiv preprint arXiv:1903.09588 (2019)."},{"key":"e_1_3_2_1_29_1","volume-title":"Ernie: Enhanced representation through knowledge integration. arXiv preprint arXiv:1904.09223","author":"Sun Yu","year":"2019","unstructured":"Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, and Hua Wu. 2019. Ernie: Enhanced representation through knowledge integration. arXiv preprint arXiv:1904.09223 (2019)."},{"key":"e_1_3_2_1_30_1","unstructured":"A. Vaswani N. Shazeer N. Parmar J. Uszkoreit L. Jones A.\u00a0N. Gomez L. Kaiser and I. Polosukhin. 2017. Attention Is All You Need. arXiv (2017)."},{"key":"e_1_3_2_1_31_1","volume-title":"Performance Evaluation of Sentiment Analysis on Text and Emoji Data Using End-to-End Transfer Learning Distributed and Explainable AI Models [J]. Journal of Advances in Information Technology 13","author":"Velampalli Sirisha","year":"2022","unstructured":"Sirisha Velampalli, Chandrashekar Muniyappa, and Ashutosh Saxena. 2022. Performance Evaluation of Sentiment Analysis on Text and Emoji Data Using End-to-End Transfer Learning Distributed and Explainable AI Models [J]. Journal of Advances in Information Technology 13 (2022)."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"crossref","unstructured":"Jin Wang Liang-Chih Yu K.\u00a0Robert Lai and Xuejie Zhang. 2016. Dimensional sentiment analysis using a regional CNN-LSTM model. 225\u2013230.","DOI":"10.18653\/v1\/P16-2037"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Xin Wang Yuanchao Liu Cheng-Jie Sun Baoxun Wang and Xiaolong Wang. 2015. Predicting polarities of tweets by composing word embeddings with long short-term memory. 1343\u20131353.","DOI":"10.3115\/v1\/P15-1130"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"crossref","unstructured":"Yequan Wang Minlie Huang Xiaoyan Zhu and Li Zhao. 2016. Attention-based LSTM for aspect-level sentiment classification. 606\u2013615.","DOI":"10.18653\/v1\/D16-1058"},{"key":"e_1_3_2_1_35_1","volume-title":"Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32","author":"Yang Zhilin","year":"2019","unstructured":"Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ\u00a0R. Salakhutdinov, and Quoc\u00a0V. Le. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32 (2019)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.12720\/jait.9.2.33-38"}],"event":{"name":"ICCPR 2023: 2023 12th International Conference on Computing and Pattern Recognition","location":"Qingdao China","acronym":"ICCPR 2023"},"container-title":["2023 12th International Conference on Computing and Pattern Recognition"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3633637.3633642","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3633637.3633642","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T17:18:52Z","timestamp":1755883132000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3633637.3633642"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,27]]},"references-count":36,"alternative-id":["10.1145\/3633637.3633642","10.1145\/3633637"],"URL":"https:\/\/doi.org\/10.1145\/3633637.3633642","relation":{},"subject":[],"published":{"date-parts":[[2023,10,27]]},"assertion":[{"value":"2024-02-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}