{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T09:56:59Z","timestamp":1773482219360,"version":"3.50.1"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T00:00:00Z","timestamp":1764633600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T00:00:00Z","timestamp":1764633600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the 14th Five-Year Plan Project of Shandong Provincial Education Science","award":["2023ZC134"],"award-info":[{"award-number":["2023ZC134"]}]},{"name":"the Heze City Science and Technology Development Plan","award":["KJTPY202310"],"award-info":[{"award-number":["KJTPY202310"]}]},{"name":"the Social Science Planning Project of Heze","award":["ZZ-2024-15"],"award-info":[{"award-number":["ZZ-2024-15"]}]},{"name":"the Artificial Intelligence Education Research Funding Project of Shandong Province","award":["SDDJ202501006"],"award-info":[{"award-number":["SDDJ202501006"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Data Sci Anal"],"published-print":{"date-parts":[[2026,6]]},"DOI":"10.1007\/s41060-025-00943-4","type":"journal-article","created":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T07:29:39Z","timestamp":1764660579000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Docsentinet: a adaptive architecture for efficient document-level sentiment analysis"],"prefix":"10.1007","volume":"21","author":[{"given":"Xiaoyang","family":"Wang","sequence":"first","affiliation":[]},{"given":"Wenfeng","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yuzhen","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Yaling","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Qiaoqiao","family":"Du","sequence":"additional","affiliation":[]},{"given":"Longqing","family":"Bao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,2]]},"reference":[{"key":"943_CR1","doi-asserted-by":"crossref","unstructured":"Zhang, L., Liu, B.: Sentiment analysis and opinion mining. In: Encyclopedia of Machine Learning and Data Mining (2017). https:\/\/api.semanticscholar.org\/CorpusID:268125336","DOI":"10.1007\/978-1-4899-7687-1_907"},{"key":"943_CR2","doi-asserted-by":"crossref","unstructured":"Cambria, E., Das, D., Bandyopadhyay, S., Feraco, A.: Affective computing and sentiment analysis. A practical guide to sentiment analysis, pp. 1\u201310 (2017)","DOI":"10.1007\/978-3-319-55394-8_1"},{"key":"943_CR3","doi-asserted-by":"publisher","first-page":"344","DOI":"10.1016\/j.inffus.2019.05.003","volume":"52","author":"JF S\u00e1nchez-Rada","year":"2019","unstructured":"S\u00e1nchez-Rada, J.F., Iglesias, C.A.: Social context in sentiment analysis: formal definition, overview of current trends and framework for comparison. Inf. Fusion 52, 344\u2013356 (2019)","journal-title":"Inf. Fusion"},{"key":"943_CR4","doi-asserted-by":"publisher","first-page":"41283","DOI":"10.1109\/ACCESS.2021.3064830","volume":"9","author":"S Tam","year":"2021","unstructured":"Tam, S., Said, R.B., Tanri\u00f6ver, \u00d6.\u00d6.: A ConvBiLSTM deep learning model-based approach for Twitter sentiment classification. IEEE Access 9, 41283\u201341293 (2021)","journal-title":"IEEE Access"},{"key":"943_CR5","unstructured":"Rodrigues, A.P., Chiplunkar, N.N.: A new big data approach for topic classification and sentiment analysis of twitter data. Evol. Intell. 1\u201311 (2022)"},{"issue":"3","key":"943_CR6","first-page":"1","volume":"8","author":"Q Wang","year":"2021","unstructured":"Wang, Q., Li, W., Jin, Z.: Review of text classification in deep learning. Open Access Library J. 8(3), 1\u20138 (2021)","journal-title":"Open Access Library J."},{"issue":"3","key":"943_CR7","first-page":"1","volume":"54","author":"S Minaee","year":"2021","unstructured":"Minaee, S., Kalchbrenner, N., Cambria, E., Nikzad, N., Chenaghlu, M., Gao, J.: Deep learning-based text classification: a comprehensive review. ACM Comput. Surv. CSUR 54(3), 1\u201340 (2021)","journal-title":"ACM Comput. Surv. CSUR"},{"key":"943_CR8","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"943_CR9","unstructured":"Tay, Y., Dehghani, M., Abnar, S., Shen, Y., Bahri, D., Pham, P., Rao, J., Yang, L., Ruder, S., Metzler, D.: Long range arena: a benchmark for efficient transformers. arXiv preprint arXiv:2011.04006 (2020)"},{"key":"943_CR10","unstructured":"Kitaev, N., Kaiser, \u0141., Levskaya, A.: Reformer: the efficient transformer. arXiv preprint arXiv:2001.04451 (2020)"},{"issue":"03","key":"943_CR11","first-page":"1","volume":"7","author":"W Li","year":"2020","unstructured":"Li, W., et al.: Review of research on text sentiment analysis based on deep learning. Open Access Library J. 7(03), 1 (2020)","journal-title":"Open Access Library J."},{"key":"943_CR12","unstructured":"Keles, F.D., Wijewardena, P.M., Hegde, C.: On the computational complexity of self-attention. In: International Conference on Algorithmic Learning Theory (2022). https:\/\/api.semanticscholar.org\/CorpusID:252198880"},{"key":"943_CR13","unstructured":"Shazeer, N.M.: Fast transformer decoding: one write-head is all you need. arXiv arXiv:1911.02150 (2019)"},{"key":"943_CR14","doi-asserted-by":"crossref","unstructured":"Peng, B., Alcaide, E., Anthony, Q.G., Albalak, A., Arcadinho, S., Biderman, S., Cao, H., Cheng, X., Chung, M., Grella, M., Kranthikiran, G., He, X., Hou, H., Kazienko, P., Koco\u0144, J., Kong, J., Koptyra, B., Lau, H., Mantri, K.S.I., Mom, F., Saito, A., Tang, X., Wang, B., Wind, J.S., Wozniak, S., Zhang, R., Zhang, Z., Zhao, Q., Zhou, P., Zhu, J., Zhu, R.: RWKV: reinventing RNNs for the transformer era. In: Conference on Empirical Methods in Natural Language Processing (2023). https:\/\/api.semanticscholar.org\/CorpusID:258832459","DOI":"10.18653\/v1\/2023.findings-emnlp.936"},{"key":"943_CR15","unstructured":"Katharopoulos, A., Vyas, A., Pappas, N., Fleuret, F.: Transformers are RNNs: fast autoregressive transformers with linear attention. In: International Conference on Machine Learning (2020). https:\/\/api.semanticscholar.org\/CorpusID:220250819"},{"key":"943_CR16","doi-asserted-by":"crossref","unstructured":"Nauen, T.C., Palacio, S., Dengel, A.: TaylorShift: shifting the complexity of self-attention from squared to linear (and Back) using Taylor-Softmax (2024)","DOI":"10.1007\/978-3-031-78172-8_1"},{"key":"943_CR17","unstructured":"Go, A., Bhayani, R., Huang, L.: Twitter sentiment classification using distant supervision. CS224N project report, Stanford, vol. 1, p. 2009 (2009)"},{"key":"943_CR18","doi-asserted-by":"crossref","unstructured":"Pang, B., Lee, L.: Seeing stars: exploiting class relationships for sentiment categorization with respect to rating scales. arXiv:cs\/0506075 (2005)","DOI":"10.3115\/1219840.1219855"},{"key":"943_CR19","doi-asserted-by":"crossref","unstructured":"Joachims, T.: Text categorization with support vector machines: learning with many relevant features. In: European Conference on Machine Learning, pp. 137\u2013142. Springer (1998)","DOI":"10.1007\/BFb0026683"},{"key":"943_CR20","unstructured":"Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013)"},{"key":"943_CR21","doi-asserted-by":"crossref","unstructured":"Tang, D., Qin, B., Liu, T.: Document modeling with gated recurrent neural network for sentiment classification. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1422\u20131432 (2015)","DOI":"10.18653\/v1\/D15-1167"},{"key":"943_CR22","doi-asserted-by":"crossref","unstructured":"Yang, Z., Yang, D., Dyer, C., He, X., Smola, A., Hovy, E.: Hierarchical attention networks for document classification. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1480\u20131489 (2016)","DOI":"10.18653\/v1\/N16-1174"},{"key":"943_CR23","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"943_CR24","unstructured":"Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"943_CR25","unstructured":"Sun, C., Huang, L., Qiu, X.: Utilizing Bert for aspect-based sentiment analysis via constructing auxiliary sentence. arXiv preprint arXiv:1903.09588 (2019)"},{"key":"943_CR26","unstructured":"Xu, H., Liu, B., Shu, L., Yu, P.S.: Bert post-training for review reading comprehension and aspect-based sentiment analysis. arXiv preprint arXiv:1904.02232 (2019)"},{"key":"943_CR27","doi-asserted-by":"crossref","unstructured":"Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., et al.: Huggingface\u2019s transformers: state-of-the-art natural language processing. arXiv preprint arXiv:1910.03771 (2019)","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"943_CR28","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., Stoyanov, V.: Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"943_CR29","unstructured":"Beltagy, I., Peters, M.E., Cohan, A.: Longformer: the long-document transformer. arXiv preprint arXiv:2004.05150 (2020)"},{"key":"943_CR30","first-page":"17283","volume":"33","author":"M Zaheer","year":"2020","unstructured":"Zaheer, M., Guruganesh, G., Dubey, K.A., Ainslie, J., Alberti, C., Ontanon, S., Pham, P., Ravula, A., Wang, Q., Yang, L., et al.: Big bird: transformers for longer sequences. Adv. Neural Inf. Process. Syst. 33, 17283\u201317297 (2020)","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"6","key":"943_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3530811","volume":"55","author":"Y Tay","year":"2022","unstructured":"Tay, Y., Dehghani, M., Bahri, D., Metzler, D.: Efficient transformers: a survey. ACM Comput. Surv. 55(6), 1\u201328 (2022)","journal-title":"ACM Comput. Surv."},{"key":"943_CR32","doi-asserted-by":"crossref","unstructured":"Abreu, J., Fred, L., Mac\u00eado, D., Zanchettin, C.: Hierarchical attentional hybrid neural networks for document classification. In: International Conference on Artificial Neural Networks, pp. 396\u2013402. Springer (2019)","DOI":"10.1007\/978-3-030-30493-5_39"},{"key":"943_CR33","unstructured":"Maas, A.L., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., Potts, C.: Learning word vectors for sentiment analysis. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 142\u2013150. Association for Computational Linguistics, Portland (2011). http:\/\/www.aclweb.org\/anthology\/P11-1015"},{"key":"943_CR34","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"943_CR35","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"943_CR36","doi-asserted-by":"crossref","unstructured":"Akiba, T., Sano, S., Yanase, T., Ohta, T., Koyama, M.: Optuna: a next-generation hyperparameter optimization framework. In: The 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 2623\u20132631 (2019)","DOI":"10.1145\/3292500.3330701"},{"issue":"8","key":"943_CR37","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)","journal-title":"OpenAI Blog"},{"key":"943_CR38","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., Liu, P.J.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(140), 1\u201367 (2020)"},{"key":"943_CR39","unstructured":"Clark, K., Luong, M.-T., Le, Q.V., Manning, C.D.: Electra: pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555 (2020)"},{"key":"943_CR40","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: Albert: a lite Bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942 (2019)"},{"key":"943_CR41","doi-asserted-by":"crossref","unstructured":"Xiao, S., Liu, Z., Zhang, P., Muennighoff, N.: C-Pack: packaged resources to advance general Chinese embedding (2023)","DOI":"10.1145\/3626772.3657878"},{"key":"943_CR42","unstructured":"He, P., Liu, X., Gao, J., Chen, W.: Deberta: decoding-enhanced Bert with disentangled attention. In: International Conference on Learning Representations (2021). https:\/\/openreview.net\/forum?id=XPZIaotutsD"},{"key":"943_CR43","volume-title":"Natural Language Processing with Python: Analyzing Text with the Natural Language Toolkit","author":"S Bird","year":"2009","unstructured":"Bird, S., Klein, E., Loper, E.: Natural Language Processing with Python: Analyzing Text with the Natural Language Toolkit. O\u2019Reilly Media Inc., Boston (2009)"}],"container-title":["International Journal of Data Science and Analytics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s41060-025-00943-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s41060-025-00943-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s41060-025-00943-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T09:35:58Z","timestamp":1773480958000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s41060-025-00943-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,2]]},"references-count":43,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,6]]}},"alternative-id":["943"],"URL":"https:\/\/doi.org\/10.1007\/s41060-025-00943-4","relation":{},"ISSN":["2364-415X","2364-4168"],"issn-type":[{"value":"2364-415X","type":"print"},{"value":"2364-4168","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,2]]},"assertion":[{"value":"24 March 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"34"}}