{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T18:40:59Z","timestamp":1772822459815,"version":"3.50.1"},"reference-count":37,"publisher":"Association for Computing Machinery (ACM)","issue":"6","license":[{"start":{"date-parts":[[2024,12,13]],"date-time":"2024-12-13T00:00:00Z","timestamp":1734048000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Korean government","award":["RS-2023-00244789"],"award-info":[{"award-number":["RS-2023-00244789"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Intell. Syst. Technol."],"published-print":{"date-parts":[[2024,12,31]]},"abstract":"<jats:p>Recent methodologies have achieved good performance in objectively summarizing important information from fact-based datasets such as Extreme Summarization and CNN Daily Mail. These methodologies involve abstractive summarization, extracting the core content from an input text and transforming it into natural sentences. Unlike fact-based documents, opinion-based documents require a thorough analysis of sentiment and understanding of the writer\u2019s intention. However, existing models do not explicitly consider these factors. Therefore, in this study, we propose a novel text summarization model that is specifically designed for opinion-based documents. Specifically, we identify the sentiment distribution of the entire document and train the summarization model to focus on major opinions that conform to the intended message while randomly masking minor opinions. Experimental results show that the proposed model outperforms existing summarization models in summarizing opinion-based documents, effectively capturing and highlighting the main opinions in the generated abstractive summaries.<\/jats:p>","DOI":"10.1145\/3696456","type":"journal-article","created":{"date-parts":[[2024,9,21]],"date-time":"2024-09-21T11:27:03Z","timestamp":1726918023000},"page":"1-20","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["User Opinion-Focused Abstractive Summarization Using Explainable Artificial Intelligence"],"prefix":"10.1145","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-3252-8560","authenticated-orcid":false,"given":"Hyunho","family":"Lee","sequence":"first","affiliation":[{"name":"Department of Data Science, Seoul National University of Science and Technology, Seoul, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4199-936X","authenticated-orcid":false,"given":"Younghoon","family":"Lee","sequence":"additional","affiliation":[{"name":"Department of Industrial Engineering, Seoul National University of Science and Technology, Seoul, South Korea"}]}],"member":"320","published-online":{"date-parts":[[2024,12,13]]},"reference":[{"key":"e_1_3_1_2_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.aci.2019.11.003"},{"key":"e_1_3_1_3_2","doi-asserted-by":"publisher","DOI":"10.1002\/widm.1424"},{"key":"e_1_3_1_4_2","doi-asserted-by":"publisher","DOI":"10.5555\/3495724.3495883"},{"key":"e_1_3_1_5_2","unstructured":"Xin Cheng Shen Gao Yuchi Zhang Yongliang Wang Xiuying Chen Mingzhe Li Dongyan Zhao and Rui Yan. 2023. Towards personalized review summarization by modeling historical reviews from customer and product separately. arXiv:2301.11682. Retrieved from https:\/\/arxiv.org\/abs\/2301.11682"},{"key":"e_1_3_1_6_2","unstructured":"Jacob Devlin Ming-Wei Chang Kenton Lee and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805. Retrieved from https:\/\/arxiv.org\/abs\/1810.04805"},{"key":"e_1_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-21743-2_7"},{"key":"e_1_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2023.3296373"},{"key":"e_1_3_1_9_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2018.10.003"},{"key":"e_1_3_1_10_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1630"},{"key":"e_1_3_1_11_2","doi-asserted-by":"crossref","unstructured":"Sebastian Gehrmann Yuntian Deng and Alexander M. Rush. 2018. Bottom-up abstractive summarization. arXiv:1808.10792. Retrieved from https:\/\/arxiv.org\/abs\/1808.10792","DOI":"10.18653\/v1\/D18-1443"},{"key":"e_1_3_1_12_2","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1168"},{"key":"e_1_3_1_13_2","unstructured":"Karthick Prasad Gunasekaran. 2023. Exploring sentiment analysis techniques in natural language processing: A comprehensive review. arXiv:2305.14842. Retrieved from https:\/\/arxiv.org\/abs\/2305.14842"},{"key":"e_1_3_1_14_2","volume-title":"Proceedings of the 28th International Conference on Neural Information Processing Systems","author":"Hermann Karl Moritz","year":"2015","unstructured":"Karl Moritz Hermann, Tomas Kocisky, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, and Phil Blunsom. 2015. Teaching machines to read and comprehend. In Proceedings of the 28th International Conference on Neural Information Processing Systems."},{"key":"e_1_3_1_15_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jretconser.2019.01.005"},{"key":"e_1_3_1_16_2","first-page":"16","volume-title":"Proceedings of the EACL Hackashop on News Media Content Analysis and Automated Report Generation","author":"Kokalj Enja","year":"2021","unstructured":"Enja Kokalj, Bla\u017e \u0160krlj, Nada Lavra\u010d, Senja Pollak, and Marko Robnik-\u0160ikonja. 2021. BERT meets shapley: Extending SHAP explanations to transformer-based classifiers. In Proceedings of the EACL Hackashop on News Media Content Analysis and Automated Report Generation, 16\u201321."},{"key":"e_1_3_1_17_2","doi-asserted-by":"crossref","unstructured":"Mike Lewis Yinhan Liu Naman Goyal Marjan Ghazvininejad Abdelrahman Mohamed Omer Levy Ves Stoyanov and Luke Zettlemoyer. 2019. Bart: Denoising sequence-to-sequence pre-training for natural language generation translation and comprehension. arXiv:1910.13461. Retrieved from https:\/\/arxiv.org\/abs\/1910.13461","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"e_1_3_1_18_2","first-page":"74","volume-title":"Text Summarization Branches Out","author":"Lin Chin-Yew","year":"2004","unstructured":"Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text Summarization Branches Out. Association for Computational Linguistics, 74\u201381."},{"key":"e_1_3_1_19_2","first-page":"4765","volume-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems","author":"Lundberg Scott M.","year":"2017","unstructured":"Scott M. Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, 4765\u20134774."},{"key":"e_1_3_1_20_2","unstructured":"Shuming Ma Xu Sun Junyang Lin and Xuancheng Ren. 2018. A hierarchical end-to-end model for jointly improving text summarization and sentiment classification. arXiv:1805.01089. Retrieved from https:\/\/arxiv.org\/abs\/1805.01089"},{"key":"e_1_3_1_21_2","volume-title":"Text Generation","author":"McKeown Kathleen","year":"1992","unstructured":"Kathleen McKeown. 1992. Text Generation. Cambridge University Press."},{"key":"e_1_3_1_22_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v31i1.10958"},{"key":"e_1_3_1_23_2","doi-asserted-by":"publisher","DOI":"10.3390\/pharmaceutics10030118"},{"key":"e_1_3_1_24_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-45275-8_22"},{"key":"e_1_3_1_25_2","volume-title":"The PageRank Citation Ranking: Bringing Order to the Web.","author":"Page Lawrence","year":"1999","unstructured":"Lawrence Page, Sergey Brin, Rajeev Motwani, and Terry Winograd. 1999. The PageRank Citation Ranking: Bringing Order to the Web. Technical Report. Stanford InfoLab."},{"key":"e_1_3_1_26_2","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401039"},{"key":"e_1_3_1_27_2","doi-asserted-by":"publisher","DOI":"10.5555\/3455716.3455856"},{"key":"e_1_3_1_28_2","doi-asserted-by":"crossref","unstructured":"Julian Salazar Davis Liang Toan Q. Nguyen and Katrin Kirchhoff. 2019. Masked language model scoring. arXiv:1910.14659. Retrieved from https:\/\/arxiv.org\/abs\/1910.14659","DOI":"10.18653\/v1\/2020.acl-main.240"},{"key":"e_1_3_1_29_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jretconser.2021.102771"},{"key":"e_1_3_1_30_2","unstructured":"Abigail See Peter J. Liu and Christopher D. Manning. 2017. Get to the point: Summarization with pointer-generator networks. arXiv:1704.04368. Retrieved from https:\/\/arxiv.org\/abs\/1704.04368"},{"key":"e_1_3_1_31_2","doi-asserted-by":"publisher","DOI":"10.5555\/3295222.3295349"},{"key":"e_1_3_1_32_2","doi-asserted-by":"publisher","DOI":"10.1007\/s12525-022-00561-z"},{"key":"e_1_3_1_33_2","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462854"},{"key":"e_1_3_1_34_2","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3615056"},{"key":"e_1_3_1_35_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.106858"},{"key":"e_1_3_1_36_2","first-page":"11328","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Zhang Jingqing","year":"2020","unstructured":"Jingqing Zhang, Yao Zhao, Mohammad Saleh, and Peter Liu. 2020. Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the International Conference on Machine Learning. PMLR, 11328\u201311339."},{"key":"e_1_3_1_37_2","first-page":"1462","volume-title":"Proceedings of the International Conference on Coling 2010: Posters","author":"Zhang Lei","year":"2010","unstructured":"Lei Zhang, Bing Liu, Suk Hwan Lim, and Eamonn O\u2019Brien-Strain. 2010. Extracting and ranking product features in opinion documents. In Proceedings of the International Conference on Coling 2010: Posters, 1462\u20131470."},{"key":"e_1_3_1_38_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.119194"}],"container-title":["ACM Transactions on Intelligent Systems and Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696456","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3696456","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:10:14Z","timestamp":1750295414000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696456"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,13]]},"references-count":37,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12,31]]}},"alternative-id":["10.1145\/3696456"],"URL":"https:\/\/doi.org\/10.1145\/3696456","relation":{},"ISSN":["2157-6904","2157-6912"],"issn-type":[{"value":"2157-6904","type":"print"},{"value":"2157-6912","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,13]]},"assertion":[{"value":"2023-09-18","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-08-24","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-12-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}