{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T05:05:56Z","timestamp":1750309556239,"version":"3.41.0"},"reference-count":44,"publisher":"Association for Computing Machinery (ACM)","issue":"3","license":[{"start":{"date-parts":[[2025,4,15]],"date-time":"2025-04-15T00:00:00Z","timestamp":1744675200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Institute of Information & Communications Technology Planning & Evaluation"},{"name":"Korea government","award":["Korea government (MSIT)"],"award-info":[{"award-number":["Korea government (MSIT)"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Intell. Syst. Technol."],"published-print":{"date-parts":[[2025,6,30]]},"abstract":"<jats:p>Most studies on abstractive summarization are conducted in a supervised learning framework, aiming to generate a golden summary from the original document. In this process, the model focuses on portions of the document that closely resemble the golden summary to produce a coherent output. Consequently, current methodologies tend to achieve higher performance on extractive datasets compared to abstractive datasets, indicating diminished effectiveness on more abstracted content. To address this, our study proposes a methodology that maintains high effectiveness on abstractive datasets. Specifically, we introduce a multi-task learning approach that incorporates both salient and non-salient information during training. This is implemented by adding a contrastive objective to the fine-tuning phase of an encoder\u2013decoder language model. Salient and non-salient parts are selected based on ROUGE-L F1 scores, and their relationships are learned through a triplet loss function. The proposed method is evaluated on five benchmark summarization datasets, including two extractive and three abstractive datasets. Experimental results demonstrate significant performance improvements on abstractive datasets, particularly those with high levels of abstraction, compared to existing abstractive summarization methods.<\/jats:p>","DOI":"10.1145\/3716851","type":"journal-article","created":{"date-parts":[[2025,2,11]],"date-time":"2025-02-11T16:00:19Z","timestamp":1739289619000},"page":"1-15","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Improving the Summarization Effectiveness of Abstractive Datasets through Contrastive Learning"],"prefix":"10.1145","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-0821-174X","authenticated-orcid":false,"given":"Junho","family":"Shin","sequence":"first","affiliation":[{"name":"Korea Electronics Technology Institute, Seongnam-si, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4199-936X","authenticated-orcid":false,"given":"Younghoon","family":"Lee","sequence":"additional","affiliation":[{"name":"Seoul National University of Science and Technology, Nowon-gu, Republic of Korea"}]}],"member":"320","published-online":{"date-parts":[[2025,4,15]]},"reference":[{"key":"e_1_3_2_2_2","first-page":"5783","volume-title":"Proceedings of the 29th International Conference on Computational Linguistics","author":"An Chenxin","year":"2022","unstructured":"Chenxin An, Ming Zhong, Zhiyong Wu, Qin Zhu, Xuan-Jing Huang, and Xipeng Qiu. 2022. CoLo: A contrastive learning based re-ranking framework for one-stage summarization. In Proceedings of the 29th International Conference on Computational Linguistics, 5783\u20135793."},{"key":"e_1_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.532"},{"key":"e_1_3_2_4_2","first-page":"1597","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Chen Ting","year":"2020","unstructured":"Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. 2020. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning. PMLR, 1597\u20131607."},{"key":"e_1_3_2_5_2","doi-asserted-by":"crossref","first-page":"120","DOI":"10.18653\/v1\/2020.nlposs-1.17","volume-title":"Proceedings of the Second Workshop for NLP Open Source Software (NLP-OSS)","author":"Deutsch Daniel","year":"2020","unstructured":"Daniel Deutsch and Dan Roth. 2020. SacreROUGE: An open-source library for using and developing summarization evaluation metrics. In Proceedings of the Second Workshop for NLP Open Source Software (NLP-OSS). ACM, 120\u2013125. Retrieved from https:\/\/www.aclweb.org\/anthology\/2020.nlposs-1.17"},{"key":"e_1_3_2_6_2","first-page":"4171","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Long and Short Papers, Vol. 1","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Long and Short Papers, Vol. 1, 4171\u20134186."},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.384"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1443"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.3390\/app13137620"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-5409"},{"key":"e_1_3_2_11_2","first-page":"21271","article-title":"Bootstrap your own latent\u2014A new approach to self-supervised learning","volume":"33","author":"Grill Jean-Bastien","year":"2020","unstructured":"Jean-Bastien Grill, Florian Strub, Florent Altch\u00e9, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. 2020. Bootstrap your own latent\u2014A new approach to self-supervised learning. In Advances in Neural Information Processing Systems, Vol. 33, 21271\u201321284.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_12_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Gunel Beliz","year":"2021","unstructured":"Beliz Gunel, Jingfei Du, Alexis Conneau, and Veselin Stoyanov. 2021. Supervised contrastive learning for pre-trained language model fine-tuning. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2006.100"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"e_1_3_2_15_2","first-page":"1","article-title":"Teaching machines to read and comprehend. In","volume":"28","author":"Hermann Karl Moritz","year":"2015","unstructured":"Karl Moritz Hermann, Tomas Kocisky, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, and Phil Blunsom. 2015. Teaching machines to read and comprehend. In Advances in Neural Information Processing Systems, Vol. 28, 1\u20139.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_16_2","first-page":"132","volume-title":"Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Long Papers, Vol. 1","author":"Hsu Wan-Ting","year":"2018","unstructured":"Wan-Ting Hsu, Chieh-Kai Lin, Ming-Ying Lee, Kerui Min, Jing Tang, and Min Sun. 2018. A unified model for extractive and abstractive summarization using inconsistency loss. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Long Papers, Vol. 1, 132\u2013141."},{"key":"e_1_3_2_17_2","first-page":"18661","article-title":"Supervised contrastive learning","volume":"33","author":"Khosla Prannay","year":"2020","unstructured":"Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. 2020. Supervised contrastive learning. In Advances in Neural Information Processing Systems, Vol. 33, 18661\u201318673.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_18_2","first-page":"2519","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Long and Short Papers, Vol. 1","author":"Kim Byeongchang","year":"2019","unstructured":"Byeongchang Kim, Hyunwoo Kim, and Gunhee Kim. 2019. Abstractive summarization of reddit posts with multi-level memory networks. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Long and Short Papers, Vol. 1, 2519\u20132531."},{"key":"e_1_3_2_19_2","first-page":"1","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Lee Seanie","year":"2021","unstructured":"Seanie Lee, Dong Bok Lee, and Sung Ju Hwang. 2021. Contrastive learning with adversarial perturbations for conditional text generation. In Proceedings of the International Conference on Learning Representations, 1\u201325."},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"e_1_3_2_21_2","first-page":"74","volume-title":"Text Summarization Branches Out.","author":"Lin Chin-Yew","year":"2004","unstructured":"Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text Summarization Branches Out. Association for Computational Linguistics, Barcelona, Spain, 74\u201381."},{"key":"e_1_3_2_22_2","unstructured":"Wei Liu Huanqin Wu Wenjing Mu Zhen Li Tao Chen and Dan Nie. 2021. CO2Sum: Contrastive learning for factual-consistent abstractive summarization. arXiv:2112.01147. Retrieved from https:\/\/arxiv.org\/abs\/2112.01147"},{"key":"e_1_3_2_23_2","first-page":"1065","volume-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Short Papers, Vol. 2","author":"Liu Yixin","year":"2021","unstructured":"Yixin Liu and Pengfei Liu. 2021. SimCLS: A simple framework for contrastive learning of abstractive summarization. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Short Papers, Vol. 2, 1065\u20131072."},{"key":"e_1_3_2_24_2","first-page":"2890","volume-title":"Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Long Papers, Vol. 1","author":"Liu Yixin","year":"2022","unstructured":"Yixin Liu, Pengfei Liu, Dragomir Radev, and Graham Neubig. 2022. BRIO: Bringing order to abstractive summarization. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Long Papers, Vol. 1, 2890\u20132903."},{"key":"e_1_3_2_25_2","first-page":"1","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Loshchilov Ilya","year":"2019","unstructured":"Ilya Loshchilov and Frank Hutter. 2019. Decoupled weight decay regularization. In Proceedings of the International Conference on Learning Representations, 1\u20138."},{"key":"e_1_3_2_26_2","doi-asserted-by":"crossref","first-page":"339","DOI":"10.1007\/978-981-13-5934-7_31","article-title":"A comprehensive survey on extractive and abstractive techniques for text summarization","volume":"1","author":"Mahajani Abhishek","year":"2019","unstructured":"Abhishek Mahajani, Vinay Pandya, Isaac Maria, and Deepak Sharma. 2019. A comprehensive survey on extractive and abstractive techniques for text summarization. Ambient Communications and Computer Systems: RACCCS-2018 1 (2019), 339\u2013351.","journal-title":"Ambient Communications and Computer Systems: RACCCS-2018"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/K16-1028"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1206"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.5555\/3455716.3455856"},{"key":"e_1_3_2_30_2","unstructured":"Marc\u2019Aurelio Ranzato Sumit Chopra Michael Auli and Wojciech Zaremba. 2015. Sequence level training with recurrent neural networks. arXiv:1511.06732. Retrieved from https:\/\/arxiv.org\/abs\/1511.06732"},{"key":"e_1_3_2_31_2","doi-asserted-by":"crossref","first-page":"4504","DOI":"10.18653\/v1\/2022.acl-long.309","volume-title":"Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Long Papers, Vol. 1","author":"Ravaut Mathieu","year":"2022","unstructured":"Mathieu Ravaut, Shafiq Joty, and Nancy Chen. 2022. SummaReranker: A multi-task mixture-of-experts re-ranking framework for abstractive summarization. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Long Papers, Vol. 1, 4504\u20134524."},{"key":"e_1_3_2_32_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D15-1044"},{"issue":"12","key":"e_1_3_2_33_2","first-page":"e26752","article-title":"The New York times annotated corpus","volume":"6","author":"Sandhaus Evan","year":"2008","unstructured":"Evan Sandhaus. 2008. The New York times annotated corpus. Linguistic Data Consortium, Philadelphia 6, 12 (2008), e26752.","journal-title":"Linguistic Data Consortium, Philadelphia"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298682"},{"key":"e_1_3_2_35_2","first-page":"1073","volume-title":"Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Long Papers","volume":"1","author":"See Abigail","year":"2017","unstructured":"Abigail See, Peter J. Liu, and Christopher D. Manning. 2017. Get to the point: Summarization with pointer-generator networks. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Long Papers, Vol. 1, 1073\u20131083."},{"key":"e_1_3_2_36_2","first-page":"1","article-title":"Attention is all you need","volume":"30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems, Vol. 30, 1\u201311.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_37_2","doi-asserted-by":"crossref","first-page":"6094","DOI":"10.18653\/v1\/2022.emnlp-main.409","volume-title":"Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing","author":"Wang Fei","year":"2022","unstructured":"Fei Wang, Kaiqiang Song, Hongming Zhang, Lifeng Jin, Sangwoo Cho, Wenlin Yao, Xiaoyang Wang, Muhao Chen, and Dong Yu. 2022. Salience allocation as guidance for abstractive summarization. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 6094\u20136106."},{"key":"e_1_3_2_38_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.451"},{"key":"e_1_3_2_39_2","first-page":"11556","volume-title":"Proceedings of the AAAI Conference on Artificial Intelligence","author":"Xu Shusheng","year":"2022","unstructured":"Shusheng Xu, Xingxing Zhang, Yi Wu, and Furu Wei. 2022. Sequence level contrastive learning for text summarization. In Proceedings of the AAAI Conference on Artificial Intelligence, 11556\u201311565."},{"key":"e_1_3_2_40_2","first-page":"11328","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Zhang Jingqing","year":"2020","unstructured":"Jingqing Zhang, Yao Zhao, Mohammad Saleh, and Peter Liu. 2020. Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the International Conference on Machine Learning. PMLR, 11328\u201311339."},{"key":"e_1_3_2_41_2","first-page":"1","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Zhang Tianyi","year":"2020","unstructured":"Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q. Weinberger, and Yoav Artzi. 2020. BERTScore: Evaluating text generation with BERT. In Proceedings of the International Conference on Learning Representations, 1\u201343."},{"key":"e_1_3_2_42_2","unstructured":"Xingxing Zhang Yiran Liu Xun Wang Pengcheng He Yang Yu Si-Qing Chen Wayne Xiong and Furu Wei. 2022. Momentum calibration for text generation. arXiv:2212.04257. Retrieved from https:\/\/arxiv.org\/abs\/2212.04257"},{"key":"e_1_3_2_43_2","first-page":"1","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"Zhao Yao","year":"2023","unstructured":"Yao Zhao, Mikhail Khalman, Rishabh Joshi, Shashi Narayan, Mohammad Saleh, and Peter J. Liu. 2023. Calibrating sequence likelihood improves conditional language generation. In Proceedings of the 11th International Conference on Learning Representations, 1\u201321."},{"key":"e_1_3_2_44_2","first-page":"1764","volume-title":"Proceedings of the 2021 IEEE International Conference on Big Data (Big Data)","author":"Zheng Chujie","year":"2021","unstructured":"Chujie Zheng, Kunpeng Zhang, Harry Jiannan Wang, Ling Fan, and Zhe Wang. 2021. Enhanced Seq2Seq autoencoder via contrastive learning for abstractive text summarization. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data). IEEE, 1764\u20131771."},{"key":"e_1_3_2_45_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.552"}],"container-title":["ACM Transactions on Intelligent Systems and Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3716851","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3716851","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:53Z","timestamp":1750295933000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3716851"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,15]]},"references-count":44,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,6,30]]}},"alternative-id":["10.1145\/3716851"],"URL":"https:\/\/doi.org\/10.1145\/3716851","relation":{},"ISSN":["2157-6904","2157-6912"],"issn-type":[{"type":"print","value":"2157-6904"},{"type":"electronic","value":"2157-6912"}],"subject":[],"published":{"date-parts":[[2025,4,15]]},"assertion":[{"value":"2023-07-23","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-12-09","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-04-15","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}