{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T18:59:11Z","timestamp":1775069951803,"version":"3.50.1"},"reference-count":63,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2021]]},"DOI":"10.1109\/access.2021.3052783","type":"journal-article","created":{"date-parts":[[2021,1,22]],"date-time":"2021-01-22T23:18:17Z","timestamp":1611357497000},"page":"13248-13265","source":"Crossref","is-referenced-by-count":83,"title":["A Survey of the State-of-the-Art Models in Neural Abstractive Text Summarization"],"prefix":"10.1109","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3113-8980","authenticated-orcid":false,"given":"Ayesha Ayub","family":"Syed","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5116-5708","authenticated-orcid":false,"given":"Ford Lumban","family":"Gaol","sequence":"additional","affiliation":[]},{"given":"Tokuro","family":"Matsuo","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","first-page":"1","article-title":"Neural machine translation by jointly learning to align and translate","author":"bahdanau","year":"2015","journal-title":"Proc 33rd Int Conf Learn Represent"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1179"},{"key":"ref33","article-title":"End-to-end abstractive summarization for meetings","author":"zhu","year":"2020","journal-title":"Microsoft Speech and Dialogue Research Group"},{"key":"ref32","first-page":"11328","article-title":"PEGASUS: Pre-training with extracted gap-sentences for abstractive summarization","volume":"119","author":"zhang","year":"2020","journal-title":"Proc 37th Int Conf Mach Learn"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1301"},{"key":"ref30","article-title":"Abstractive text summarization with neural sequence-to-sequence models","author":"aksenov","year":"2019"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1016\/0364-0213(90)90002-E"},{"key":"ref36","first-page":"1","article-title":"Abstractive method of text summarization with sequence to sequence RNNs","author":"masum","year":"2019","journal-title":"Proc 10th Int Conf Comput Commun Netw Technol (ICCCNT)"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i01.5328"},{"key":"ref34","article-title":"Length-controllable abstractive summarization by guiding with summary prototype","author":"saito","year":"2020","journal-title":"arXiv 2001 07331"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1623"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-018-5749-3"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/78.650093"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1007\/s11431-020-1647-3"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1443"},{"key":"ref27","first-page":"5998","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst NIPS"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P18-1063"},{"key":"ref2","author":"zhang","year":"2019","journal-title":"Evaluating the Factual Correctness for Abstractive Summarization"},{"key":"ref1","first-page":"1","article-title":"Text summarization techniques: A brief survey","volume":"8","author":"allahyari","year":"2017","journal-title":"Int J Adv Comput Sci Appl"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D15-1044"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D15-1229"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N16-1012"},{"key":"ref24","article-title":"Distraction-based neural networks for document summarization","author":"chen","year":"2016","journal-title":"Proc 25th Int Joint Conf Artif Intell"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/K16-1028"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1099"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P16-1154"},{"key":"ref50","first-page":"65","article-title":"METEOR: An automatic metric for MT evaluation with improved correlation with human judgments","author":"banerjee","year":"2007","journal-title":"Proc of the Workshop on Statistical Machine Translation"},{"key":"ref51","first-page":"13","article-title":"Analyzing the past to prepare for the future: Writing a literature review","volume":"26","author":"webster","year":"2002","journal-title":"MIS Quart"},{"key":"ref59","first-page":"1","article-title":"A deep reinforced model for abstractive summarization","author":"paulus","year":"2018","journal-title":"Proc 6th Int Conf Learn Represent (ICLR)"},{"key":"ref58","article-title":"Summary-aware attention for social media short text abstractive summarization","author":"wang","year":"0","journal-title":"Neurocomputing"},{"key":"ref57","doi-asserted-by":"crossref","first-page":"581","DOI":"10.1016\/j.procs.2019.09.017","article-title":"Indonesian abstractive text summarization using bidirectional gated recurrent unit","volume":"157","author":"adelia","year":"2019","journal-title":"Procedia Comput Sci"},{"key":"ref56","article-title":"Automatic text summarization of COVID-19 medical research articles using BERT and GPT-2","author":"kieuvongngam","year":"2020","journal-title":"arXiv 2006 01997"},{"key":"ref55","article-title":"CO-search: COVID-19 information retrieval with semantic search, question answering, and abstractive summarization","author":"esteva","year":"2020","journal-title":"arXiv 2006 09595"},{"key":"ref54","first-page":"10384","article-title":"MASS: Masked sequence to sequence pre-training for language generation","author":"song","year":"2019","journal-title":"Proc 36th Int Conf Mach Learn (ICML)"},{"key":"ref53","article-title":"Abstractive text summarization with attention-based mechanism","author":"sanjabi","year":"2018"},{"key":"ref52","article-title":"Bidirectional attentional encoder-decoder model and bidirectional beam search for abstractive summarization","author":"al-sabahi","year":"2018","journal-title":"arXiv 1809 06662"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"ref11","first-page":"1","article-title":"AMHARIC abstractive text summarization","volume":"14","author":"zaki","year":"2020","journal-title":"J Xidian Univ"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref12","doi-asserted-by":"crossref","first-page":"159","DOI":"10.1147\/rd.22.0159","article-title":"The automatic creation of literature abstracts","volume":"2","author":"luhn","year":"1958","journal-title":"IBM J Res Develop"},{"key":"ref13","article-title":"Recent trends in deep learning based abstractive text summarization","volume":"8","author":"rane","year":"2019","journal-title":"Int J Recent Technol Eng"},{"key":"ref14","year":"2019","journal-title":"Towards Automatic Summarization Part 2 Abstractive Methods"},{"key":"ref15","author":"sanad","year":"2019","journal-title":"A Comprehensive Guide to Build your own Language Model in Python"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511809071"},{"key":"ref17","article-title":"A survey on neural network language models","author":"jing","year":"2019","journal-title":"arXiv 1906 03591"},{"key":"ref18","first-page":"1137","article-title":"A neural probabilistic language model","volume":"3","author":"bengio","year":"2003","journal-title":"J Mach Learn Res"},{"key":"ref19","first-page":"1","article-title":"Efficient estimation of word representations in vector space","author":"mikolov","year":"2013","journal-title":"Proc Int Conf Learn Represent (ICLR)ICLR"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-011-9216-z"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1561\/9781601984715"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.5220\/0009723306480655"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33019815"},{"key":"ref8","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"devlin","year":"2019","journal-title":"Proc Conf North Amer Chapter Assoc Comput Linguistics Hum Lang Technol"},{"key":"ref7","article-title":"Neural abstractive text summarization with sequence-to-sequence models","author":"shi","year":"2018","journal-title":"arXiv 1812 02303"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.3115\/1073083.1073135"},{"key":"ref9","article-title":"Improving language understanding by generative pre-training","author":"radford","year":"0","journal-title":"The Open Journal"},{"key":"ref46","author":"khandelwal","year":"2019","journal-title":"Overview of different Optimizers for neural networks"},{"key":"ref45","article-title":"An overview of gradient descent optimization algorithms","author":"ruder","year":"2016","journal-title":"Dblp computer science bibliography"},{"key":"ref48","first-page":"8","article-title":"ROUGE: A package for automatic evaluation of summaries","volume":"34","author":"lin","year":"2004","journal-title":"Proc Jpn Circulat J Conf"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1065"},{"key":"ref42","article-title":"Empirical evaluation of gated recurrent neural networks on sequence modeling","author":"chung","year":"2014","journal-title":"NIPS Deep Learning Workshop"},{"key":"ref41","first-page":"3104","article-title":"Sequence to sequence learning with neural networks","volume":"4","author":"sutskever","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref44","author":"venkatraman","year":"2015","journal-title":"Improving Multi-Step Prediction of Learned Time Series Models"},{"key":"ref43","first-page":"1","article-title":"Sequence level training with recurrent neural networks","author":"ranzato","year":"2016","journal-title":"Proc Int Conf Learn Represent"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9312710\/09328413.pdf?arnumber=9328413","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,26]],"date-time":"2022-01-26T07:18:03Z","timestamp":1643181483000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9328413\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"references-count":63,"URL":"https:\/\/doi.org\/10.1109\/access.2021.3052783","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]}}}