{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,28]],"date-time":"2025-11-28T12:35:24Z","timestamp":1764333324974,"version":"3.37.3"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T00:00:00Z","timestamp":1690588800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T00:00:00Z","timestamp":1690588800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SN COMPUT. SCI."],"DOI":"10.1007\/s42979-023-02007-5","type":"journal-article","created":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T13:02:16Z","timestamp":1690635736000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["A Comprehensive Survey on Summarization Techniques"],"prefix":"10.1007","volume":"4","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8393-8157","authenticated-orcid":false,"given":"Padma Jyothi","family":"Uppalapati","sequence":"first","affiliation":[]},{"given":"Madhavi","family":"Dabbiru","sequence":"additional","affiliation":[]},{"given":"K. Venkata","family":"Rao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,29]]},"reference":[{"key":"2007_CR1","first-page":"9815","volume":"33","author":"H Lin","year":"2019","unstructured":"Lin H, Ng V. Abstractive summarization: a survey of the state of the art. Proc AAAI Conf Artif Intell. 2019;33:9815\u201322.","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"2007_CR2","doi-asserted-by":"crossref","unstructured":"Nallapati R, Zhai F, Zhou B, Summarunner: a recurrent neural network based sequence model for extractive summarization of documents. In: Thirty-first AAAI Conference on Artificial Intelligence; 2017.","DOI":"10.1609\/aaai.v31i1.10958"},{"key":"2007_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2020.113679","volume":"165","author":"WS El-Kassas","year":"2021","unstructured":"El-Kassas WS, Salama CR, Rafea AA, Mohamed HK. Automatic text summarization: a comprehensive survey. Expert Syst Appl. 2021;165: 113679.","journal-title":"Expert Syst Appl"},{"issue":"2","key":"2007_CR4","doi-asserted-by":"publisher","first-page":"159","DOI":"10.1147\/rd.22.0159","volume":"2","author":"HP Luhn","year":"1958","unstructured":"Luhn HP. The automatic creation of literature abstracts. IBM J Res Dev. 1958;2(2):159\u201365.","journal-title":"IBM J Res Dev"},{"key":"2007_CR5","doi-asserted-by":"crossref","unstructured":"Manjari KU, Rousha S, Sumanth D, Devi JS. Extractive text summarization from web pages using selenium and tf-idf algorithm. In: 2020 4th international conference on trends in electronics and informatics (ICOEI)(48184); 2020. IEEE. pp. 648\u201352","DOI":"10.1109\/ICOEI48184.2020.9142938"},{"issue":"4","key":"2007_CR6","first-page":"16253","volume":"5","author":"P Savyanavar","year":"2016","unstructured":"Savyanavar P, Mehta B, Marathe V, Padvi P, Shewale M. Multi-document summarization using TF-IDF algorithm. Int J Eng Comput Sci. 2016;5(4):16253\u20136.","journal-title":"Int J Eng Comput Sci"},{"issue":"3","key":"2007_CR7","first-page":"1512","volume":"26","author":"RA Albeer","year":"2022","unstructured":"Albeer RA, Al-Shahad HF, Aleqabie HJ, Al-shakarchy ND. Automatic summarization of youtube video transcription text using term frequency-inverse document frequency. Indones J Electr Eng Comput Sci. 2022;26(3):1512\u20139.","journal-title":"Indones J Electr Eng Comput Sci"},{"key":"2007_CR8","doi-asserted-by":"publisher","first-page":"558","DOI":"10.1016\/j.procs.2021.01.040","volume":"179","author":"R Siautama","year":"2021","unstructured":"Siautama R, IA AC, Suhartono D, et al. Extractive hotel review summarization based on TF\/IDF and adjective-noun pairing by considering annual sentiment trends. Procedia Comput Sci. 2021;179:558\u201365.","journal-title":"Procedia Comput Sci"},{"key":"2007_CR9","doi-asserted-by":"crossref","unstructured":"El-Haj M, Kruschwitz U, Fox C. Multi-document Arabic text summarisation. In: 2011 3rd computer science and electronic engineering conference (CEEC); 2011. IEEE. p. 40\u20134","DOI":"10.1109\/CEEC.2011.5995822"},{"key":"2007_CR10","doi-asserted-by":"crossref","unstructured":"Andhale N, Bewoor LA. An overview of text summarization techniques. In: 2016 international conference on computing communication control and automation (ICCUBEA), 2016. p. 1\u20137","DOI":"10.1109\/ICCUBEA.2016.7860024"},{"key":"2007_CR11","doi-asserted-by":"crossref","unstructured":"Allahyari M, Pouriyeh S, Assefi M, Safaei S, Trippe ED, Gutierrez JB, Kochut K. Text summarization techniques: a brief survey. 2017. arXiv preprint arXiv:1707.02268.","DOI":"10.14569\/IJACSA.2017.081052"},{"key":"2007_CR12","doi-asserted-by":"crossref","unstructured":"Lebanoff L, Song K, Liu F. Adapting the neural encoder-decoder framework from single to multi-document summarization. 2018. arXiv preprint arXiv:1808.06218.","DOI":"10.18653\/v1\/D18-1446"},{"key":"2007_CR13","unstructured":"Hermann KM, Kocisky T, Grefenstette E, Espeholt L, Kay W, Suleyman M, Blunsom P. Teaching machines to read and comprehend. Advances in neural information processing systems. 2015;28."},{"key":"2007_CR14","doi-asserted-by":"crossref","unstructured":"Chen D, Bolton J, Manning CD. A thorough examination of the CNN\/daily mail reading comprehension task. 2016. arXiv preprint arXiv:1606.02858.","DOI":"10.18653\/v1\/P16-1223"},{"key":"2007_CR15","unstructured":"Okonda J, Kinini R, Wang P. Neural abstractive summarization on the gigaword dataset."},{"key":"2007_CR16","doi-asserted-by":"crossref","unstructured":"Narayan S, Cohen SB, Lapata M. Don\u2019t give me the details, just the summary! Topic-aware convolutional neural networks for extreme summarization. 2018. arXiv preprint arXiv:1808.08745.","DOI":"10.18653\/v1\/D18-1206"},{"key":"2007_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2020\/9365340","volume":"2020","author":"D Suleiman","year":"2020","unstructured":"Suleiman D, Awajan A. Deep learning based abstractive text summarization: approaches, datasets, evaluation measures, and challenges. Math Probl Eng. 2020;2020:1\u201329.","journal-title":"Math Probl Eng"},{"key":"2007_CR18","unstructured":"Kim B, Kim H, Kim G. Abstractive summarization of reddit posts with multi-level memory networks. 2018. arXiv preprint arXiv:1811.00783."},{"key":"2007_CR19","doi-asserted-by":"crossref","unstructured":"Cohan A, Dernoncourt F, Kim DS, Bui T, Kim S, Chang W, Goharian N. A discourse-aware attention model for abstractive summarization of long documents. 2018. arXiv preprint arXiv:1804.05685.","DOI":"10.18653\/v1\/N18-2097"},{"key":"2007_CR20","doi-asserted-by":"crossref","unstructured":"Sharma E, Li C, Wang L. Bigpatent: a large-scale dataset for abstractive and coherent summarization. 2019. arXiv preprint arXiv:1906.03741.","DOI":"10.18653\/v1\/P19-1212"},{"issue":"5","key":"2007_CR21","doi-asserted-by":"publisher","first-page":"178","DOI":"10.3390\/info10050178","volume":"10","author":"D Maurel","year":"2019","unstructured":"Maurel D, Morale E, Thouvenin N, Ringot P, Turri A. Istex: a database of twenty million scientific papers with a mining tool which uses named entities. Information. 2019;10(5):178.","journal-title":"Information"},{"key":"2007_CR22","doi-asserted-by":"crossref","unstructured":"Gygli M, Grabner H, Riemenschneider H, Gool LV. Creating summaries from user videos. In: European conference on computer vision; 2014. Springer. p. 505\u201320.","DOI":"10.1007\/978-3-319-10584-0_33"},{"key":"2007_CR23","unstructured":"Lv T, Cui L, Vasilijevic M, Wei F. Vt-ssum: a benchmark dataset for video transcript segmentation and summarization. 2021. arXiv preprint arXiv:2106.05606."},{"key":"2007_CR24","unstructured":"Song Y, Vallmitjana J, Stent A, Jaimes A. Tvsum: Summarizing web videos using titles. InProceedings of the IEEE conference on computer vision and pattern recognition; 2015. pp. 5179\u2013187."},{"key":"2007_CR25","doi-asserted-by":"crossref","unstructured":"Kaushal V, Kothawade S, Iyer R, Ramakrishnan G. Realistic video summarization through visiocity: a new benchmark and evaluation framework. In: Proceedings of the 2nd international workshop on AI for smart TV content production, access and delivery; 2020. p. 37\u201344.","DOI":"10.1145\/3422839.3423064"},{"key":"2007_CR26","doi-asserted-by":"publisher","first-page":"3017","DOI":"10.1109\/TIP.2022.3163855","volume":"31","author":"W Zhu","year":"2022","unstructured":"Zhu W, Han Y, Lu J, Zhou J. Relational reasoning over spatial-temporal graphs for video summarization. IEEE Trans Image Process. 2022;31:3017\u201331.","journal-title":"IEEE Trans Image Process"},{"key":"2007_CR27","doi-asserted-by":"crossref","unstructured":"Bonthu S, Sree SR, Krishna PM. Text2pycode: machine translation of natural language intent to python source code. In: International cross-domain conference for machine learning and knowledge extraction; 2021. Springer. p. 51\u201360.","DOI":"10.1007\/978-3-030-84060-0_4"},{"key":"2007_CR28","doi-asserted-by":"crossref","unstructured":"Bonthu S, Rama SS, Krishna PM. Automated short answer grading using deep learning: a survey. In: International cross-domain conference for machine learning and knowledge extraction; 2021. Springer. p. 61\u201378.","DOI":"10.1007\/978-3-030-84060-0_5"},{"key":"2007_CR29","doi-asserted-by":"crossref","unstructured":"Liu Y, Liu P, Radev D, Neubig G. Brio: bringing order to abstractive summarization. 2022. arXiv preprint arXiv:2203.16804.","DOI":"10.18653\/v1\/2022.acl-long.207"},{"key":"2007_CR30","doi-asserted-by":"crossref","unstructured":"Tenney I, Das D, Pavlick E. Bert rediscovers the classical nlp pipeline. 2019. arXiv preprint arXiv:1905.05950.","DOI":"10.18653\/v1\/P19-1452"},{"key":"2007_CR31","unstructured":"Kumar V, Choudhary A, Cho E. Data augmentation using pre-trained transformer models. 2020. arXiv preprint arXiv:2003.02245."},{"key":"2007_CR32","unstructured":"Zhang J, Zhao Y, Saleh M, Liu P. Pegasus: pre-training with extracted gap-sentences for abstractive summarization. In: International conference on machine learning; 2020. PMLR. p. 11328\u20131339."},{"key":"2007_CR33","first-page":"17283","volume":"33","author":"M Zaheer","year":"2020","unstructured":"Zaheer M, Guruganesh G, Dubey KA, Ainslie J, Alberti C, Ontanon S, Pham P, Ravula A, Wang Q, Yang L, et al. Big bird: transformers for longer sequences. Adv Neural Inf Process Syst. 2020;33:17283\u201397.","journal-title":"Adv Neural Inf Process Syst"},{"key":"2007_CR34","doi-asserted-by":"crossref","unstructured":"Ravaut M, Joty S, Chen NF. Summareranker: a multi-task mixture-of-experts re-ranking framework for abstractive summarization. 2022. arXiv preprint arXiv:2203.06569.","DOI":"10.18653\/v1\/2022.acl-long.309"},{"key":"2007_CR35","doi-asserted-by":"publisher","first-page":"24205","DOI":"10.1109\/ACCESS.2018.2829199","volume":"6","author":"K Al-Sabahi","year":"2018","unstructured":"Al-Sabahi K, Zuping Z, Nadher M. A hierarchical structured self-attentive model for extractive document summarization (HSSAS). IEEE Access. 2018;6:24205\u201312.","journal-title":"IEEE Access"},{"key":"2007_CR36","doi-asserted-by":"crossref","unstructured":"Xiao D, Zhang H, Li Y, Sun Y, Tian H, Wu H, Wang H. Ernie-gen: an enhanced multi-flow pre-training and fine-tuning framework for natural language generation. 2020. arXiv preprint arXiv:2001.11314.","DOI":"10.24963\/ijcai.2020\/553"},{"key":"2007_CR37","unstructured":"Aghajanyan A, Shrivastava A, Gupta A, Goyal N, Zettlemoyer L, Gupta S. Better fine-tuning by reducing representational collapse. 2020. arXiv preprint arXiv:2008.03156."},{"key":"2007_CR38","doi-asserted-by":"crossref","unstructured":"Aghajanyan A, Gupta A, Shrivastava A, Chen X, Zettlemoyer L, Gupta S. Muppet: massive multi-task representations with pre-finetuning. 2021. arXiv preprint arXiv:2101.11038.","DOI":"10.18653\/v1\/2021.emnlp-main.468"},{"key":"2007_CR39","unstructured":"Wang P, Yang A, Men R, Lin J, Bai S, Li Z, Ma J, Zhou C, Zhou J, Yang H. Unifying architectures, tasks, and modalities through a simple sequence-to-sequence learning framework. 2022. arXiv preprint arXiv:2202.03052."},{"key":"2007_CR40","first-page":"17283","volume":"33","author":"M Zaheer","year":"2020","unstructured":"Zaheer M, Guruganesh G, Dubey KA, Ainslie J, Alberti C, Ontanon S, Pham P, Ravula A, Wang Q, Yang L, et al. Big bird: transformers for longer sequences. Adv Neural Inf Process Syst. 2020;33:17283\u201397.","journal-title":"Adv Neural Inf Process Syst"},{"key":"2007_CR41","doi-asserted-by":"crossref","unstructured":"Takase S, Kiyono S. Rethinking perturbations in encoder\u2013decoders for fast training. 2021. arXiv preprint arXiv:2104.01853.","DOI":"10.18653\/v1\/2021.naacl-main.460"},{"key":"2007_CR42","doi-asserted-by":"crossref","unstructured":"Yasunaga M, Zhang R, Meelu K, Pareek A, Srinivasan K, Radev D. Graph-based neural multi-document summarization. 2017. arXiv preprint arXiv:1706.06681.","DOI":"10.18653\/v1\/K17-1045"},{"key":"2007_CR43","doi-asserted-by":"crossref","unstructured":"Takase S, Okazaki N. Positional encoding to control output sequence length. 2019. arXiv preprint arXiv:1904.07418.","DOI":"10.18653\/v1\/N19-1401"}],"container-title":["SN Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-023-02007-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s42979-023-02007-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-023-02007-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,29]],"date-time":"2023-07-29T13:28:10Z","timestamp":1690637290000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s42979-023-02007-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,29]]},"references-count":43,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2023,9]]}},"alternative-id":["2007"],"URL":"https:\/\/doi.org\/10.1007\/s42979-023-02007-5","relation":{},"ISSN":["2661-8907"],"issn-type":[{"type":"electronic","value":"2661-8907"}],"subject":[],"published":{"date-parts":[[2023,7,29]]},"assertion":[{"value":"27 September 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 January 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 July 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The study was not supported by any funding.All authors declare that they have no conflicts of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}},{"value":"This article does not contain any studies with human participants or animals performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}},{"value":"With the own interest of authors, we have conducted this survey.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed Consent"}}],"article-number":"560"}}