{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,20]],"date-time":"2026-04-20T10:23:13Z","timestamp":1776680593478,"version":"3.51.2"},"reference-count":33,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2020,12,1]],"date-time":"2020-12-01T00:00:00Z","timestamp":1606780800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2020,12,11]],"date-time":"2020-12-11T00:00:00Z","timestamp":1607644800000},"content-version":"vor","delay-in-days":10,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Big Data"],"published-print":{"date-parts":[[2020,12]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Natural language processing has witnessed remarkable progress with the advent of deep learning techniques. Text summarization, along other tasks like text translation and sentiment analysis, used deep neural network models to enhance results. The new methods of text summarization are subject to a sequence-to-sequence framework of encoder\u2013decoder model, which is composed of neural networks trained jointly on both input and output. Deep neural networks take advantage of big datasets to improve their results. These networks are supported by the attention mechanism, which can deal with long texts more efficiently by identifying focus points in the text. They are also supported by the copy mechanism that allows the model to copy words from the source to the summary directly. In this research, we are re-implementing the basic summarization model that applies the sequence-to-sequence framework on the Arabic language, which has not witnessed the employment of this model in the text summarization before. Initially, we build an Arabic data set of summarized article headlines. This data set consists of approximately 300 thousand entries, each consisting of an article introduction and the headline corresponding to this introduction. We then apply baseline summarization models to the previous data set and compare the results using the ROUGE scale.<\/jats:p>","DOI":"10.1186\/s40537-020-00386-7","type":"journal-article","created":{"date-parts":[[2020,12,11]],"date-time":"2020-12-11T07:18:49Z","timestamp":1607671129000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":58,"title":["Arabic text summarization using deep learning approach"],"prefix":"10.1186","volume":"7","author":[{"given":"Molham","family":"Al-Maleh","sequence":"first","affiliation":[]},{"given":"Said","family":"Desouki","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,12,11]]},"reference":[{"key":"386_CR1","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1007\/978-3-319-16486-1_23","volume-title":"New contributions in information systems and technologies","author":"R Belkebir","year":"2015","unstructured":"Belkebir R, Guessoum A. A supervised approach to Arabic text summarization using adaboost. In: New contributions in information systems and technologies. New York: Springer; 2015. p. 227\u201336."},{"issue":"17","key":"386_CR2","doi-asserted-by":"publisher","first-page":"38","DOI":"10.5120\/12980-0237","volume":"74","author":"I Imam","year":"2013","unstructured":"Imam I, Nounou N, Hamouda A, Abdul Khalek HA. An ontology-based summarization system for Arabic documents (ossad). IJCA. 2013;74(17):38\u201343.","journal-title":"IJCA"},{"issue":"1","key":"386_CR3","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1016\/S0004-3702(02)00222-9","volume":"139","author":"K Knight","year":"2002","unstructured":"Knight K, Marcu D. Summarization beyond sentence extraction: a probabilistic approach to sentence compression. Artif Intell. 2002;139(1):91\u2013107.","journal-title":"Artif Intell"},{"issue":"4","key":"386_CR4","doi-asserted-by":"publisher","first-page":"527","DOI":"10.1162\/089120102762671972","volume":"28","author":"H Jing","year":"2002","unstructured":"Jing H. Using hidden markov modeling to decompose human-written summaries. Comput Linguist. 2002;28(4):527\u201343.","journal-title":"Comput Linguist"},{"key":"386_CR5","doi-asserted-by":"crossref","unstructured":"Dorr B, Zajic D, Schwartz R. Hedge trimmer: a parse-and-trim approach to headline generation. In: Proceedings of the HLTNAACL03 on text summarization workshop, vol. 5. Edmonton: Association for Computational Linguistics; 2003. p. 1\u20138.","DOI":"10.3115\/1119467.1119468"},{"key":"386_CR6","unstructured":"Zajic D, Dorr B, Schwartz R. Bbn\/umd at duc-2004: Topiary. In: Proceedings of the HLT-NAACL 2004 document understanding workshop. Boston: Association for Computational Linguistics; 2004. p. 112\u20139."},{"key":"386_CR7","doi-asserted-by":"crossref","unstructured":"Cohn T, Lapata M. Sentence compression beyond word deletion. In: Proceedings of the 22nd international conference on computational linguistics, vol. 1. Manchester: Association for Computational Linguistics; 2008. p. 137\u201344.","DOI":"10.3115\/1599081.1599099"},{"key":"386_CR8","unstructured":"Woodsend K, Feng Y, Lapata M. Generation with quasi-synchronous grammar. In: Proceedings of the 2010 conference on empirical methods in natural language processing. Cambridge: Association for Computational Linguistics; 2010. p. 513\u201323."},{"key":"386_CR9","doi-asserted-by":"crossref","unstructured":"Rush AM, Chopra S, Weston J. A neural attention model for abstractive sentence summarization. arXiv Prepr arXiv:1509.00685. 2015.","DOI":"10.18653\/v1\/D15-1044"},{"key":"386_CR10","doi-asserted-by":"crossref","unstructured":"Gu J, Lu Z, Li H, Li VO. Incorporating copying mechanism in sequence-to-sequence learning. arXiv Prepr arXiv:1603.06393. 2016.","DOI":"10.18653\/v1\/P16-1154"},{"key":"386_CR11","doi-asserted-by":"crossref","unstructured":"See A, Liu PJ, Manning CD. Get to the point: summarization with pointer-generator networks. arXiv Prepr arXiv:1704.04368. 2017.","DOI":"10.18653\/v1\/P17-1099"},{"key":"386_CR12","unstructured":"Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate. arXiv Prepr arXiv:1409.0473. 2014."},{"key":"386_CR13","unstructured":"El-Shishtawy T, El-Ghannam F. Keyphrase based Arabic summarizer (KPAS). arXiv Prepr arXiv:1206.5384. 2012."},{"key":"386_CR14","doi-asserted-by":"crossref","unstructured":"Azmi A. Al-thanyyan S. Ikhtasir\u2014a user selected compression ratio Arabic text summarization system. In: 2009 international conference on natural language processing and knowledge engineering. Dalian: IEEE; 2009. p. 1\u20137.","DOI":"10.1109\/NLPKE.2009.5313732"},{"key":"386_CR15","unstructured":"AlSanie W. Towards an infrastructure for Arabic text summarization using rhetorical structure theory. M.Sc. Thesis, Dept. of Computer Science, King Saud University, Riyadh, Saudi Arabia. 2005."},{"key":"386_CR16","unstructured":"Douzidia FS, Lapalme G. Lakhas, an Arabic summarization system. In: Proceedings of 2004 document understanding conference (DUC2004). Boston: NIST; 2004. p. 260\u201373."},{"key":"386_CR17","first-page":"62","volume":"2","author":"A Haboush","year":"2012","unstructured":"Haboush A, Al-Zoubi M. Arabic text summarization model using clustering techniques. World Comput Sci Inf Technol J. 2012;2:62\u20137.","journal-title":"World Comput Sci Inf Technol J"},{"key":"386_CR18","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1186\/s40537-019-0219-y","volume":"6","author":"IH Sarker","year":"2019","unstructured":"Sarker IH, Kayes ASM, Watters P. Effectiveness analysis of machine learning classification models for predicting personalized context-aware smartphone usage. J Big Data. 2019;6:57.","journal-title":"J Big Data"},{"key":"386_CR19","unstructured":"Sobh I, Darwish N, Fayek M. An optimized dual classification system for arabic extractive generic text summarization. M.Sc. Thesis, Faculty of engineering, Cairo University, Giza, Egypt. 2009."},{"key":"386_CR20","first-page":"3104","volume":"4","author":"I Sutskever","year":"2014","unstructured":"Sutskever I, Vinyals O, Le QV. Sequence to sequence learning with neural networks. Adv Neural Inf Process Syst. 2014;4:3104\u201312.","journal-title":"Adv Neural Inf Process Syst"},{"key":"386_CR21","unstructured":"Nallapati R, Zhou B, Ma M. Classify or select: neural architectures for extractive document summarization. arXiv Prepr arXiv:1611.04244. 2016."},{"key":"386_CR22","doi-asserted-by":"crossref","unstructured":"Chopra S, Auli M, Rush AM. Abstractive sentence summarization with attentive recurrent neural networks. In: Proceedings of the 2016 conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. San Diego: Association for Computational Linguistics; 2016. p. 93\u20138.","DOI":"10.18653\/v1\/N16-1012"},{"key":"386_CR23","unstructured":"Vinyals O, Fortunato M, Jaitly N. Pointer networks. In: Proceedings of the 28th international conference on neural information processing systems, vol. 2. Montreal: MIT Press; 2015. p. 2692\u2013700."},{"key":"386_CR24","doi-asserted-by":"crossref","unstructured":"Tu Z, Lu Z, Liu Y, Liu X, Li H. Modeling coverage for neural machine translation. arXiv Prepr arXiv:1601.04811. 2016.","DOI":"10.18653\/v1\/P16-1008"},{"key":"386_CR25","unstructured":"Elmadani K N, Elgezouli M, Showk A. BERT fine-tuning for Arabic text summarization. arXiv Prepr arXiv:2004.14135. 2020."},{"key":"386_CR26","unstructured":"Devlin J, Chang MW, Lee K, Toutanova K. Bert: pre-training of deep bidirectional transformers for language understanding. arXiv Prepr arXiv:1810.04805. 2018."},{"key":"386_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2018.10.486","author":"M Helmy","year":"2018","unstructured":"Helmy M, Vigneshram RM, Serra G, Tasso C. Applying deep learning for Arabic keyphrase extraction. Procedia Comput Sci. 2018. https:\/\/doi.org\/10.1016\/j.procs.2018.10.486.","journal-title":"Procedia Comput Sci"},{"key":"386_CR28","unstructured":"Wu Y, Schuster M, Chen Z, Le QV, Norouzi M, Macherey W, et al. Googles neural machine translation system: bridging the gap between human and machine translation. arXiv Prepr arXiv:1609.08144. 2016."},{"key":"386_CR29","doi-asserted-by":"crossref","unstructured":"Gehrmann S, Deng Y, Rush A. Bottom-up abstractive summarization. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Brussels: Association for Computational Linguistics; 2018. p. 4098\u2013109.","DOI":"10.18653\/v1\/D18-1443"},{"key":"386_CR30","unstructured":"Mawdoo3. https:\/\/mawdoo3.com\/. Accessed 02 Feb 2019."},{"key":"386_CR31","first-page":"2121","volume":"12","author":"J Duchi","year":"2011","unstructured":"Duchi J, Hazan E, Singer Y. Adaptive subgradient methods for online learning and stochastic optimization. J Mach Learn Res. 2011;12:2121\u201359.","journal-title":"J Mach Learn Res"},{"key":"386_CR32","doi-asserted-by":"crossref","unstructured":"Klein G, Kim Y, Deng Y, Senellart J, Rush AM. Opennmt: open-source toolkit for neural machine translation. arXiv Prepr arXiv:1701.02810. 2017.","DOI":"10.18653\/v1\/P17-4012"},{"key":"386_CR33","doi-asserted-by":"crossref","unstructured":"Manning CD, Surdeanu M, Bauer J, Finkel J, Bethard SJ, McClosky D. The Stanford CoreNLP natural language processing toolkit. In: Proceedings of 52nd annual meeting of the association for computational linguistics: system demonstrations. Baltimore: Association for Computational Linguistics; 2014. p. 55\u201360.","DOI":"10.3115\/v1\/P14-5010"}],"updated-by":[{"DOI":"10.1186\/s40537-021-00440-y","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2021,4,9]],"date-time":"2021-04-09T00:00:00Z","timestamp":1617926400000}}],"container-title":["Journal of Big Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s40537-020-00386-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1186\/s40537-020-00386-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s40537-020-00386-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,4,9]],"date-time":"2021-04-09T07:03:52Z","timestamp":1617951832000},"score":1,"resource":{"primary":{"URL":"https:\/\/journalofbigdata.springeropen.com\/articles\/10.1186\/s40537-020-00386-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,12]]},"references-count":33,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2020,12]]}},"alternative-id":["386"],"URL":"https:\/\/doi.org\/10.1186\/s40537-020-00386-7","relation":{"correction":[{"id-type":"doi","id":"10.1186\/s40537-021-00440-y","asserted-by":"object"}]},"ISSN":["2196-1115"],"issn-type":[{"value":"2196-1115","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,12]]},"assertion":[{"value":"11 February 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 November 2020","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 December 2020","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 April 2021","order":4,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":5,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"A Correction to this paper has been published:","order":6,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"https:\/\/doi.org\/10.1186\/s40537-021-00440-y","URL":"https:\/\/doi.org\/10.1186\/s40537-021-00440-y","order":7,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare that they have no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"109"}}