{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T16:21:27Z","timestamp":1772554887988,"version":"3.50.1"},"reference-count":33,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,10,26]],"date-time":"2023-10-26T00:00:00Z","timestamp":1698278400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,10,26]],"date-time":"2023-10-26T00:00:00Z","timestamp":1698278400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,10,26]]},"DOI":"10.1109\/wincom59760.2023.10322921","type":"proceedings-article","created":{"date-parts":[[2023,11,22]],"date-time":"2023-11-22T19:07:27Z","timestamp":1700680047000},"page":"1-6","source":"Crossref","is-referenced-by-count":5,"title":["A Comprehensive Survey On Efficient Transformers"],"prefix":"10.1109","author":[{"given":"Yasser","family":"Elouargui","sequence":"first","affiliation":[{"name":"LEYTON,Casablanca,Morocco"}]},{"given":"Mahmoud","family":"Zyate","sequence":"additional","affiliation":[{"name":"LEYTON,Casablanca,Morocco"}]},{"given":"Abdellatif","family":"Sassioui","sequence":"additional","affiliation":[{"name":"LEYTON,Casablanca,Morocco"}]},{"given":"Meriyem","family":"Chergui","sequence":"additional","affiliation":[{"name":"Hassan II University,C3S Laboratory,Casablanca,Morocco"}]},{"given":"Mohamed","family":"El Kamili","sequence":"additional","affiliation":[{"name":"C3S Laboratory, Hassan II University\nCasablanca, Morocco"}]},{"given":"Mohammed","family":"Ouzzif","sequence":"additional","affiliation":[{"name":"Hassan II University,C3S Laboratory,Casablanca,Morocco"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33013159"},{"key":"ref2","article-title":"Diagnosing and enhancing vae models","author":"Dai","year":"2019","journal-title":"arXiv preprint arXiv:1903.05789"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-4009"},{"issue":"8","key":"ref4","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI blog"},{"key":"ref5","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv preprint arXiv:1810.04805"},{"key":"ref6","article-title":"Improving language understanding by generative pre-training","author":"Radford","year":"2018"},{"key":"ref7","article-title":"Xlnet: Generalized autoregressive pretraining for language understanding","volume":"32","author":"Yang","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref9","article-title":"Longformer: The long-document transformer","author":"Beltagy","year":"2020","journal-title":"arXiv preprint arXiv:2004.05150"},{"key":"ref10","first-page":"17 283","article-title":"Big bird: Transformers for longer sequences","volume":"33","author":"Zaheer","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.19"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i16.17664"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00353"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1133"},{"key":"ref15","article-title":"Reformer: The efficient transformer","author":"Kitaev","year":"2020","journal-title":"arXiv preprint arXiv:2001.04451"},{"key":"ref16","article-title":"Rethinking attention with performers","author":"Choromanski","year":"2020","journal-title":"arXiv preprint arXiv:2009.14794"},{"key":"ref17","first-page":"5156","article-title":"Transformers are rnns: Fast autoregressive transformers with linear attention","volume-title":"International Conference on Machine Learning","author":"Katharopoulos"},{"key":"ref18","article-title":"cosformer: Rethinking softmax in attention","author":"Qin","year":"2022","journal-title":"arXiv preprint arXiv:2202.08791"},{"key":"ref19","article-title":"Linformer: Self-attention with linear complexity","author":"Wang","year":"2020","journal-title":"arXiv preprint arXiv:2006.04768"},{"key":"ref20","first-page":"17 723","article-title":"Long-short transformer: Efficient transformers for language and vision","volume":"34","author":"Zhu","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP40776.2020.9053878"},{"key":"ref22","article-title":"Long range arena: A benchmark for efficient transformers","author":"Tay","year":"2020","journal-title":"arXiv preprint arXiv:2011.04006"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-4013"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1016\/j.cose.2020.101740"},{"key":"ref25","article-title":"bytesteady: Fast classification using byte-level n-gram embeddings","author":"Zhang","year":"2021","journal-title":"arXiv preprint arXiv:2106.13302"},{"key":"ref26","first-page":"142","article-title":"Learning word vectors for sentiment analysis","volume-title":"Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies. Portland","author":"Maas"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/2983323.2983769"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.3115\/1699750.1699759"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.32470\/CCN.2018.1116-0"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00276"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1259"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1521"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1147"}],"event":{"name":"2023 10th International Conference on Wireless Networks and Mobile Communications (WINCOM)","location":"Istanbul, Turkiye","start":{"date-parts":[[2023,10,26]]},"end":{"date-parts":[[2023,10,28]]}},"container-title":["2023 10th International Conference on Wireless Networks and Mobile Communications (WINCOM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10322877\/10322879\/10322921.pdf?arnumber=10322921","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,2]],"date-time":"2024-03-02T21:43:40Z","timestamp":1709415820000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10322921\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,26]]},"references-count":33,"URL":"https:\/\/doi.org\/10.1109\/wincom59760.2023.10322921","relation":{},"subject":[],"published":{"date-parts":[[2023,10,26]]}}}