{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,8]],"date-time":"2024-09-08T11:39:00Z","timestamp":1725795540255},"reference-count":23,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,5,27]],"date-time":"2024-05-27T00:00:00Z","timestamp":1716768000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,5,27]],"date-time":"2024-05-27T00:00:00Z","timestamp":1716768000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,5,27]]},"DOI":"10.1109\/iwcmc61514.2024.10592458","type":"proceedings-article","created":{"date-parts":[[2024,7,17]],"date-time":"2024-07-17T17:18:34Z","timestamp":1721236714000},"page":"903-908","source":"Crossref","is-referenced-by-count":0,"title":["Improving Response Prediction: Transformer Integration and Optimizer Evaluation"],"prefix":"10.1109","author":[{"given":"Boutheina Ben","family":"Ismail","sequence":"first","affiliation":[{"name":"SERCOM Laboratory,Tunisia"}]},{"given":"Mohamed","family":"Hamroun","sequence":"additional","affiliation":[{"name":"University of Limoges, Avenue Albert Thomas,XLIM-Lab, UMR CNRS 7252,Limoges,France,87060"}]},{"given":"Bassem","family":"Bensalah","sequence":"additional","affiliation":[{"name":"SERCOM Laboratory,Tunisia"}]},{"given":"Hedi","family":"Sakli","sequence":"additional","affiliation":[{"name":"EITA Consulting 7 rue du chant des oiseaux,Montesson,France,78360"}]}],"member":"263","reference":[{"key":"ref1","first-page":"16","article-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","author":"Jacob Devlin","year":"2019","journal-title":"Computer Science > Computation and Language"},{"key":"ref2","first-page":"17","article-title":"ALBERT: A Lite BERT for Self-supervised Learning of Language Representations","volume-title":"ICLR 2020","author":"Zhenzhong Lan"},{"key":"ref3","first-page":"16","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","author":"Toutanova","year":"2019"},{"first-page":"11","article-title":"Transformer Transducer: A Streamable Speech Recognition Model with Transformer Encoders and RNN-T Loss","volume-title":"ICASSP 2020","key":"ref4"},{"key":"ref5","first-page":"33","article-title":"A CNN-BiLSTM-AM method for stock price prediction","author":"Wenjie Lu","year":"2021","journal-title":"Neural Computing and Applications"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00256"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP40776.2020.9053896"},{"key":"ref8","first-page":"52","article-title":"Machine Reading Comprehension:The Role of Contextualized Language Models","author":"Zhuosheng","year":"2020","journal-title":"Association for Computational Linguistics"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref10","first-page":"17","article-title":"UnifiedQA: Crossing Format Boundaries With a Single QA System","author":"Daniel Khashabi","year":"2020","journal-title":"Findings of the Association for Computational Linguistics 2020"},{"key":"ref11","first-page":"21","article-title":"Recent Advances in Recurrent Neural Networks","author":"Hojjat Salehinejad","year":"2018","journal-title":"Neural and Evolutionary Computing"},{"key":"ref12","first-page":"16","article-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","author":"Jacob Devlin","year":"2019","journal-title":"Computer Science > Computation and Language"},{"key":"ref13","first-page":"10","volume-title":"SQuAD Computer Science Department Stanford University","author":"Rajpurkar","year":"2016"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1264"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"volume-title":"Improving language understanding by generative pretraining","year":"2018","author":"Radford","key":"ref16"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00276"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"ref19","article-title":"How to Develop a Bidirectional LSTM For Sequence Classification in Python with Keras","author":"Brownlee","year":"2019","journal-title":"Machine Learning Mastery"},{"key":"ref20","article-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","author":"Devlin","year":"2018","journal-title":"arXiv preprint arXiv:1810.04805"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1202"},{"volume-title":"ELECTRA: Pre-training Text Encoders as Discriminators","year":"2020","author":"Clark","key":"ref22"},{"volume-title":"RoBERTa: A Robustly Optimized BERT Approach","year":"2019","author":"Liu","key":"ref23"}],"event":{"name":"2024 International Wireless Communications and Mobile Computing (IWCMC)","start":{"date-parts":[[2024,5,27]]},"location":"Ayia Napa, Cyprus","end":{"date-parts":[[2024,5,31]]}},"container-title":["2024 International Wireless Communications and Mobile Computing (IWCMC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10592105\/10592309\/10592458.pdf?arnumber=10592458","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,19]],"date-time":"2024-07-19T05:09:12Z","timestamp":1721365752000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10592458\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,27]]},"references-count":23,"URL":"https:\/\/doi.org\/10.1109\/iwcmc61514.2024.10592458","relation":{},"subject":[],"published":{"date-parts":[[2024,5,27]]}}}