{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T20:19:44Z","timestamp":1770495584168,"version":"3.49.0"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031353192","type":"print"},{"value":"9783031353208","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-35320-8_44","type":"book-chapter","created":{"date-parts":[[2023,6,13]],"date-time":"2023-06-13T05:01:53Z","timestamp":1686632513000},"page":"577-587","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["RoBERTweet: A BERT Language Model for\u00a0Romanian Tweets"],"prefix":"10.1007","author":[{"given":"Iulian-Marius","family":"T\u0103iatu","sequence":"first","affiliation":[]},{"given":"Andrei-Marius","family":"Avram","sequence":"additional","affiliation":[]},{"given":"Dumitru-Clementin","family":"Cercel","sequence":"additional","affiliation":[]},{"given":"Florin","family":"Pop","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,14]]},"reference":[{"key":"44_CR1","unstructured":"Avram, A.M., et al.: Distilling the knowledge of Romanian BERTs using multiple teachers. In: Proceedings of the Thirteenth Language Resources and Evaluation Conference, Marseille, France, pp. 374\u2013384. European Language Resources Association (2022)"},{"key":"44_CR2","unstructured":"Ciobotaru, A., Constantinescu, M.V., Dinu, L.P., Dumitrescu, S.: RED v2: enhancing red dataset for multi-label emotion detection. In: Proceedings of the Thirteenth Language Resources and Evaluation Conference, pp. 1392\u20131399 (2022)"},{"key":"44_CR3","doi-asserted-by":"crossref","unstructured":"Ciobotaru, A., Dinu, L.P.: RED: a novel dataset for Romanian emotion detection from tweets. In: Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP 2021), pp. 291\u2013300 (2021)","DOI":"10.26615\/978-954-452-072-4_034"},{"key":"44_CR4","unstructured":"Clark, K., Luong, M.T., Le, Q.V., Manning, C.D.: ELECTRA: pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555 (2020)"},{"key":"44_CR5","doi-asserted-by":"publisher","first-page":"273","DOI":"10.1007\/BF00994018","volume":"20","author":"C Cortes","year":"1995","unstructured":"Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20, 273\u2013297 (1995)","journal-title":"Mach. Learn."},{"key":"44_CR6","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171\u20134186 (2019)"},{"key":"44_CR7","doi-asserted-by":"crossref","unstructured":"Dumitrescu, S., Avram, A.M., Pyysalo, S.: The birth of Romanian BERT. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 4324\u20134328 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.387"},{"key":"44_CR8","unstructured":"Dumitrescu, S.D., et al.: LiRo: benchmark and leaderboard for Romanian language tasks. In: Thirty-Fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1) (2021)"},{"key":"44_CR9","doi-asserted-by":"crossref","unstructured":"Guo, Y., Rennard, V., Xypolopoulos, C., Vazirgiannis, M.: BERTweetFR: domain adaptation of pre-trained language models for French tweets. In: Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021), pp. 445\u2013450 (2021)","DOI":"10.18653\/v1\/2021.wnut-1.49"},{"key":"44_CR10","unstructured":"Hoefels, D.C., \u00c7\u00f6ltekin, \u00c7., M\u0103droane, I.D.: CoRoSeOf-an annotated corpus of Romanian sexist and offensive tweets. In: Proceedings of the Thirteenth Language Resources and Evaluation Conference, pp. 2269\u20132281 (2022)"},{"key":"44_CR11","doi-asserted-by":"crossref","unstructured":"Joulin, A., Grave, \u00c9., Bojanowski, P., Mikolov, T.: Bag of tricks for efficient text classification. In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, pp. 427\u2013431 (2017)","DOI":"10.18653\/v1\/E17-2068"},{"key":"44_CR12","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Proceedings of the 3rd International Conference on Learning Representations. In: ICLR 2015 (2015)"},{"key":"44_CR13","doi-asserted-by":"crossref","unstructured":"Koto, F., Lau, J.H., Baldwin, T.: IndoBERTweet: a pretrained language model for Indonesian twitter with effective domain-specific vocabulary initialization. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 10660\u201310668 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.833"},{"key":"44_CR14","unstructured":"Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: ALBERT: a lite BERT for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942 (2019)"},{"key":"44_CR15","unstructured":"Liu, Y., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"44_CR16","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. In: International Conference on Learning Representations (2019)"},{"key":"44_CR17","doi-asserted-by":"crossref","unstructured":"Masala, M., et al.: jurBERT: a Romanian BERT model for legal judgement prediction. In: Proceedings of the Natural Legal Language Processing Workshop 2021, pp. 86\u201394 (2021)","DOI":"10.18653\/v1\/2021.nllp-1.8"},{"key":"44_CR18","doi-asserted-by":"crossref","unstructured":"Masala, M., Ruseti, S., Dascalu, M.: RoBERT-a Romanian BERT model. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 6626\u20136637 (2020)","DOI":"10.18653\/v1\/2020.coling-main.581"},{"key":"44_CR19","doi-asserted-by":"crossref","unstructured":"Nguyen, D.Q., Vu, T., Nguyen, A.T.: BERTweet: a pre-trained language model for English tweets. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp. 9\u201314 (2020)","DOI":"10.18653\/v1\/2020.emnlp-demos.2"},{"issue":"4","key":"44_CR20","doi-asserted-by":"publisher","first-page":"57","DOI":"10.3390\/computers11040057","volume":"11","author":"DC Nicolae","year":"2022","unstructured":"Nicolae, D.C., Yadav, R.K., Tufi\u015f, D.: A lite Romanian BERT: ALR-BERT. Computers 11(4), 57 (2022)","journal-title":"Computers"},{"key":"44_CR21","unstructured":"P\u0103i\u015f, V., Mititelu, V.B., Irimia, E., Mitrofan, M., Gasan, C.L., Micu, R.: Romanian micro-blogging named entity recognition including health-related entities. In: Proceedings of The Seventh Workshop on Social Media Mining for Health Applications, Workshop & Shared Task, pp. 190\u2013196 (2022)"},{"key":"44_CR22","unstructured":"P\u00e9rez, J.M., Furman, D.A., Alemany, L.A., Luque, F.M.: RoBERTuito: a pre-trained language model for social media text in Spanish. In: Proceedings of the Thirteenth Language Resources and Evaluation Conference, pp. 7235\u20137243 (2022)"},{"issue":"8","key":"44_CR23","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)","journal-title":"OpenAI Blog"},{"key":"44_CR24","unstructured":"Su\u00e1rez, P.J.O., Sagot, B., Romary, L.: Asynchronous pipeline for processing huge corpora on medium to low resource infrastructures. In: 7th Workshop on the Challenges in the Management of Large Corpora (CMLC-7). Leibniz-Institut f\u00fcr Deutsche Sprache (2019)"},{"key":"44_CR25","unstructured":"Tiedemann, J.: Parallel data, tools and interfaces in opus. In: Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC 2012), pp. 2214\u20132218 (2012)"},{"key":"44_CR26","doi-asserted-by":"crossref","unstructured":"Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., Bowman, S.: GLUE: a multi-task benchmark and analysis platform for natural language understanding. In: Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pp. 353\u2013355 (2018)","DOI":"10.18653\/v1\/W18-5446"},{"key":"44_CR27","unstructured":"Wolf, T., et al.: Transformers: state-of-the-art natural language processing. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp. 38\u201345 (2020)"},{"key":"44_CR28","unstructured":"Zampieri, M., et al.: Language variety identification with true labels. arXiv preprint arXiv:2303.01490 (2023)"},{"key":"44_CR29","doi-asserted-by":"crossref","unstructured":"Zhang, X., et al.: TwHIN-BERT: a socially-enriched pre-trained language model for multilingual tweet representations. arXiv preprint arXiv:2209.07562 (2022)","DOI":"10.1145\/3580305.3599921"}],"container-title":["Lecture Notes in Computer Science","Natural Language Processing and Information Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-35320-8_44","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,15]],"date-time":"2023-12-15T00:18:42Z","timestamp":1702599522000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-35320-8_44"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031353192","9783031353208"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-35320-8_44","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"14 June 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"NLDB","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Applications of Natural Language to Information Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Derby","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 June 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 June 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"nldb2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.derby.ac.uk\/events\/latest-events\/nldb-2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy Chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"89","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"31","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"14","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}