{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T19:00:51Z","timestamp":1763665251687,"version":"3.40.3"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031424472"},{"type":"electronic","value":"9783031424489"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-42448-9_11","type":"book-chapter","created":{"date-parts":[[2023,9,10]],"date-time":"2023-09-10T23:02:21Z","timestamp":1694386941000},"page":"121-132","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Supervised Machine-Generated Text Detectors: Family and\u00a0Scale Matters"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-4055-1601","authenticated-orcid":false,"given":"Areg Mikael","family":"Sarvazyan","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3812-5792","authenticated-orcid":false,"given":"Jos\u00e9 \u00c1ngel","family":"Gonz\u00e1lez","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8922-1242","authenticated-orcid":false,"given":"Paolo","family":"Rosso","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7946-6601","authenticated-orcid":false,"given":"Marc","family":"Franco-Salvador","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,11]]},"reference":[{"key":"11_CR1","unstructured":"Bakhtin, A., Gross, S., Ott, M., Deng, Y., Ranzato, M., Szlam, A.: Real or fake? learning to discriminate machine from human generated text. arXiv preprint arXiv:1906.03351 (2019)"},{"key":"11_CR2","unstructured":"Brown, T., et al.: Language models are few-shot learners. In: Advances in Neural Information Processing Systems, pp. 1877\u20131901 (2020)"},{"key":"11_CR3","doi-asserted-by":"crossref","unstructured":"Conneau, A., et al.: Unsupervised cross-lingual representation learning at scale. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 8440\u20138451 (2020)","DOI":"10.18653\/v1\/2020.acl-main.747"},{"key":"11_CR4","doi-asserted-by":"crossref","unstructured":"Eloundou, T., Manning, S., Mishkin, P., Rock, D.: GPTs are GPTs: an early look at the labor market impact potential of large language models. arXiv preprint arXiv:2303.10130 (2023)","DOI":"10.1126\/science.adj0998"},{"key":"11_CR5","unstructured":"Fandi\u00f1o, A.G., et al.: MarIA: Spanish language models. Procesamiento del Lenguaje Natural (2022)"},{"key":"11_CR6","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1037\/h0057532","volume":"32","author":"R Flesch","year":"1948","unstructured":"Flesch, R.: A new readability yardstick. J. Appl. Psychol. 32, 221 (1948)","journal-title":"J. Appl. Psychol."},{"key":"11_CR7","doi-asserted-by":"crossref","unstructured":"Gehrmann, S., Strobelt, H., Rush, A.: GLTR: statistical detection and visualization of generated text. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 111\u2013116 (2019)","DOI":"10.18653\/v1\/P19-3019"},{"key":"11_CR8","unstructured":"Gu, C., Huang, C., Zheng, X., Chang, K.W., Hsieh, C.J.: Watermarking pre-trained language models with backdooring. arXiv preprint arXiv:2210.07543 (2022)"},{"key":"11_CR9","unstructured":"He, P., Gao, J., Chen, W.: DeBERTav3: improving deBERTa using ELECTRA-style pre-training with gradient-disentangled embedding sharing. In: The Eleventh International Conference on Learning Representations (2023)"},{"key":"11_CR10","doi-asserted-by":"crossref","unstructured":"Ippolito, D., Duckworth, D., Callison-Burch, C., Eck, D.: Automatic detection of generated text is easiest when humans are fooled. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 1808\u20131822 (2020)","DOI":"10.18653\/v1\/2020.acl-main.164"},{"key":"11_CR11","doi-asserted-by":"crossref","unstructured":"Jawahar, G., Abdul-Mageed, M., Lakshmanan, V.S., L.: Automatic detection of machine generated text: a critical survey. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 2296\u20132309 (2020)","DOI":"10.18653\/v1\/2020.coling-main.208"},{"key":"11_CR12","doi-asserted-by":"publisher","first-page":"102274","DOI":"10.1016\/j.lindif.2023.102274","volume":"103","author":"E Kasneci","year":"2023","unstructured":"Kasneci, E., et al.: ChatGPT for good? on opportunities and challenges of large language models for education. Learn. Individ. Differ. 103, 102274 (2023)","journal-title":"Learn. Individ. Differ."},{"key":"11_CR13","unstructured":"Kirchenbauer, J., Geiping, J., Wen, Y., Katz, J., Miers, I., Goldstein, T.: A watermark for large language models. arXiv preprint arXiv:2301.10226 (2023)"},{"key":"11_CR14","unstructured":"Liang, P., et al.: Holistic evaluation of language models. arXiv preprint arXiv:2211.09110 (2022)"},{"key":"11_CR15","doi-asserted-by":"crossref","unstructured":"Maronikolakis, A., Sch\u00fctze, H., Stevenson, M.: Identifying automatically generated headlines using transformers. In: Proceedings of the Fourth Workshop on NLP for Internet Freedom: Censorship, Disinformation, and Propaganda, pp. 1\u20136 (2021)","DOI":"10.18653\/v1\/2021.nlp4if-1.1"},{"key":"11_CR16","unstructured":"Mitchell, E., Lee, Y., Khazatsky, A., Manning, C.D., Finn, C.: DetectGPT: zero-shot machine-generated text detection using probability curvature. arXiv preprint arXiv:2301.11305 (2023)"},{"key":"11_CR17","unstructured":"OpenAI: AI text classifier. OpenAI Blog (2023)"},{"key":"11_CR18","unstructured":"Ouyang, L., et al.: Training language models to follow instructions with human feedback. In: Advances in Neural Information Processing Systems (2022)"},{"key":"11_CR19","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. In: OpenAI (2019)"},{"key":"11_CR20","doi-asserted-by":"crossref","unstructured":"Rodriguez, J., Hay, T., Gros, D., Shamsi, Z., Srinivasan, R.: Cross-domain detection of GPT-2-generated technical text. In: Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1213\u20131233 (2022)","DOI":"10.18653\/v1\/2022.naacl-main.88"},{"key":"11_CR21","unstructured":"Sadasivan, V.S., Kumar, A., Balasubramanian, S., Wang, W., Feizi, S.: Can AI-generated text be reliably detected? arXiv preprint arXiv:2303.11156 (2023)"},{"key":"11_CR22","unstructured":"Sarvazyan, A.M., Gonz\u00e1lez, J.\u00c1., Franco-Salvador, M., Rangel, F., Chulvi, B., Rosso, P.: Overview of AuTexTification at IberLEF 2023: detection and attribution of machine-generated text in multiple domains. In: Procesamiento del Lenguaje Natural (2023)"},{"key":"11_CR23","unstructured":"Scao, T.L., et al.: BLOOM: A 176B-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100 (2022)"},{"key":"11_CR24","unstructured":"Solaiman, I., et al.: Release strategies and the social impacts of language models. arXiv preprint arXiv:1908.09203 (2019)"},{"key":"11_CR25","doi-asserted-by":"crossref","unstructured":"Tourille, J., Sow, B., Popescu, A.: Automatic detection of bot-generated tweets. In: Proceedings of the 1st International Workshop on Multimedia AI against Disinformation, p. 44\u201351 (2022)","DOI":"10.1145\/3512732.3533584"},{"key":"11_CR26","unstructured":"Touvron, H., et al.: LLaMA: open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)"},{"key":"11_CR27","doi-asserted-by":"crossref","unstructured":"Uchendu, A., Le, T., Shu, K., Lee, D.: Authorship attribution for neural text generation. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 8384\u20138395 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.673"},{"key":"11_CR28","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems (2017)"},{"key":"11_CR29","unstructured":"Wolf, T., et al.: Transformers: state-of-the-art natural language processing. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp. 38\u201345 (2020)"},{"key":"11_CR30","unstructured":"Zellers, R., et al.: Defending against neural fake news. In: Advances in Neural Information Processing Systems (2019)"}],"container-title":["Lecture Notes in Computer Science","Experimental IR Meets Multilinguality, Multimodality, and Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-42448-9_11","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,27]],"date-time":"2024-10-27T21:05:54Z","timestamp":1730063154000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-42448-9_11"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031424472","9783031424489"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-42448-9_11","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"11 September 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CLEF","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference of the Cross-Language Evaluation Forum for European Languages","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Thessaloniki","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"clef2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/clef2023.clef-initiative.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"10","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"29% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7 Best of 2022 Labs + 13 Lab Overviews","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}