{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,20]],"date-time":"2026-04-20T20:41:39Z","timestamp":1776717699839,"version":"3.51.2"},"publisher-location":"Cham","reference-count":20,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031363351","type":"print"},{"value":"9783031363368","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-36336-8_67","type":"book-chapter","created":{"date-parts":[[2023,6,29]],"date-time":"2023-06-29T14:02:43Z","timestamp":1688047363000},"page":"428-434","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["A Quantitative Study of\u00a0NLP Approaches to\u00a0Question Difficulty Estimation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5113-4696","authenticated-orcid":false,"given":"Luca","family":"Benedetto","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,30]]},"reference":[{"key":"67_CR1","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"29","DOI":"10.1007\/978-3-030-78292-4_3","volume-title":"Artificial Intelligence in Education","author":"S AlKhuzaey","year":"2021","unstructured":"AlKhuzaey, S., Grasso, F., Payne, T.R., Tamma, V.: A systematic review of data-driven approaches to item difficulty prediction. In: Roll, I., McNamara, D., Sosnovsky, S., Luckin, R., Dimitrova, V. (eds.) AIED 2021. LNCS (LNAI), vol. 12748, pp. 29\u201341. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-78292-4_3"},{"key":"67_CR2","doi-asserted-by":"crossref","unstructured":"Beinborn, L., Zesch, T., Gurevych, I.: Candidate evaluation strategies for improved difficulty prediction of language tests. In: Proceedings of the Tenth Workshop on Innovative Use of NLP for Building Educational Applications, pp. 1\u201311 (2015)","DOI":"10.3115\/v1\/W15-0601"},{"key":"67_CR3","unstructured":"Benedetto, L., Aradelli, G., Cremonesi, P., Cappelli, A., Giussani, A., Turrin, R.: On the application of transformers for estimating the difficulty of multiple-choice questions from text. In: Proceedings of the 16th Workshop on Innovative Use of NLP for Building Educational Applications, pp. 147\u2013157 (2021)"},{"key":"67_CR4","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"43","DOI":"10.1007\/978-3-030-52237-7_4","volume-title":"Artificial Intelligence in Education","author":"L Benedetto","year":"2020","unstructured":"Benedetto, L., Cappelli, A., Turrin, R., Cremonesi, P.: Introducing a framework to assess newly created questions with natural language processing. In: Bittencourt, I.I., Cukurova, M., Muldner, K., Luckin, R., Mill\u00e1n, E. (eds.) AIED 2020. LNCS (LNAI), vol. 12163, pp. 43\u201354. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-52237-7_4"},{"key":"67_CR5","doi-asserted-by":"crossref","unstructured":"Benedetto, L., Cappelli, A., Turrin, R., Cremonesi, P.: R2DE: a NLP approach to estimating IRT parameters of newly generated questions. In: Proceedings of the Tenth International Conference on Learning Analytics & Knowledge, pp. 412\u2013421 (2020)","DOI":"10.1145\/3375462.3375517"},{"key":"67_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3556538","volume":"55","author":"L Benedetto","year":"2022","unstructured":"Benedetto, L., et al.: A survey on recent approaches to question difficulty estimation from text. ACM Comput. Surv. (CSUR) 55, 1\u201337 (2022)","journal-title":"ACM Comput. Surv. (CSUR)"},{"issue":"4","key":"67_CR7","doi-asserted-by":"publisher","first-page":"503","DOI":"10.1177\/0265532215572268","volume":"32","author":"B Culligan","year":"2015","unstructured":"Culligan, B.: A comparison of three test formats to assess word difficulty. Lang. Test. 32(4), 503\u2013520 (2015)","journal-title":"Lang. Test."},{"key":"67_CR8","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (2019)"},{"key":"67_CR9","unstructured":"Ehara, Y.: Building an English vocabulary knowledge dataset of Japanese English-as-a-second-language learners using crowdsourcing. In: Proceedings of the Eleventh International Conference on Language Resources and Evaluation (2018)"},{"issue":"3","key":"67_CR10","doi-asserted-by":"publisher","first-page":"243","DOI":"10.1007\/s11257-009-9063-7","volume":"19","author":"M Feng","year":"2009","unstructured":"Feng, M., Heffernan, N., Koedinger, K.: Addressing the assessment challenge with an online system that tutors as it assesses. User Model. User-Adapt. Interact. 19(3), 243\u2013266 (2009)","journal-title":"User Model. User-Adapt. Interact."},{"key":"67_CR11","doi-asserted-by":"crossref","unstructured":"Hou, J., Maximilian, K., Quecedo, J.M.H., Stoyanova, N., Yangarber, R.: Modeling language learning using specialized Elo rating. In: Proceedings of the Fourteenth Workshop on Innovative Use of NLP for Building Educational Applications (2019)","DOI":"10.18653\/v1\/W19-4451"},{"key":"67_CR12","unstructured":"Huang, Y.T., Chen, M.C., Sun, Y.S.: Development and evaluation of a personalized computer-aided question generation for English learners to improve proficiency and correct mistakes. arXiv preprint arXiv:1808.09732 (2018)"},{"key":"67_CR13","doi-asserted-by":"crossref","unstructured":"Lai, G., Xie, Q., Liu, H., Yang, Y., Hovy, E.: RACE: large-scale reading comprehension dataset from examinations. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 785\u2013794 (2017)","DOI":"10.18653\/v1\/D17-1082"},{"key":"67_CR14","unstructured":"Liang, Y., Li, J., Yin, J.: A new multi-choice reading comprehension dataset for curriculum learning. In: Asian Conference on Machine Learning. PMLR (2019)"},{"key":"67_CR15","doi-asserted-by":"crossref","unstructured":"Loginova, E., Benedetto, L., Benoit, D., Cremonesi, P.: Towards the application of calibrated Transformers to the unsupervised estimation of question difficulty from text. In: RANLP 2021, pp. 846\u2013855. INCOMA (2021)","DOI":"10.26615\/978-954-452-072-4_097"},{"key":"67_CR16","unstructured":"Sanh, V., Debut, L., Chaumond, J., Wolf, T.: DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019)"},{"issue":"2","key":"67_CR17","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1177\/0265532215623581","volume":"34","author":"J Trace","year":"2017","unstructured":"Trace, J., Brown, J.D., Janssen, G., Kozhevnikova, L.: Determining cloze item difficulty from item and passage characteristics across different learner backgrounds. Lang. Test. 34(2), 151\u2013174 (2017)","journal-title":"Lang. Test."},{"key":"67_CR18","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NIPS (2017)"},{"key":"67_CR19","unstructured":"Yaneva, V., Baldwin, P., Mee, J., et al.: Predicting the difficulty of multiple choice questions in a high-stakes medical exam. In: Proceedings of the Fourteenth Workshop on Innovative Use of NLP for Building Educational Applications (2019)"},{"key":"67_CR20","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Tao, C.: Multi-task BERT for problem difficulty prediction. In: 2020 International Conference on Communications, Information System and Computer Engineering (CISCE), pp. 213\u2013216. IEEE (2020)","DOI":"10.1109\/CISCE50729.2020.00048"}],"container-title":["Communications in Computer and Information Science","Artificial Intelligence in Education. Posters and Late Breaking Results, Workshops and Tutorials, Industry and Innovation Tracks, Practitioners, Doctoral Consortium and Blue Sky"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-36336-8_67","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,29]],"date-time":"2023-06-29T14:08:30Z","timestamp":1688047710000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-36336-8_67"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031363351","9783031363368"],"references-count":20,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-36336-8_67","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"30 June 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"AIED","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Intelligence in Education","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tokyo","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 July 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 July 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"aied2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.aied2023.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"311","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"53","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"26","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"17% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}