{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T20:43:01Z","timestamp":1775594581133,"version":"3.50.1"},"reference-count":74,"publisher":"Association for Computing Machinery (ACM)","issue":"2","license":[{"start":{"date-parts":[[2025,2,15]],"date-time":"2025-02-15T00:00:00Z","timestamp":1739577600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key R & D Program of China","doi-asserted-by":"crossref","award":["2022YFC3303600"],"award-info":[{"award-number":["2022YFC3303600"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Key Laboratory of Smart Education of Guangdong Higher Education Institutes, Jinan University","award":["2022LSYS003"],"award-info":[{"award-number":["2022LSYS003"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Knowl. Discov. Data"],"published-print":{"date-parts":[[2025,2,28]]},"abstract":"<jats:p>\n            Knowledge tracing (KT) plays a crucial role in predicting students\u2019 future performance by analyzing their historical learning processes. Deep neural networks (DNNs) have shown great potential in solving the KT problem. However, there still exist some important challenges when applying deep learning techniques to model the KT process. The first challenge lies in modeling the individual question information. This is crucial because students\u2019 knowledge acquisition on questions that share the same set of knowledge components (KCs) may vary significantly. However, due to the large question bank, the average number of interactions per question may not be sufficient. This limitation can potentially result in overfitting of the question embedding and inaccurate question knowledge acquisition state that relies on its corresponding question representation. Furthermore, there is a considerable portion of questions receiving relatively less interaction from students in comparison to the majority of questions. This can further increase the risk of overfitting and lower the accuracy of the obtained question knowledge acquisition state. The second challenge lies in interpreting the prediction results from existing deep learning-based KT models. In real-world applications, while it may not be necessary to have complete transparency and interpretability of the model parameters, it is crucial to present the model\u2019s prediction results in a manner that teachers find interpretable. This makes teachers accept the rationale behind the prediction results and utilize them to design teaching activities and tailored learning strategies for students. However, the inherent black-box nature of deep learning techniques often poses a hurdle for teachers to fully embrace the model\u2019s prediction results. To address these challenges, we propose a Question-centric Multi-experts Contrastive Learning framework for KT called Q-MCKT. This framework explicitly models students\u2019 knowledge acquisition state at both the question and concept levels. It leverages the mixture of experts technique to capture a more robust and accurate knowledge acquisition state in both question and concept levels for prediction. Additionally, a fine-grained question-centric contrastive learning task is introduced to enhance the representations of less interactive questions and improve the accuracy of their corresponding question knowledge acquisition states. Moreover, Q-MCKT utilizes an item response theory-based prediction layer to generate interpretable prediction results based on the knowledge acquisition states obtained from the question and concept knowledge acquisition modules. We evaluate the proposed Q-MCKT framework on four public real-world educational datasets. The experimental results demonstrate that our approach outperforms a wide range of deep learning-based KT models in terms of prediction accuracy while maintaining better model interpretability. To ensure reproducibility, we have provided all the datasets and code on our website at\n            <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" ext-link-type=\"uri\" xlink:href=\"https:\/\/github.com\/rattlesnakey\/Q-MCKT\">https:\/\/github.com\/rattlesnakey\/Q-MCKT<\/jats:ext-link>\n            .\n          <\/jats:p>","DOI":"10.1145\/3674840","type":"journal-article","created":{"date-parts":[[2024,6,28]],"date-time":"2024-06-28T18:04:16Z","timestamp":1719597856000},"page":"1-25","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["A Question-centric Multi-experts Contrastive Learning Framework for Improving the Accuracy and Interpretability of Deep Sequential Knowledge Tracing Models"],"prefix":"10.1145","volume":"19","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-1991-2163","authenticated-orcid":false,"given":"Hengyuan","family":"Zhang","sequence":"first","affiliation":[{"name":"Shenzhen International Graduate School, Tsinghua University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0491-307X","authenticated-orcid":false,"given":"Zitao","family":"Liu","sequence":"additional","affiliation":[{"name":"Guangdong Institute of Smart Education, Jinan University, Guangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-8945-2165","authenticated-orcid":false,"given":"Chenming","family":"Shang","sequence":"additional","affiliation":[{"name":"Shenzhen International Graduate School, Tsinghua University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-4027-6091","authenticated-orcid":false,"given":"Dawei","family":"Li","sequence":"additional","affiliation":[{"name":"Halicioglu Data Science Institute, University of California, San Diego, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4260-1395","authenticated-orcid":false,"given":"Yong","family":"Jiang","sequence":"additional","affiliation":[{"name":"Shenzhen International Graduate School, Tsinghua University, Shenzhen, China"}]}],"member":"320","published-online":{"date-parts":[[2025,2,15]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1145\/3331184.3331195"},{"key":"e_1_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.coling-main.449"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.804"},{"key":"e_1_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-69132-7_44"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1007\/11774303_17"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i12.26661"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2018.00019"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1145\/3386527.3405945"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-52240-7_13"},{"key":"e_1_3_2_11_2","volume-title":"Proceedings of the NIPS 2014 Workshop on Deep Learning","author":"Chung Junyoung","year":"2014","unstructured":"Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. In Proceedings of the NIPS 2014 Workshop on Deep Learning."},{"key":"e_1_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-78270-2_20"},{"key":"e_1_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.1007\/BF01099821"},{"key":"e_1_3_2_14_2","first-page":"5547","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Du Nan","year":"2022","unstructured":"Nan Du, Yanping Huang, Andrew M Dai, Simon Tong, Dmitry Lepikhin, Yuanzhong Xu, Maxim Krikun, Yanqi Zhou, Adams Wei Yu, Orhan Firat, Barret Zoph, Liam Fedus, Maarten Bosma, Zongwei Zhou, Tao Wang, Yu Emma Wang, Kellie Webster, Marie Pellat, Kevin Robinson, Kathleen Meier-Hellstern, Toju Duke, Lucas Dixon, Kun Zhang, Quoc V Le, Yonghui Wu, Zhifeng Chen, and Claire Cui. 2022. Glam: Efficient scaling of language models with mixture-of-experts. In Proceedings of the International Conference on Machine Learning. PMLR, 5547\u20135569."},{"key":"e_1_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.5555\/3586589.3586709"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.552"},{"key":"e_1_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403282"},{"key":"e_1_3_2_18_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Gunel Beliz","year":"2020","unstructured":"Beliz Gunel, Jingfei Du, Alexis Conneau, and Ves Stoyanov. 2020. Supervised contrastive learning for pre-trained language model fine-tuning. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475554"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-07221-0_18"},{"key":"e_1_3_2_21_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1991.3.1.79"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1109\/TLT.2017.2689017"},{"key":"e_1_3_2_25_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Kingma Diederik P.","year":"2015","unstructured":"Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_26_2","first-page":"424","volume-title":"Proceedings of the 9th International Conference on Educational Data Mining","author":"Lan Andrew S.","year":"2016","unstructured":"Andrew S. Lan and Richard G. Baraniuk. 2016. A contextual bandits framework for personalized learning action selection. In Proceedings of the 9th International Conference on Educational Data Mining. 424\u2013429."},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.1109\/TLT.2018.2823710"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1145\/3303772.3303786"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3512105"},{"key":"e_1_3_2_30_2","volume-title":"Proceedings of the International Conference on Learning Representations","author":"Lepikhin Dmitry","year":"2020","unstructured":"Dmitry Lepikhin, HyoukJoong Lee, Yuanzhong Xu, Dehao Chen, Orhan Firat, Yanping Huang, Maxim Krikun, Noam Shazeer, and Zhifeng Chen. 2020. Gshard: Scaling giant models with conditional computation and automatic sharding. In Proceedings of the International Conference on Learning Representations."},{"key":"e_1_3_2_31_2","first-page":"6265","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Lewis Mike","year":"2021","unstructured":"Mike Lewis, Shruti Bhosale, Tim Dettmers, Naman Goyal, and Luke Zettlemoyer. 2021. Base layers: Simplifying training of large, sparse models. In Proceedings of the International Conference on Machine Learning. PMLR, 6265\u20136274."},{"key":"e_1_3_2_32_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58577-8_8"},{"key":"e_1_3_2_33_2","doi-asserted-by":"publisher","DOI":"10.1145\/3437963.3441743"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2019.2924374"},{"key":"e_1_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583866"},{"key":"e_1_3_2_36_2","volume-title":"Proceedings of the 11th International Conference on Learning","author":"Liu Zitao","year":"2023","unstructured":"Zitao Liu, Qiongqiong Liu, Jiahao Chen, Shuyan Huang, and Weiqi Luo. 2023a. simpleKT: a simple but tough-to-beat baseline for knowledge tracing. In Proceedings of the 11th International Conference on Learning Representations."},{"key":"e_1_3_2_37_2","first-page":"18542","volume-title":"Proceedings of the 36th Conference on Neural Information Processing Systems Datasets and Benchmarks Track","author":"Liu Zitao","year":"2022","unstructured":"Zitao Liu, Qiongqiong Liu, Jiahao Chen, Shuyan Huang, Jiliang Tang, and Weiqi Luo. 2022. pyKT: A python library to benchmark deep learning based knowledge tracing models. In Proceedings of the 36th Conference on Neural Information Processing Systems Datasets and Benchmarks Track. 18542\u201318555."},{"key":"e_1_3_2_38_2","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462827"},{"issue":"7","key":"e_1_3_2_39_2","first-page":"1","article-title":"Interpreting deep learning models for knowledge tracing","volume":"33","author":"Lu Yu","year":"2022","unstructured":"Yu Lu, Deliang Wang, Penghe Chen, Qinggang Meng, and Shengquan Yu. 2022. Interpreting deep learning models for knowledge tracing. International Journal of Artificial Intelligence in Education 33, 7 (2022), 1\u201324.","journal-title":"International Journal of Artificial Intelligence in Education"},{"key":"e_1_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-52240-7_34"},{"key":"e_1_3_2_41_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2018.00156"},{"key":"e_1_3_2_42_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-39112-5_48"},{"key":"e_1_3_2_43_2","doi-asserted-by":"publisher","DOI":"10.1145\/3308558.3313565"},{"key":"e_1_3_2_44_2","doi-asserted-by":"publisher","DOI":"10.1145\/3350546.3352513"},{"key":"e_1_3_2_45_2","series-title":"International Educational Data Mining Society","first-page":"384","volume-title":"Proceedings of the 12th International Conference on Educational Data Mining","author":"Pandey Shalini","year":"2019","unstructured":"Shalini Pandey and George Karypis. 2019. A self-attentive model for knowledge tracing. In Proceedings of the 12th International Conference on Educational Data Mining. International Educational Data Mining Society, 384\u2013389."},{"key":"e_1_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3411994"},{"key":"e_1_3_2_47_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-22362-4_21"},{"key":"e_1_3_2_48_2","first-page":"505","article-title":"Deep knowledge tracing","volume":"28","author":"Piech Chris","year":"2015","unstructured":"Chris Piech, Jonathan Bassen, Jonathan Huang, Surya Ganguli, Mehran Sahami, Leonidas J. Guibas, and Jascha Sohl-Dickstein. 2015a. Deep knowledge tracing. Advances in Neural Information Processing Systems 28, 505\u2013513.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_49_2","first-page":"1093","volume-title":"Proceedings of the International conference on machine Learning","author":"Piech Chris","year":"2015","unstructured":"Chris Piech, Jonathan Huang, Andy Nguyen, Mike Phulsuksombati, Mehran Sahami, and Leonidas Guibas. 2015b. Learning program embeddings to propagate feedback on student code. In Proceedings of the International conference on machine Learning. PMLR, 1093\u20131102."},{"key":"e_1_3_2_50_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-52240-7_46"},{"key":"e_1_3_2_51_2","volume-title":"EAKT: Embedding Cognitive Framework with Attention for Interpretable Knowledge Tracing","author":"Pu Yanjun","year":"2022","unstructured":"Yanjun Pu, Wenjun Wu, Tianhao Peng, Fang Liu, Yu Liang, Xin Yu, Ruibo Chen, and Pu Feng. 2022. EAKT: Embedding Cognitive Framework with Attention for Interpretable Knowledge Tracing. Scientific Reports."},{"key":"e_1_3_2_52_2","first-page":"8748","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. 2021. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning. PMLR, 8748\u20138763."},{"key":"e_1_3_2_53_2","volume-title":"Probabilistic Models for Some Intelligence and Attainment Tests","author":"Rasch Georg","year":"1993","unstructured":"Georg Rasch. 1993. Probabilistic Models for Some Intelligence and Attainment Tests. ERIC."},{"key":"e_1_3_2_54_2","first-page":"8583","article-title":"Scaling vision with sparse mixture of experts","volume":"34","author":"Riquelme Carlos","year":"2021","unstructured":"Carlos Riquelme, Joan Puigcerver, Basil Mustafa, Maxim Neumann, Rodolphe Jenatton, Andr\u00e9 Susano Pinto, Daniel Keysers, and Neil Houlsby. 2021. Scaling vision with sparse mixture of experts. Advances in Neural Information Processing Systems 34, 8583\u20138595.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_55_2","first-page":"3659","article-title":"Towards crowdsourced training of large neural networks using decentralized mixture-of-experts","volume":"33","author":"Ryabinin Max","year":"2020","unstructured":"Max Ryabinin and Anton Gusev. 2020. Towards crowdsourced training of large neural networks using decentralized mixture-of-experts. Advances in Neural Information Processing Systems 33, 3659\u20133672.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_56_2","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467237"},{"key":"e_1_3_2_57_2","first-page":"677","volume-title":"Proceedings of The 13th International Conference on Educational Data Mining (EDM\u201920)","author":"Sonkar Shashank","year":"2020","unstructured":"Shashank Sonkar, Andrew E. Waters, Andrew S. Lan, Phillip J. Grimaldi, and Richard G. Baraniuk. 2020. qDKT: Question-centric deep knowledge tracing. In Proceedings of The 13th International Conference on Educational Data Mining (EDM\u201920). 677\u2013681."},{"key":"e_1_3_2_58_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11864"},{"key":"e_1_3_2_59_2","doi-asserted-by":"publisher","DOI":"10.4018\/978-1-61350-489-5.ch006"},{"key":"e_1_3_2_60_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM50108.2020.00063"},{"key":"e_1_3_2_61_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.3301750"},{"key":"e_1_3_2_62_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.6080"},{"key":"e_1_3_2_63_2","unstructured":"Zichao Wang Angus Lamb Evgeny Saveliev Pashmina Cameron Yordan Zaykov Jos\u00e9 Miguel Hern\u00e1ndez-Lobato Richard E. Turner Richard G. Baraniuk Craig Barton Simon Peyton Jones Simon Woodhead and Cheng Zhang. 2020a. Instructions and guide for diagnostic questions: The NeurIPS 2020 education challenge. arXiv:2007.12061. Retrieved from https:\/\/arxiv.org\/abs\/2007.12061"},{"key":"e_1_3_2_64_2","volume-title":"Building Intelligent Interactive Tutors: Student-Centered Strategies for Revolutionizing","author":"Beverly Park Woolf","year":"2010","unstructured":"Beverly Park Woolf. 2010. Building Intelligent Interactive Tutors: Student-Centered Strategies for Revolutionizing e-Learning. Morgan Kaufmann."},{"key":"e_1_3_2_65_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2020.106481"},{"key":"e_1_3_2_66_2","first-page":"299","volume-title":"Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases","author":"Yang Yang","year":"2020","unstructured":"Yang Yang, Jian Shen, Yanru Qu, Yunfei Liu, Kerong Wang, Yaoming Zhu, Weinan Zhang, and Yong Yu. 2020. GIKT: A graph-based interaction model for knowledge tracing. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 299\u2013315."},{"key":"e_1_3_2_67_2","first-page":"683","volume-title":"Proceedings of the 12th International Conference on Educational Data Mining (EDM\u201919)","author":"Yeung Chun-Kit","year":"2019","unstructured":"Chun-Kit Yeung. 2019a. Deep-IRT: Make deep learning based knowledge tracing explainable using item response theory. In Proceedings of the 12th International Conference on Educational Data Mining (EDM\u201919). 683\u2013686."},{"key":"e_1_3_2_68_2","unstructured":"Chun-Kit Yeung. 2019b. Deep-IRT: Make deep learning based knowledge tracing explainable using item response theory. arXiv:1904.11738. Retrieved from https:\/\/arxiv.org\/abs\/1904.11738"},{"key":"e_1_3_2_69_2","doi-asserted-by":"publisher","DOI":"10.1145\/3231644.3231647"},{"key":"e_1_3_2_70_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-39112-5_18"},{"key":"e_1_3_2_71_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.bea-1.23"},{"key":"e_1_3_2_72_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.aacl-main.73"},{"key":"e_1_3_2_73_2","doi-asserted-by":"publisher","DOI":"10.1145\/3038912.3052580"},{"key":"e_1_3_2_74_2","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3482372"},{"key":"e_1_3_2_75_2","doi-asserted-by":"publisher","DOI":"10.1145\/3386527.3406739"}],"container-title":["ACM Transactions on Knowledge Discovery from Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3674840","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3674840","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:05:56Z","timestamp":1750291556000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3674840"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,15]]},"references-count":74,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,2,28]]}},"alternative-id":["10.1145\/3674840"],"URL":"https:\/\/doi.org\/10.1145\/3674840","relation":{},"ISSN":["1556-4681","1556-472X"],"issn-type":[{"value":"1556-4681","type":"print"},{"value":"1556-472X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,15]]},"assertion":[{"value":"2023-09-15","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-06-09","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-02-15","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}