{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,3]],"date-time":"2025-10-03T12:43:37Z","timestamp":1759495417565,"version":"build-2065373602"},"publisher-location":"Berlin, Heidelberg","reference-count":58,"publisher":"Springer Berlin Heidelberg","isbn-type":[{"value":"9783662722428","type":"print"},{"value":"9783662722435","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,4]],"date-time":"2025-10-04T00:00:00Z","timestamp":1759536000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,4]],"date-time":"2025-10-04T00:00:00Z","timestamp":1759536000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-662-72243-5_18","type":"book-chapter","created":{"date-parts":[[2025,10,3]],"date-time":"2025-10-03T12:14:07Z","timestamp":1759493647000},"page":"310-327","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Longitudinal Surveys Are Texts: LLM-Enhanced Analysis of\u00a0School Attendance in\u00a0New Zealand"],"prefix":"10.1007","author":[{"given":"Tingrui","family":"Qiao","sequence":"first","affiliation":[]},{"given":"Caroline","family":"Walker","sequence":"additional","affiliation":[]},{"given":"Chris","family":"Cunningham","sequence":"additional","affiliation":[]},{"given":"Adam","family":"Jang-Jones","sequence":"additional","affiliation":[]},{"given":"Susan","family":"Morton","sequence":"additional","affiliation":[]},{"given":"Kane","family":"Meissel","sequence":"additional","affiliation":[]},{"given":"Yun","family":"Sing Koh","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,4]]},"reference":[{"issue":"4","key":"18_CR1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0266516","volume":"17","author":"DA Adler","year":"2022","unstructured":"Adler, D.A., Wang, F., Mohr, D.C., Choudhury, T.: Machine learning for passive mental health symptom prediction: generalization across different longitudinal mobile sensing studies. PLoS ONE 17(4), e0266516 (2022)","journal-title":"PLoS ONE"},{"key":"18_CR2","unstructured":"Atkinson, J., Salmond, C., Crampton, P.: Nzdep2013 index of deprivation. Wellington: Department of Public Health, University of Otago 5541, 1\u201364 (2014)"},{"key":"18_CR3","unstructured":"Bai, J., et al.: Qwen technical report. arXiv preprint arXiv:2309.16609 (2023)"},{"key":"18_CR4","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1023\/A:1010933404324","volume":"45","author":"L Breiman","year":"2001","unstructured":"Breiman, L.: Random forests. Mach. Learn. 45, 5\u201332 (2001)","journal-title":"Mach. Learn."},{"key":"18_CR5","unstructured":"Cao, D., et al.: Tempo: prompt-based generative pre-trained transformer for time series forecasting. arXiv preprint arXiv:2310.04948 (2023)"},{"key":"18_CR6","doi-asserted-by":"crossref","unstructured":"Chen, T., Guestrin, C.: Xgboost: a scalable tree boosting system. In: SIGKDD, pp. 785\u2013794 (2016)","DOI":"10.1145\/2939672.2939785"},{"key":"18_CR7","doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Burstein, J., Doran, C., Solorio, T. (eds.) Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, vol. 1 (Long and Short Papers), pp. 4171\u20134186. Association for Computational Linguistics, Minneapolis (2019). https:\/\/doi.org\/10.18653\/v1\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Dey, R., Salem, F.M.: Gate-variants of gated recurrent unit (gru) neural networks. In: MWSCAS, pp. 1597\u20131600. IEEE (2017)","DOI":"10.1109\/MWSCAS.2017.8053243"},{"issue":"7","key":"18_CR9","doi-asserted-by":"publisher","first-page":"591","DOI":"10.1037\/0003-066X.63.7.591","volume":"63","author":"DM Erceg-Hurn","year":"2008","unstructured":"Erceg-Hurn, D.M., Mirosevich, V.M.: Modern robust statistical methods: an easy way to maximize the accuracy and power of your research. Am. Psychol. 63(7), 591 (2008)","journal-title":"Am. Psychol."},{"key":"18_CR10","doi-asserted-by":"crossref","unstructured":"Finning, K., et al.: The association between anxiety and poor attendance at school\u2013a systematic review. Child Adolesc. Mental Health 24(3), 205\u2013216 (2019)","DOI":"10.1111\/camh.12322"},{"key":"18_CR11","doi-asserted-by":"crossref","unstructured":"Harder, F., Bauer, M., Park, M.: Interpretable and differentially private predictions. In: Proceedings of AAAI, vol.\u00a034, pp. 4083\u20134090 (2020)","DOI":"10.1609\/aaai.v34i04.5827"},{"key":"18_CR12","unstructured":"Heyen, H., Widdicombe, A., Siegel, N.Y., Perez-Ortiz, M., Treleaven, P.: The effect of model size on llm post-hoc explainability via lime. arXiv preprint arXiv:2405.05348 (2024)"},{"issue":"8","key":"18_CR13","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735\u20131780 (1997)","journal-title":"Neural Comput."},{"key":"18_CR14","unstructured":"Houlsby, N., et al.: Parameter-efficient transfer learning for nlp. In: ICML, pp. 2790\u20132799. PMLR (2019)"},{"issue":"2","key":"18_CR15","first-page":"3","volume":"1","author":"EJ Hu","year":"2022","unstructured":"Hu, E.J., et al.: Lora: low-rank adaptation of large language models. ICLR 1(2), 3 (2022)","journal-title":"ICLR"},{"key":"18_CR16","unstructured":"Huang, S., Mamidanna, S., Jangam, S., Zhou, Y., Gilpin, L.H.: Can large language models explain themselves? A study of llm-generated self-explanations. arXiv preprint arXiv:2310.11207 (2023)"},{"key":"18_CR17","unstructured":"Jiang, A.Q., et al.: Mistral 7b (2023). https:\/\/arxiv.org\/abs\/2310.06825"},{"issue":"7","key":"18_CR18","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0271487","volume":"17","author":"BT Jin","year":"2022","unstructured":"Jin, B.T., Choi, M.H., Moyer, M.F., Kim, D.A.: Predicting malnutrition from longitudinal patient trajectories with deep learning. PLoS ONE 17(7), e0271487 (2022)","journal-title":"PLoS ONE"},{"key":"18_CR19","unstructured":"Jin, M., et\u00a0al.: Time-llm: time series forecasting by reprogramming large language models. arXiv preprint arXiv:2310.01728 (2023)"},{"key":"18_CR20","doi-asserted-by":"crossref","unstructured":"Kearney, C.A., Benoit, L., Gonz\u00e1lvez, C., Keppens, G.: School attendance and school absenteeism: a primer for the past, present, and theory of change for the future. In: Frontiers in Education, vol.\u00a07, p. 1044608. Frontiers (2022)","DOI":"10.3389\/feduc.2022.1044608"},{"key":"18_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.childyouth.2020.105432","volume":"118","author":"M Klein","year":"2020","unstructured":"Klein, M., Sosu, E.M., Dare, S.: Mapping inequalities in school attendance: the relationship between dimensions of socioeconomic status and forms of school absence. Child Youth Serv. Rev. 118, 105432 (2020)","journal-title":"Child Youth Serv. Rev."},{"key":"18_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.avb.2022.101722","volume":"64","author":"R Laith","year":"2022","unstructured":"Laith, R., Vaillancourt, T.: The temporal sequence of bullying victimization, academic achievement, and school attendance: A review of the literature. Aggress. Violent. Beh. 64, 101722 (2022)","journal-title":"Aggress. Violent. Beh."},{"key":"18_CR23","doi-asserted-by":"crossref","unstructured":"Lester, B., Al-Rfou, R., Constant, N.: The power of scale for parameter-efficient prompt tuning. arXiv preprint arXiv:2104.08691 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"Li, X.L., Liang, P.: Prefix-tuning: optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190 (2021)","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"18_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s12859-014-0346-6","volume":"15","author":"SG Liao","year":"2014","unstructured":"Liao, S.G., et al.: Missing value imputation in high-dimensional phenomic data: imputable or not, and how? BMC Bioinformatics 15, 1\u201312 (2014)","journal-title":"BMC Bioinformatics"},{"key":"18_CR26","unstructured":"Liu, Y., et al.: itransformer: inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625 (2023)"},{"key":"18_CR27","first-page":"9881","volume":"35","author":"Y Liu","year":"2022","unstructured":"Liu, Y., Wu, H., Wang, J., Long, M.: Non-stationary transformers: exploring the stationarity in time series forecasting. NeuraIPS 35, 9881\u20139893 (2022)","journal-title":"NeuraIPS"},{"key":"18_CR28","doi-asserted-by":"crossref","unstructured":"Madhu, G., Bharadwaj, B.L., Nagachandrika, G., Vardhan, K.S.: A novel algorithm for missing data imputation on machine learning. In: 2019 International Conference on Smart Systems and Inventive Technology (ICSSIT), pp. 173\u2013177. IEEE (2019)","DOI":"10.1109\/ICSSIT46314.2019.8987895"},{"key":"18_CR29","doi-asserted-by":"crossref","unstructured":"Maltezou, H.C., Ledda, C., Sipsas, N.V.: Absenteeism of healthcare personnel in the covid-19 era: a systematic review of the literature and implications for the post-pandemic seasons. In: Healthcare, vol.\u00a011, p.\u00a02950. MDPI (2023)","DOI":"10.3390\/healthcare11222950"},{"key":"18_CR30","unstructured":"Mattei, P.A., Frellsen, J.: Miwae: deep generative modelling and imputation of incomplete data sets. In: ICML, pp. 4413\u20134423. PMLR (2019)"},{"issue":"4","key":"18_CR31","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1080\/1045988X.2013.821649","volume":"58","author":"BM McConnell","year":"2014","unstructured":"McConnell, B.M., Kubina, R.M.: Connecting with families to improve students\u2019 school attendance: a review of the literature. Prevent. School Fail. Alternat. Educ. Children Youth 58(4), 249\u2013256 (2014)","journal-title":"Prevent. School Fail. Alternat. Educ. Children Youth"},{"issue":"64\u201367","key":"18_CR32","first-page":"2","volume":"5","author":"LR Medsker","year":"2001","unstructured":"Medsker, L.R., Jain, L., et al.: Recurrent neural networks. Des. Appl. 5(64\u201367), 2 (2001)","journal-title":"Des. Appl."},{"key":"18_CR33","unstructured":"Mohammadi, B.: Explaining large language models decisions using shapley values. arXiv preprint arXiv:2404.01332 (2024)"},{"key":"18_CR34","doi-asserted-by":"crossref","unstructured":"Morton, S.M., et\u00a0al.: Cohort profile: growing up in New Zealand. Int. J. Epidemiol. 42(1), 65\u201375 (2013)","DOI":"10.1093\/ije\/dyr206"},{"key":"18_CR35","unstructured":"Nie, Y., Nguyen, N.H., Sinthong, P., Kalagnanam, J.: A time series is worth 64 words: long-term forecasting with transformers. arXiv preprint arXiv:2211.14730 (2022)"},{"issue":"5","key":"18_CR36","doi-asserted-by":"publisher","first-page":"e295","DOI":"10.1016\/S2589-7500(21)00040-6","volume":"3","author":"O Nitski","year":"2021","unstructured":"Nitski, O., et al.: Long-term mortality risk stratification of liver transplant recipients: real-time application of deep learning algorithms on longitudinal data. Lancet Digital Health 3(5), e295\u2013e305 (2021)","journal-title":"Lancet Digital Health"},{"key":"18_CR37","unstructured":"Pan, Z., Jiang, Y., Garg, S., Schneider, A., Nevmyvaka, Y., Song, D.: $$s^2$$ip-llm: semantic space informed prompt learning with llm for time series forecasting. In: ICML (2024)"},{"key":"18_CR38","doi-asserted-by":"publisher","unstructured":"Pang, B., Qiao, T., Walker, C., Cunningham, C., Koh, Y.S.: Libra: measuring bias of large language model from a local context. In: European Conference on Information Retrieval, pp. 1\u201316. Springer, Heidelberg (2025). https:\/\/doi.org\/10.1007\/978-3-031-88708-6_1","DOI":"10.1007\/978-3-031-88708-6_1"},{"key":"18_CR39","doi-asserted-by":"crossref","unstructured":"Qiao, T., Walker, C., Cunningham, C., Koh, Y.S.: Thematic-lm: a llm-based multi-agent system for large-scale thematic analysis. In: Proceedings of the ACM on Web Conference 2025, pp. 649\u2013658 (2025)","DOI":"10.1145\/3696410.3714595"},{"issue":"8","key":"18_CR40","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)","journal-title":"OpenAI Blog"},{"key":"18_CR41","doi-asserted-by":"crossref","unstructured":"Sheetal, A., Jiang, Z., Di\u00a0Milia, L.: Using machine learning to analyze longitudinal data: a tutorial guide and best-practice recommendations for social science researchers. Appl. Psychol. (2023)","DOI":"10.1111\/apps.12435"},{"key":"18_CR42","doi-asserted-by":"publisher","DOI":"10.1016\/j.physd.2019.132306","volume":"404","author":"A Sherstinsky","year":"2020","unstructured":"Sherstinsky, A.: Fundamentals of recurrent neural network (rnn) and long short-term memory (lstm) network. Physica D 404, 132306 (2020)","journal-title":"Physica D"},{"key":"18_CR43","unstructured":"Srinivas, S., Fleuret, F.: Rethinking the role of gradient-based attribution methods for model interpretability. arXiv preprint arXiv:2006.09128 (2020)"},{"key":"18_CR44","unstructured":"Sun, C., Li, H., Li, Y., Hong, S.: Test: text prototype aligned embedding to activate llm\u2019s ability for time series. arXiv preprint arXiv:2308.08241 (2023)"},{"key":"18_CR45","doi-asserted-by":"crossref","unstructured":"Tan, Z., Chen, T., Zhang, Z., Liu, H.: Sparsity-guided holistic explanation for llms with interpretable inference-time intervention. In: Proceedings of AAAI, vol.\u00a038, pp. 21619\u201321627 (2024)","DOI":"10.1609\/aaai.v38i19.30160"},{"key":"18_CR46","unstructured":"Touvron, H., et\u00a0al.: Llama: open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)"},{"key":"18_CR47","unstructured":"Vaswani, A., et al.: Attention is all you need. NeuraIPS 30 (2017)"},{"issue":"8","key":"18_CR48","doi-asserted-by":"publisher","first-page":"876","DOI":"10.3390\/electronics8080876","volume":"8","author":"R Wan","year":"2019","unstructured":"Wan, R., Mei, S., Wang, J., Liu, M., Yang, F.: Multivariate temporal convolutional network: a deep neural networks approach for multivariate time series forecasting. Electronics 8(8), 876 (2019)","journal-title":"Electronics"},{"key":"18_CR49","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1007\/s11165-004-3437-y","volume":"35","author":"RT White","year":"2005","unstructured":"White, R.T., Arzi, H.J.: Longitudinal studies: designs, validity, practicality, and value. Res. Sci. Educ. 35, 137\u2013149 (2005)","journal-title":"Res. Sci. Educ."},{"key":"18_CR50","first-page":"22419","volume":"34","author":"H Wu","year":"2021","unstructured":"Wu, H., Xu, J., Wang, J., Long, M.: Autoformer: decomposition transformers with auto-correlation for long-term series forecasting. NeuraIPS 34, 22419\u201322430 (2021)","journal-title":"NeuraIPS"},{"issue":"11","key":"18_CR51","first-page":"6851","volume":"36","author":"H Xue","year":"2023","unstructured":"Xue, H., Salim, F.D.: Promptcast: a new prompt-based learning paradigm for time series forecasting. IEEE TKDE 36(11), 6851\u20136864 (2023)","journal-title":"IEEE TKDE"},{"key":"18_CR52","unstructured":"Yoon, J., Jordon, J., Schaar, M.: Gain: missing data imputation using generative adversarial nets. In: ICML, pp. 5689\u20135698. PMLR (2018)"},{"key":"18_CR53","doi-asserted-by":"crossref","unstructured":"Zeng, A., Chen, M., Zhang, L., Xu, Q.: Are transformers effective for time series forecasting? In: Proceedings of AAAI, vol.\u00a037, pp. 11121\u201311128 (2023)","DOI":"10.1609\/aaai.v37i9.26317"},{"key":"18_CR54","unstructured":"Zhang, J., Bareinboim, E.: Equality of opportunity in classification: a causal approach. NeuraIPS 31 (2018)"},{"key":"18_CR55","unstructured":"Zhang, Y., Yan, J.: Crossformer: transformer utilizing cross-dimension dependency for multivariate time series forecasting. In: ICLR (2023)"},{"key":"18_CR56","doi-asserted-by":"crossref","unstructured":"Zhou, H., et al.: Informer: beyond efficient transformer for long sequence time-series forecasting. In: Proceedings of AAAI. vol.\u00a035, pp. 11106\u201311115 (2021)","DOI":"10.1609\/aaai.v35i12.17325"},{"key":"18_CR57","unstructured":"Zhou, T., Ma, Z., Wen, Q., Wang, X., Sun, L., Jin, R.: Fedformer: frequency enhanced decomposed transformer for long-term series forecasting. In: Proceedings of ICML, pp. 27268\u201327286. PMLR (2022)"},{"key":"18_CR58","first-page":"43322","volume":"36","author":"T Zhou","year":"2023","unstructured":"Zhou, T., Niu, P., Sun, L., Jin, R., et al.: One fits all: power general time series analysis by pretrained lm. NeuraIPS 36, 43322\u201343355 (2023)","journal-title":"NeuraIPS"}],"container-title":["Lecture Notes in Computer Science","Machine Learning and Knowledge Discovery in Databases. Research Track and Applied Data Science Track"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-662-72243-5_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,3]],"date-time":"2025-10-03T12:14:29Z","timestamp":1759493669000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-662-72243-5_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,4]]},"ISBN":["9783662722428","9783662722435"],"references-count":58,"URL":"https:\/\/doi.org\/10.1007\/978-3-662-72243-5_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,4]]},"assertion":[{"value":"4 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECML PKDD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Joint European Conference on Machine Learning and Knowledge Discovery in Databases","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Porto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Portugal","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecml2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ecmlpkdd.org\/2025\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}