{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T02:45:18Z","timestamp":1767321918118,"version":"3.48.0"},"publisher-location":"Singapore","reference-count":38,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819541577","type":"print"},{"value":"9789819541584","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-4158-4_5","type":"book-chapter","created":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T02:42:39Z","timestamp":1767321759000},"page":"71-87","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Harnessing LLMs Explanations to\u00a0Boost Surrogate Models in\u00a0Tabular Data Classification"],"prefix":"10.1007","author":[{"given":"Ruxue","family":"Shi","sequence":"first","affiliation":[]},{"given":"Hengrui","family":"Gu","sequence":"additional","affiliation":[]},{"given":"Xu","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,2]]},"reference":[{"key":"5_CR1","doi-asserted-by":"crossref","unstructured":"Arik, S.\u00d6., Pfister, T.: Tabnet: Attentive interpretable tabular learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a035, pp. 6679\u20136687 (2021)","DOI":"10.1609\/aaai.v35i8.16826"},{"key":"5_CR2","unstructured":"Asuncion, A., Newman, D.: Uci machine learning repository (2007)"},{"key":"5_CR3","unstructured":"Badirli, S., Liu, X., Xing, Z., Bhowmik, A., Doan, K., Keerthi, S.S.: Gradient boosting neural networks: Grownet. arXiv preprint arXiv:2002.07971 (2020)"},{"key":"5_CR4","unstructured":"Bahri, D., Jiang, H., Tay, Y., Metzler, D.: Scarf: self-supervised contrastive learning using random feature corruption. arXiv preprint arXiv:2106.15147 (2021)"},{"key":"5_CR5","unstructured":"Bahri, D., Jiang, H., Tay, Y., Metzler, D.: Scarf: self-supervised contrastive learning using random feature corruption. In: International Conference on Learning Representations (2022)"},{"key":"5_CR6","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1023\/A:1010933404324","volume":"45","author":"L Breiman","year":"2001","unstructured":"Breiman, L.: Random forests. Mach. Learn. 45, 5\u201332 (2001)","journal-title":"Mach. Learn."},{"key":"5_CR7","unstructured":"Chen, S., Wu, J., Hovakimyan, N., Yao, H.: Recontab: regularized contrastive representation learning for tabular data. arXiv preprint arXiv:2310.18541 (2023)"},{"key":"5_CR8","doi-asserted-by":"crossref","unstructured":"Chen, T., Guestrin, C.: Xgboost: A scalable tree boosting system. In: Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining. pp. 785\u2013794 (2016)","DOI":"10.1145\/2939672.2939785"},{"key":"5_CR9","first-page":"11763","volume":"35","author":"T Dinh","year":"2022","unstructured":"Dinh, T., et al.: Lift: language-interfaced fine-tuning for non-language machine learning tasks. Adv. Neural. Inf. Process. Syst. 35, 11763\u201311784 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"5_CR10","doi-asserted-by":"crossref","unstructured":"Fan, Z., Gao, X., Mirchev, M., Roychoudhury, A., Tan, S.H.: Automated repair of programs from large language models. In: 2023 IEEE\/ACM 45th International Conference on Software Engineering (ICSE), pp. 1469\u20131481. IEEE (2023)","DOI":"10.1109\/ICSE48619.2023.00128"},{"key":"5_CR11","doi-asserted-by":"crossref","unstructured":"Finch, S.E., Paek, E.S., Choi, J.D.: Leveraging large language models for automated dialogue analysis. In: Proceedings of the 24th Meeting of the Special Interest Group on Discourse and Dialogue, pp. 202\u2013215 (2023)","DOI":"10.18653\/v1\/2023.sigdial-1.20"},{"key":"5_CR12","first-page":"18932","volume":"34","author":"Y Gorishniy","year":"2021","unstructured":"Gorishniy, Y., Rubachev, I., Khrulkov, V., Babenko, A.: Revisiting deep learning models for tabular data. Adv. Neural. Inf. Process. Syst. 34, 18932\u201318943 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"5_CR13","doi-asserted-by":"crossref","unstructured":"Hastie, T.J., Pregibon, D.: Generalized linear models. In: Statistical models in S, pp. 195\u2013247. Routledge (2017)","DOI":"10.1201\/9780203738535-6"},{"key":"5_CR14","doi-asserted-by":"crossref","unstructured":"Ho, T.K.: Random decision forests. In: Proceedings of 3rd International Conference on Document Analysis and Recognition, vol.\u00a01, pp. 278\u2013282. IEEE (1995)","DOI":"10.1109\/ICDAR.1995.598994"},{"key":"5_CR15","unstructured":"Hollmann, N., M\u00fcller, S., Eggensperger, K., Hutter, F.: Tabpfn: a transformer that solves small tabular classification problems in a second. In: The Eleventh International Conference on Learning Representations (2023)"},{"key":"5_CR16","unstructured":"Huang, X., Khetan, A., Cvitkovic, M., Karnin, Z.: Tabtransformer: tabular data modeling using contextual embeddings. arXiv preprint arXiv:2012.06678 (2020)"},{"key":"5_CR17","first-page":"23928","volume":"34","author":"A Kadra","year":"2021","unstructured":"Kadra, A., Lindauer, M., Hutter, F., Grabocka, J.: Well-tuned simple nets excel on tabular datasets. Adv. Neural. Inf. Process. Syst. 34, 23928\u201323941 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"5_CR18","unstructured":"Ke, G., et al.: Lightgbm: a highly efficient gradient boosting decision tree. Advances in neural information processing systems 30 (2017)"},{"key":"5_CR19","doi-asserted-by":"crossref","unstructured":"Ke, G., Xu, Z., Zhang, J., Bian, J., Liu, T.Y.: Deepgbm: a deep learning framework distilled by gbdt for online prediction tasks. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 384\u2013394 (2019)","DOI":"10.1145\/3292500.3330858"},{"key":"5_CR20","unstructured":"Ke, G., Zhang, J., Xu, Z., Bian, J., Liu, T.Y.: Tabnn: A universal neural network solution for tabular data (2018)"},{"key":"5_CR21","unstructured":"Klambauer, G., Unterthiner, T., Mayr, A., Hochreiter, S.: Self-normalizing neural networks. Advances in neural information processing systems 30 (2017)"},{"issue":"18","key":"5_CR22","doi-asserted-by":"publisher","first-page":"2395","DOI":"10.1161\/CIRCULATIONAHA.106.682658","volume":"117","author":"MP LaValley","year":"2008","unstructured":"LaValley, M.P.: Logistic regression. Circulation 117(18), 2395\u20132399 (2008)","journal-title":"Circulation"},{"issue":"9","key":"5_CR23","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3560815","volume":"55","author":"P Liu","year":"2023","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., Neubig, G.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9), 1\u201335 (2023)","journal-title":"ACM Comput. Surv."},{"issue":"1","key":"5_CR24","first-page":"14","volume":"1","author":"WY Loh","year":"2011","unstructured":"Loh, W.Y.: Classification and regression trees. Wiley interdisciplinary reviews: data mining and knowledge discovery 1(1), 14\u201323 (2011)","journal-title":"Wiley interdisciplinary reviews: data mining and knowledge discovery"},{"key":"5_CR25","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1016\/j.dss.2014.03.001","volume":"62","author":"S Moro","year":"2014","unstructured":"Moro, S., Cortez, P., Rita, P.: A data-driven approach to predict the success of bank telemarketing. Decision Support Syst. 62, 22\u201331 (2014)","journal-title":"Decision Support Syst."},{"key":"5_CR26","unstructured":"Nam, J., et al.: Semi-supervised tabular classification via in-context learning of large language models. In: Workshop on Efficient Systems for Foundation Models@ ICML2023 (2023)"},{"key":"5_CR27","unstructured":"Nam, J., Tack, J., Lee, K., Lee, H., Shin, J.: Stunt: few-shot tabular learning with self-generated tasks from unlabeled tables. In: The Eleventh International Conference on Learning Representations (2023)"},{"key":"5_CR28","unstructured":"Popov, S., Morozov, S., Babenko, A.: Neural oblivious decision ensembles for deep learning on tabular data. arXiv preprint arXiv:1909.06312 (2019)"},{"key":"5_CR29","unstructured":"Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A.V., Gulin, A.: Catboost: unbiased boosting with categorical features. Advances in neural information processing systems 31 (2018)"},{"key":"5_CR30","unstructured":"Slack, D., Singh, S.: Tablet: learning from instructions for tabular data. arXiv preprint arXiv:2304.13188 (2023)"},{"key":"5_CR31","unstructured":"Somepalli, G., Goldblum, M., Schwarzschild, A., Bruss, C.B., Goldstein, T.: Saint: Improved neural networks for tabular data via row attention and contrastive pre-training. arXiv preprint arXiv:2106.01342 (2021)"},{"key":"5_CR32","doi-asserted-by":"crossref","unstructured":"Song, W., et al.: Autoint: automatic feature interaction learning via self-attentive neural networks. In: Proceedings of the 28th ACM international conference on information and knowledge management. pp. 1161\u20131170 (2019)","DOI":"10.1145\/3357384.3357925"},{"key":"5_CR33","unstructured":"Touvron, H., et\u00a0al.: Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)"},{"key":"5_CR34","doi-asserted-by":"crossref","unstructured":"Wang, R., et al.: Dcn v2: Improved deep & cross network and practical lessons for web-scale learning to rank systems. In: Proceedings of the Web Conference 2021, pp. 1785\u20131797 (2021)","DOI":"10.1145\/3442381.3450078"},{"key":"5_CR35","unstructured":"Wei, J., et\u00a0al.: Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022)"},{"key":"5_CR36","first-page":"24824","volume":"35","author":"J Wei","year":"2022","unstructured":"Wei, J., et al.: Chain-of-thought prompting elicits reasoning in large language models. Adv. Neural. Inf. Process. Syst. 35, 24824\u201324837 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"5_CR37","first-page":"11033","volume":"33","author":"J Yoon","year":"2020","unstructured":"Yoon, J., Zhang, Y., Jordon, J., van der Schaar, M.: Vime: Extending the success of self-and semi-supervised learning to tabular domain. Adv. Neural. Inf. Process. Syst. 33, 11033\u201311043 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"5_CR38","unstructured":"Zhang, Z., Zhang, A., Li, M., Smola, A.: Automatic chain of thought prompting in large language models. In: The Eleventh International Conference on Learning Representations"}],"container-title":["Lecture Notes in Computer Science","Database Systems for Advanced Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-4158-4_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T02:42:44Z","timestamp":1767321764000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-4158-4_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819541577","9789819541584"],"references-count":38,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-4158-4_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"2 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DASFAA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Database Systems for Advanced Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 May 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 May 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"dasfaa2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/dasfaa2025.github.io","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}