{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T23:05:12Z","timestamp":1775257512134,"version":"3.50.1"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2024,8,5]],"date-time":"2024-08-05T00:00:00Z","timestamp":1722816000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,8,5]],"date-time":"2024-08-05T00:00:00Z","timestamp":1722816000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["DEB-1933497"],"award-info":[{"award-number":["DEB-1933497"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Mach Learn"],"published-print":{"date-parts":[[2024,10]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Model distillation has been a popular method for producing interpretable machine learning. It uses an interpretable \u201cstudent\u201d model to mimic the predictions made by the black box \u201cteacher\u201d model. However, when the student model is sensitive to the variability of the data sets used for training even when keeping the teacher fixed, the corresponded interpretation is not reliable. Existing strategies stabilize model distillation by checking whether a large enough sample of pseudo-data is generated to reliably reproduce student models, but methods to do so have so far been developed separately for each specific class of student model. In this paper, we develop a generic approach for stable model distillation based on central limit theorem for the estimated fidelity of the student to the teacher. We start with a collection of candidate student models and search for candidates that reasonably agree with the teacher. Then we construct a multiple testing framework to select a sample size such that the consistent student model would be selected under different pseudo samples. We demonstrate the application of our proposed approach on three commonly used intelligible models: decision trees, falling rule lists and symbolic regression. Finally, we conduct simulation experiments on Mammographic Mass and Breast Cancer datasets and illustrate the testing procedure throughout a theoretical analysis with Markov process. The code is publicly available at <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" ext-link-type=\"uri\" xlink:href=\"https:\/\/github.com\/yunzhe-zhou\/GenericDistillation\">https:\/\/github.com\/yunzhe-zhou\/GenericDistillation<\/jats:ext-link>.<\/jats:p>","DOI":"10.1007\/s10994-024-06597-w","type":"journal-article","created":{"date-parts":[[2024,8,5]],"date-time":"2024-08-05T22:07:12Z","timestamp":1722895632000},"page":"7645-7688","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["A generic approach for reproducible model distillation"],"prefix":"10.1007","volume":"113","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6869-2942","authenticated-orcid":false,"given":"Yunzhe","family":"Zhou","sequence":"first","affiliation":[]},{"given":"Peiru","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Giles","family":"Hooker","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,5]]},"reference":[{"key":"6597_CR1","doi-asserted-by":"publisher","first-page":"175","DOI":"10.1007\/978-1-4939-0375-7_10","volume-title":"Genetic programming theory and practice xi","author":"M Affenzeller","year":"2014","unstructured":"Affenzeller, M., Winkler, S. M., Kronberger, G., Kommenda, M., Burlacu, B., & Wagner, S. (2014). Gaining deeper insights in symbolic regression. Genetic programming theory and practice xi (pp. 175\u2013190). Berlin: Springer."},{"key":"6597_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10710-022-09435-x","volume":"23","author":"GSI Aldeia","year":"2022","unstructured":"Aldeia, G. S. I., & de Fran\u00e7a, F. O. (2022). Interpretability in symbolic regression: A benchmark of explanatory methods using the feynman data set. Genetic Programming and Evolvable Machines, 23, 1\u201341.","journal-title":"Genetic Programming and Evolvable Machines"},{"key":"6597_CR3","doi-asserted-by":"crossref","unstructured":"Augusto, D. A. & Barbosa, H. J. (2000). Symbolic regression via genetic programming. In Proceedings. vol. 1. Sixth Brazilian Symposium on Neural Networks (pp. 173\u2013178).","DOI":"10.1109\/SBRN.2000.889734"},{"key":"6597_CR4","unstructured":"Bishop, C. M. (1994). Mixture density networks."},{"issue":"24","key":"6597_CR5","doi-asserted-by":"publisher","first-page":"9943","DOI":"10.1073\/pnas.0609476104","volume":"104","author":"J Bongard","year":"2007","unstructured":"Bongard, J., & Lipson, H. (2007). Automated reverse engineering of nonlinear dynamical systems. Proceedings of the National Academy of Sciences, 104(24), 9943\u20139948.","journal-title":"Proceedings of the National Academy of Sciences"},{"issue":"1","key":"6597_CR6","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1023\/A:1010933404324","volume":"45","author":"L Breiman","year":"2001","unstructured":"Breiman, L. (2001). Random forests. Machine Learning, 45(1), 5\u201332.","journal-title":"Machine Learning"},{"issue":"58","key":"6597_CR7","first-page":"3","volume":"1","author":"L Breiman","year":"2002","unstructured":"Breiman, L. (2002). Manual on setting up, using, and understanding random forests v3.1. Statistics Department University of California Berkeley, CA, USA, 1(58), 3\u201342.","journal-title":"Statistics Department University of California Berkeley, CA, USA"},{"key":"6597_CR8","volume-title":"Classification and regression trees. brooks","author":"L Breiman","year":"1984","unstructured":"Breiman, L., Friedman, J. H., Olshen, R. A., & Stone, C. J. (1984). Classification and regression trees. brooks. Monterey, CA: Wadsworth and Brooks."},{"key":"6597_CR9","doi-asserted-by":"publisher","DOI":"10.1201\/9781315139470","volume-title":"Classification and regression trees","author":"L Breiman","year":"2017","unstructured":"Breiman, L., Friedman, J. H., Olshen, R. A., & Stone, C. J. (2017). Classification and regression trees. Oxfordshire: Routledge."},{"issue":"1","key":"6597_CR10","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1109\/MSP.2017.2765202","volume":"35","author":"A Creswell","year":"2018","unstructured":"Creswell, A., White, T., Dumoulin, V., Arulkumaran, K., Sengupta, B., & Bharath, A. A. (2018). Generative adversarial networks: An overview. IEEE Signal Processing Magazine, 35(1), 53\u201365.","journal-title":"IEEE Signal Processing Magazine"},{"key":"6597_CR11","unstructured":"Donsker, M. D. (1951). An invariance principle for certain probability limit theorems."},{"key":"6597_CR12","unstructured":"Doob, J. L. (1949). Application of the theory of martingales. Le Calcul des Probabilites et ses Applications (pp. 23\u201327)."},{"key":"6597_CR13","unstructured":"Doshi-Velez, F. & Kim, B. (2017). Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608"},{"issue":"1","key":"6597_CR14","doi-asserted-by":"publisher","first-page":"68","DOI":"10.1145\/3359786","volume":"63","author":"M Du","year":"2019","unstructured":"Du, M., Liu, N., & Hu, X. (2019). Techniques for interpretable machine learning. Communications of the ACM, 63(1), 68\u201377.","journal-title":"Communications of the ACM"},{"key":"6597_CR15","unstructured":"Duembgen, L. (2010). Bounding standard gaussian tail probabilities. arXiv preprint arXiv:1012.2063"},{"issue":"2","key":"6597_CR16","doi-asserted-by":"publisher","first-page":"407","DOI":"10.1214\/009053604000000067","volume":"32","author":"B Efron","year":"2004","unstructured":"Efron, B., Hastie, T., Johnstone, I., & Tibshirani, R. (2004). Least angle regression. The Annals of Statistics, 32(2), 407\u2013499.","journal-title":"The Annals of Statistics"},{"issue":"11","key":"6597_CR17","doi-asserted-by":"publisher","first-page":"4164","DOI":"10.1118\/1.2786864","volume":"34","author":"M Elter","year":"2007","unstructured":"Elter, M., Schulz-Wendtland, R., & Wittenberg, T. (2007). The prediction of breast cancer biopsy outcomes using two cad approaches that both emphasize an intelligible decision process. Medical Physics, 34(11), 4164\u201372.","journal-title":"Medical Physics"},{"key":"6597_CR18","doi-asserted-by":"crossref","unstructured":"Ferreira, L. A., Guimar\u00e3es, F. G., & Silva, R. (2020). Applying genetic programming to improve interpretability in machine learning models. In 2020 IEEE Congress on Evolutionary Computation (cec) (pp. 1\u20138).","DOI":"10.1109\/CEC48606.2020.9185620"},{"key":"6597_CR19","unstructured":"Ghosal, I., Zhou, Y., & Hooker, G. (2022). The infinitesimal jackknife and combinations of models. arXiv preprint arXiv:2209.00147"},{"issue":"7","key":"6597_CR20","doi-asserted-by":"publisher","first-page":"3579","DOI":"10.4088\/JCP.12m08338","volume":"74","author":"RD Gibbons","year":"2013","unstructured":"Gibbons, R. D., Hooker, G., Finkelman, M. D., Weiss, D. J., Pilkonis, P. A., Frank, E., & Kupfer, D. J. (2013). The computerized adaptive diagnostic test for major depressive disorder (cad-mdd): A screening tool for depression. The Journal of Clinical Psychiatry, 74(7), 3579.","journal-title":"The Journal of Clinical Psychiatry"},{"key":"6597_CR21","unstructured":"Hinton, G., Vinyals, O., & Dean, J. et al. (2015). Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 2(7)."},{"key":"6597_CR22","doi-asserted-by":"crossref","unstructured":"Johansson, U., S\u00f6nstr\u00f6d, C. & L\u00f6fstr\u00f6m, T. (2011). One tree to explain them all. In 2011 IEEE Congress of Evolutionary Computation (CEC) (pp. 1444\u20131451).","DOI":"10.1109\/CEC.2011.5949785"},{"key":"6597_CR23","unstructured":"Kingma, D.P. & Welling, M. (2013). Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114."},{"key":"6597_CR24","unstructured":"Koza, J. (1992). On the programming of computers by means of natural selection. Genetic Programming."},{"key":"6597_CR25","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4614-6849-3","volume-title":"Applied predictive modeling","author":"M Kuhn","year":"2013","unstructured":"Kuhn, M., Johnson, K., et al. (2013). Applied predictive modeling (Vol. 26). Berlin: Springer."},{"key":"6597_CR26","doi-asserted-by":"publisher","DOI":"10.1002\/9781118914564","volume-title":"Combining pattern classifiers: Methods and algorithms","author":"LI Kuncheva","year":"2014","unstructured":"Kuncheva, L. I. (2014). Combining pattern classifiers: Methods and algorithms. London: Wiley."},{"issue":"1","key":"6597_CR27","first-page":"14","volume":"1","author":"W-Y Loh","year":"2011","unstructured":"Loh, W.-Y. (2011). Classification and regression trees. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 1(1), 14\u201323.","journal-title":"Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery"},{"key":"6597_CR28","unstructured":"Lundberg, S. M., & Lee, S.-I. (2017). A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems 30."},{"issue":"4","key":"6597_CR29","doi-asserted-by":"publisher","first-page":"570","DOI":"10.1287\/opre.43.4.570","volume":"43","author":"OL Mangasarian","year":"1995","unstructured":"Mangasarian, O. L., Street, W. N., & Wolberg, W. H. (1995). Breast cancer diagnosis and prognosis via linear programming. Operations Research, 43(4), 570\u2013577.","journal-title":"Operations Research"},{"key":"6597_CR30","doi-asserted-by":"publisher","DOI":"10.7717\/peerj-cs.103","volume":"3","author":"A Meurer","year":"2017","unstructured":"Meurer, A., Smith, C. P., Paprocki, M., \u010cert\u00edk, O., Kirpichev, S. B., & Rocklin, M. (2017). Sympy: Symbolic computing in python. PeerJ Computer Science, 3, e103.","journal-title":"PeerJ Computer Science"},{"key":"6597_CR31","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1016\/j.dss.2014.03.001","volume":"62","author":"S Moro","year":"2014","unstructured":"Moro, S., Cortez, P., & Rita, P. (2014). A data-driven approach to predict the success of bank telemarketing. Decision Support Systems, 62, 22\u201331.","journal-title":"Decision Support Systems"},{"issue":"44","key":"6597_CR32","doi-asserted-by":"publisher","first-page":"22071","DOI":"10.1073\/pnas.1900654116","volume":"116","author":"WJ Murdoch","year":"2019","unstructured":"Murdoch, W. J., Singh, C., Kumbier, K., Abbasi-Asl, R., & Yu, B. (2019). Definitions, methods, and applications in interpretable machine learning. Proceedings of the National Academy of Sciences, 116(44), 22071\u201322080.","journal-title":"Proceedings of the National Academy of Sciences"},{"key":"6597_CR33","unstructured":"National Academies of Sciences, E. & Medicine. (2019). Reproducibility and replicability in science. National Academies Press."},{"key":"6597_CR34","first-page":"2825","volume":"12","author":"F Pedregosa","year":"2011","unstructured":"Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., & Duchesnay, E. (2011). Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12, 2825\u20132830.","journal-title":"Journal of Machine Learning Research"},{"key":"6597_CR35","first-page":"304","volume":"87","author":"JR Quinlan","year":"1987","unstructured":"Quinlan, J. R. (1987). Generating production rules from decision trees. Ijcai, 87, 304\u2013307.","journal-title":"Ijcai"},{"key":"6597_CR36","volume-title":"C4.5: Programs for machine learning","author":"JR Quinlan","year":"2014","unstructured":"Quinlan, J. R. (2014). C4.5: Programs for machine learning. Amsterdam: Elsevier."},{"key":"6597_CR37","doi-asserted-by":"crossref","unstructured":"Ribeiro, M. T., Singh, S. & Guestrin, C. (2016). \u201c why should i trust you?\u201d explaining the predictions of any classifier. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining (pp. 1135\u20131144).","DOI":"10.1145\/2939672.2939778"},{"key":"6597_CR38","unstructured":"Ruiqurm (2018). Kaggle. https:\/\/www.kaggle.com\/datasets\/ruiqurm\/lianjia (House price in Beijing)"},{"key":"6597_CR39","doi-asserted-by":"crossref","unstructured":"Tan, S., Caruana, R., Hooker, G., & Lou, Y. (2018). Distill-and-compare: Auditing black-box models using transparent model distillation. In Proceedings of the 2018 Aaai\/acm Conference on Ai, Ethics, and Society (pp. 303\u2013310).","DOI":"10.1145\/3278721.3278725"},{"issue":"1","key":"6597_CR40","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s43586-020-00001-2","volume":"1","author":"R van de Schoot","year":"2021","unstructured":"van de Schoot, R., Depaoli, S., King, R., Kramer, B., M\u00e4rtens, K., & Tadesse, M. G. (2021). Bayesian statistics and modelling. Nature Reviews Methods Primers, 1(1), 1\u201326.","journal-title":"Nature Reviews Methods Primers"},{"key":"6597_CR41","volume-title":"Python 3 reference manual","author":"G Van Rossum","year":"2009","unstructured":"Van Rossum, G., & Drake, F. L. (2009). Python 3 reference manual. Scotts Valley: CreateSpace."},{"key":"6597_CR42","doi-asserted-by":"publisher","DOI":"10.1201\/b14876","volume-title":"Kernel smoothing","author":"MP Wand","year":"1994","unstructured":"Wand, M. P., & Jones, M. C. (1994). Kernel smoothing. Boca Raton: CRC Press."},{"key":"6597_CR43","unstructured":"Wang, F., & Rudin, C. (2015). Falling rule lists. Artificial Intelligence and Statistics (pp. 1013\u20131022)."},{"issue":"1","key":"6597_CR44","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10115-007-0114-2","volume":"14","author":"X Wu","year":"2008","unstructured":"Wu, X., Kumar, V., Ross Quinlan, J., Ghosh, J., Yang, Q., & Motoda, H. (2008). Top 10 algorithms in data mining. Knowledge and Information Systems, 14(1), 1\u201337.","journal-title":"Knowledge and Information Systems"},{"key":"6597_CR45","unstructured":"Zhou, Y., Zhou, Z., & Hooker, G. (2018). Approximation trees: Statistical stability in model distillation. arXiv preprint arXiv:1808.07573"},{"key":"6597_CR46","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Hooker, G., & Wang, F. (2021). S-lime: Stabilized-lime for model explanation. In Proceedings of the 27th Acm Sigkdd Conference on Knowledge Discovery & Data Mining (pp. 2429\u20132438).","DOI":"10.1145\/3447548.3467274"}],"container-title":["Machine Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-024-06597-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10994-024-06597-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-024-06597-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,17]],"date-time":"2024-10-17T21:09:08Z","timestamp":1729199348000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10994-024-06597-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,5]]},"references-count":46,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["6597"],"URL":"https:\/\/doi.org\/10.1007\/s10994-024-06597-w","relation":{},"ISSN":["0885-6125","1573-0565"],"issn-type":[{"value":"0885-6125","type":"print"},{"value":"1573-0565","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,8,5]]},"assertion":[{"value":"2 May 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 July 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 July 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 August 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The presented research did not involve human or animal participants and required no ethics approval.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"The presented research did not involve human participants and thus consents to participate do not apply.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"All authors consent for submission, and publication.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}]}}