{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T00:16:29Z","timestamp":1769818589476,"version":"3.49.0"},"reference-count":23,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2026,1,3]],"date-time":"2026-01-03T00:00:00Z","timestamp":1767398400000},"content-version":"vor","delay-in-days":2,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100016378","name":"Technische Universit\u00e4t Dortmund","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100016378","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Comput Stat"],"published-print":{"date-parts":[[2026,1]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>\n                    The induction of additional randomness in parallel and sequential ensemble methods has proven to be worthwhile in many aspects. In this manuscript, we propose a novel random tree depth approach for sequential and parallel tree-based approaches. In particular, we apply the concept of a random tree depth for the representative methods of Boosting (MART) and Random Forests. Both approaches are then investigated with respect to their runtime and prediction performance. We call the resulting methods\n                    <jats:italic>Random Depth Boosting<\/jats:italic>\n                    and\n                    <jats:italic>Random Depth Forest<\/jats:italic>\n                    . Initially, an exemplary experiment on a simple data set indicates that combining\n                    <jats:italic>Random Depth<\/jats:italic>\n                    with\n                    <jats:italic>MART<\/jats:italic>\n                    can enhance prediction performance, while the impact on\n                    <jats:italic>Random Forests<\/jats:italic>\n                    remains limited. This observation aligns with a heuristically intuitive understanding of how randomizing tree depth interacts with bagging and boosting dynamics. Though a full theoretical analysis lies beyond the scope of this work, the underlying mechanisms can be well-motivated and heuristically explained. Building on these insights, a Monte Carlo simulation study investigates the effects on both artificial tree-shaped data sets with varying numbers of final partitions and on a selection of real-world classification and regression datasets. The results show that\n                    <jats:italic>Random Depth Boosting<\/jats:italic>\n                    offers relevant improvements for MART-based models. Additionally, the randomization of tree depth can reduce computation time by up to\n                    <jats:inline-formula>\n                      <jats:tex-math>$$40\\%$$<\/jats:tex-math>\n                    <\/jats:inline-formula>\n                    , making the approach an attractive option in time-sensitive applications.\n                  <\/jats:p>","DOI":"10.1007\/s00180-025-01697-0","type":"journal-article","created":{"date-parts":[[2026,1,3]],"date-time":"2026-01-03T08:34:37Z","timestamp":1767429277000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["The impact of random tree depth\u2014a novel randomization process for ensemble methods"],"prefix":"10.1007","volume":"41","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5208-0482","authenticated-orcid":false,"given":"Daniel","family":"Horn","sequence":"first","affiliation":[]},{"given":"Tobias Markus","family":"Krabel","sequence":"additional","affiliation":[]},{"given":"Thi Ngoc Tien","family":"Tran","sequence":"additional","affiliation":[]},{"given":"Andreas","family":"Groll","sequence":"additional","affiliation":[]},{"given":"Carsten","family":"Jentsch","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,3]]},"reference":[{"key":"1697_CR1","volume-title":"Classification and regression trees","author":"L Breiman","year":"1984","unstructured":"Breiman L, Friedman J, Stone CJ, Olshen RA (1984) Classification and regression trees. CRC Press, Boca Raton"},{"issue":"2","key":"1697_CR2","doi-asserted-by":"publisher","first-page":"123","DOI":"10.1023\/A:1018054314350","volume":"24","author":"L Breiman","year":"1996","unstructured":"Breiman L (1996) Bagging predictors. Mach Learn 24(2):123\u2013140","journal-title":"Mach Learn"},{"issue":"6","key":"1697_CR3","doi-asserted-by":"publisher","first-page":"2350","DOI":"10.1214\/aos\/1032181158","volume":"24","author":"L Breiman","year":"1996","unstructured":"Breiman L (1996) Heuristics of instability and stabilization in model selection. Ann Stat 24(6):2350\u20132383","journal-title":"Ann Stat"},{"issue":"1","key":"1697_CR4","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1023\/A:1010933404324","volume":"45","author":"L Breiman","year":"2001","unstructured":"Breiman L (2001) Random forests. Mach Learn 45(1):5\u201332","journal-title":"Mach Learn"},{"issue":"4","key":"1697_CR5","doi-asserted-by":"publisher","first-page":"927","DOI":"10.1214\/aos\/1031689014","volume":"30","author":"P B\u00fchlmann","year":"2002","unstructured":"B\u00fchlmann P, Yu B (2002) Analyzing bagging. Ann Stat 30(4):927\u2013961","journal-title":"Ann Stat"},{"issue":"2","key":"1697_CR6","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1109\/4235.996017","volume":"6","author":"K Deb","year":"2002","unstructured":"Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic algorithm: Nsga-ii. Trans Evol Comp 6(2):182\u2013197","journal-title":"Trans Evol Comp"},{"key":"1697_CR7","volume-title":"The elements of statistical learning","author":"J Friedman","year":"2001","unstructured":"Friedman J, Hastie T, Tibshirani R (2001) The elements of statistical learning. Springer, Berlin"},{"issue":"5","key":"1697_CR8","doi-asserted-by":"publisher","first-page":"1189","DOI":"10.1214\/aos\/1013203451","volume":"29","author":"J Friedman","year":"2001","unstructured":"Friedman J (2001) Greedy function approximation: a gradient boosting machine. Ann Stat 29(5):1189\u20131232","journal-title":"Ann Stat"},{"issue":"4","key":"1697_CR9","doi-asserted-by":"publisher","first-page":"367","DOI":"10.1016\/S0167-9473(01)00065-2","volume":"38","author":"J Friedman","year":"2002","unstructured":"Friedman J (2002) Stochastic gradient boosting. Comput Stat Data Anal 38(4):367\u2013378","journal-title":"Comput Stat Data Anal"},{"key":"1697_CR10","unstructured":"Freund Y, Schapire RE (1996) Experiments with a new boosting algorithm. In: Proceedings of the Thirteenth International Conference on International Conference on Machine Learning, pp. 148\u2013156. Morgan Kaufmann Publishers Inc"},{"issue":"1","key":"1697_CR11","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/s10994-006-6226-1","volume":"63","author":"P Geurts","year":"2006","unstructured":"Geurts P, Ernst D, Wehenkel L (2006) Extremely randomized trees. Mach Learn 63(1):3\u201342","journal-title":"Mach Learn"},{"issue":"9","key":"1697_CR12","doi-asserted-by":"publisher","first-page":"1344","DOI":"10.1016\/j.neunet.2007.12.046","volume":"21","author":"N Garc\u00eda-Pedrajas","year":"2008","unstructured":"Garc\u00eda-Pedrajas N, Ortiz-Boyer D (2008) Boosting random subspace method. Neural Netw 21(9):1344\u20131362","journal-title":"Neural Netw"},{"issue":"1","key":"1697_CR13","first-page":"183","volume":"2","author":"D Horn","year":"2017","unstructured":"Horn D, Bischl B, Demircioglu A, Glasmachers T, Wagner T, Weihs C (2017) Multi-objective selection of algorithm portfolios. Arch Data Sci Ser A (Online First) 2(1):183\u2013196","journal-title":"Arch Data Sci Ser A (Online First)"},{"issue":"8","key":"1697_CR14","doi-asserted-by":"publisher","first-page":"832","DOI":"10.1109\/34.709601","volume":"20","author":"TK Ho","year":"1998","unstructured":"Ho TK (1998) The random subspace method for constructing decision forests. IEEE Trans Pattern Anal Mach Intell 20(8):832\u2013844","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1697_CR15","doi-asserted-by":"publisher","DOI":"10.5281\/zenodo.16994355","author":"D Horn","year":"2025","unstructured":"Horn D (2025) Supplementary programming code and data for \u201cThe impact of random tree depth\u2014a novel randomization process for ensemble methods\u2019\u2019. Zenodo. https:\/\/doi.org\/10.5281\/zenodo.16994355","journal-title":"Zenodo"},{"issue":"1","key":"1697_CR16","doi-asserted-by":"publisher","first-page":"51","DOI":"10.1214\/aos\/1015362184","volume":"30","author":"W Jiang","year":"2002","unstructured":"Jiang W (2002) On weak base hypotheses and their implications for boosting regression and classification. Ann Stat 30(1):51\u201373","journal-title":"Ann Stat"},{"key":"1697_CR17","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4614-6849-3","volume-title":"Applied predictive modeling","author":"M Kuhn","year":"2013","unstructured":"Kuhn M, Johnson K (2013) Applied predictive modeling. Springer, Berlin"},{"key":"1697_CR18","doi-asserted-by":"publisher","first-page":"319","DOI":"10.1023\/A:1022645801436","volume":"3","author":"J Mingers","year":"1989","unstructured":"Mingers J (1989) An empirical comparison of selection measures for decision-tree induction. Mach Learn 3:319\u2013342","journal-title":"Mach Learn"},{"key":"1697_CR19","doi-asserted-by":"publisher","first-page":"197","DOI":"10.1023\/A:1022648800760","volume":"5","author":"RE Schapire","year":"1990","unstructured":"Schapire RE (1990) The strength of weak learnability. Mach Learn 5:197\u2013227","journal-title":"Mach Learn"},{"key":"1697_CR20","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1007\/s100440200011","volume":"5","author":"M Skurichina","year":"2002","unstructured":"Skurichina M, Duin RPW (2002) Bagging, boosting and the random subspace method for linear classifiers. Pattern Anal Appl 5:121\u2013135","journal-title":"Pattern Anal Appl"},{"key":"1697_CR21","unstructured":"Segal MR (2004) Machine learning benchmarks and random forest regression. Center Bioinform Mol Biostat"},{"key":"1697_CR22","unstructured":"Vinayak RK, Gilad-Bachrach R (2015) Dart: Dropouts meet multiple additive regression trees. In: Artificial Intelligence and Statistics, pp. 489\u2013497"},{"issue":"2","key":"1697_CR23","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1145\/2641190.2641198","volume":"15","author":"J Vanschoren","year":"2013","unstructured":"Vanschoren J, Rijn JN, Bischl B, Torgo L (2013) Openml: Networked science in machine learning. SIGKDD Explor 15(2):49\u201360","journal-title":"SIGKDD Explor"}],"container-title":["Computational Statistics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00180-025-01697-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00180-025-01697-0","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00180-025-01697-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T12:14:06Z","timestamp":1769775246000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00180-025-01697-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1]]},"references-count":23,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["1697"],"URL":"https:\/\/doi.org\/10.1007\/s00180-025-01697-0","relation":{},"ISSN":["0943-4062","1613-9658"],"issn-type":[{"value":"0943-4062","type":"print"},{"value":"1613-9658","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1]]},"assertion":[{"value":"10 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors report there are no competing interests to declare.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interest"}}],"article-number":"25"}}