{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T07:45:17Z","timestamp":1740123917507,"version":"3.37.3"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T00:00:00Z","timestamp":1666396800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T00:00:00Z","timestamp":1666396800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100004543","name":"China Scholarship Council","doi-asserted-by":"publisher","award":["202006220228"],"award-info":[{"award-number":["202006220228"]}],"id":[{"id":"10.13039\/501100004543","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001711","name":"Schweizerischer Nationalfonds zur F\u00f6rderung der Wissenschaftlichen Forschung","doi-asserted-by":"publisher","award":["197226"],"award-info":[{"award-number":["197226"]}],"id":[{"id":"10.13039\/501100001711","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000002","name":"National Institutes of Health","doi-asserted-by":"publisher","award":["DE-AC5206NA25396"],"award-info":[{"award-number":["DE-AC5206NA25396"]}],"id":[{"id":"10.13039\/100000002","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Stat Comput"],"published-print":{"date-parts":[[2022,12]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>To fit sparse linear associations, a LASSO sparsity inducing penalty with a single hyperparameter provably allows to recover the important features (needles) with high probability in certain regimes even if the sample size is smaller than the dimension of the input vector (haystack). More recently learners known as artificial neural networks (ANN) have shown great successes in many machine learning tasks, in particular fitting nonlinear associations. Small learning rate, stochastic gradient descent algorithm and large training set help to cope with the explosion in the number of parameters present in deep neural networks. Yet few ANN learners have been developed and studied to find needles in nonlinear haystacks. Driven by a single hyperparameter, our ANN learner, like for sparse linear associations, exhibits a phase transition in the probability of retrieving the needles, which we do not observe with other ANN learners. To select our penalty parameter, we generalize the universal threshold of Donoho and Johnstone (Biometrika 81(3):425\u2013455, 1994) which is a better rule than the conservative (too many false detections) and expensive cross-validation. In the spirit of simulated annealing, we propose a warm-start sparsity inducing algorithm to solve the high-dimensional, non-convex and non-differentiable optimization problem. We perform simulated and real data Monte Carlo experiments to quantify the effectiveness of our approach.<\/jats:p>","DOI":"10.1007\/s11222-022-10169-0","type":"journal-article","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T09:11:53Z","timestamp":1666429913000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A phase transition for finding needles in nonlinear haystacks with LASSO artificial neural networks"],"prefix":"10.1007","volume":"32","author":[{"given":"Xiaoyu","family":"Ma","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4935-2595","authenticated-orcid":false,"given":"Sylvain","family":"Sardy","sequence":"additional","affiliation":[]},{"given":"Nick","family":"Hengartner","sequence":"additional","affiliation":[]},{"given":"Nikolai","family":"Bobenko","sequence":"additional","affiliation":[]},{"given":"Yen Ting","family":"Lin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,22]]},"reference":[{"key":"10169_CR1","unstructured":"Adcock, B., Brugiapaglia, S., Dexter, N., Morage, S.: Deep neural networks are effective at learning high-dimensional Hilbert-valued functions from limited data. In: Proceedings of the 2nd Mathematical and Scientific Machine Learning Conference, vol. 145, pp. 1\u201336. PMLR (2022)"},{"issue":"2","key":"10169_CR2","doi-asserted-by":"publisher","first-page":"624","DOI":"10.1137\/20M131309X","volume":"3","author":"B Adcock","year":"2021","unstructured":"Adcock, B., Dexter, N.: The gap between theory and practice in function approximation with deep neural networks. SIAM J. Math. Data Sci. 3(2), 624\u2013655 (2021)","journal-title":"SIAM J. Math. Data Sci."},{"key":"10169_CR3","doi-asserted-by":"publisher","first-page":"428","DOI":"10.1016\/j.neunet.2020.08.022","volume":"132","author":"MS Advani","year":"2020","unstructured":"Advani, M.S., Saxe, A.M., Sompolinsky, H.: High-dimensional dynamics of generalization error in neural networks. Neural Netw. 132, 428\u2013446 (2020)","journal-title":"Neural Netw."},{"key":"10169_CR4","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1214\/09-SS054","volume":"4","author":"S Arlot","year":"2010","unstructured":"Arlot, S., Celisse, A.: A survey of cross-validation procedures for model selection. Stat. Surv. 4, 40\u201379 (2010)","journal-title":"Stat. Surv."},{"issue":"1","key":"10169_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1561\/2200000015","volume":"4","author":"F Bach","year":"2011","unstructured":"Bach, F., Jenatton, R., Mairal, J., Obozinski, G.: Optimization with sparsity-inducing penalties. Found. Trends Mach. Learn. 4(1), 1\u2013106 (2011)","journal-title":"Found. Trends Mach. Learn."},{"issue":"3","key":"10169_CR6","doi-asserted-by":"publisher","first-page":"930","DOI":"10.1109\/18.256500","volume":"39","author":"AR Barron","year":"1993","unstructured":"Barron, A.R.: Universal approximation bounds for superpositions of a sigmoidal function. IEEE Trans. Inf. Theory 39(3), 930\u2013945 (1993)","journal-title":"IEEE Trans. Inf. Theory"},{"key":"10169_CR7","unstructured":"Bastounis, A., Hansen, A.C., Vlacic, V.: The mathematics of adversarial attacks in AI\u2014why deep learning is unstable despite the existence of stable neural networks. arXiv:2109.06098 (2021b)"},{"key":"10169_CR8","unstructured":"Bastounis, A., Hansen, A.C., Vlavcic, V.: The extended Smale\u2019s 9th problem\u2014on computational barriers and paradoxes in estimation, regularisation, computer-assisted proofs and learning (2021a)"},{"key":"10169_CR9","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1137\/080716542","volume":"2","author":"A Beck","year":"2009","unstructured":"Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imag. Sci. 2, 183\u2013202 (2009)","journal-title":"SIAM J. Imag. Sci."},{"issue":"4","key":"10169_CR10","doi-asserted-by":"publisher","first-page":"791","DOI":"10.1093\/biomet\/asr043","volume":"98","author":"A Belloni","year":"2011","unstructured":"Belloni, A., Chernozhukov, V., Wang, L.: Square-root lasso: pivotal recovery of sparse signals via conic programming. Biometrika 98(4), 791\u2013806 (2011)","journal-title":"Biometrika"},{"issue":"1","key":"10169_CR11","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1137\/18M118709X","volume":"1","author":"H B\u00f6lcskei","year":"2019","unstructured":"B\u00f6lcskei, H., Grohs, P., Kutyniok, G., Petersen, P.: Optimal approximation with sparsely connected deep neural networks. SIAM J. Math. Data Sci. 1(1), 8\u201345 (2019)","journal-title":"SIAM J. Math. Data Sci."},{"issue":"1","key":"10169_CR12","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1023\/A:1010933404324","volume":"45","author":"L Breiman","year":"2001","unstructured":"Breiman, L.: Random forests. Mach. Learn. 45(1), 5\u201332 (2001)","journal-title":"Mach. Learn."},{"key":"10169_CR13","volume-title":"Classification and Regression Trees","author":"L Breiman","year":"1984","unstructured":"Breiman, L., Friedman, J., Olshen, R., Stone, C.: Classification and Regression Trees. Routledge, Boca Raton (1984)"},{"key":"10169_CR14","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-20192-9","volume-title":"Statistics for High-Dimensional Data: Methods, Theory and Applications","author":"P B\u00fchlmann","year":"2011","unstructured":"B\u00fchlmann, P., van de Geer, S.: Statistics for High-Dimensional Data: Methods, Theory and Applications. Springer, Heidelberg (2011)"},{"key":"10169_CR15","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1146\/annurev-statistics-022513-115545","volume":"1","author":"P B\u00fchlmann","year":"2014","unstructured":"B\u00fchlmann, P., Kalisch, M., Meier, L.: High-dimensional statistics with a view toward applications in biology. Ann. Rev. Stat. Appl. 1, 255\u2013278 (2014)","journal-title":"Ann. Rev. Stat. Appl."},{"key":"10169_CR16","doi-asserted-by":"publisher","first-page":"4203","DOI":"10.1109\/TIT.2005.858979","volume":"51","author":"EJ Cand\u00e8s","year":"2005","unstructured":"Cand\u00e8s, E.J., Tao, T.: Decoding by linear programming. IEEE Trans. Inf. Theory 51, 4203\u20134215 (2005)","journal-title":"IEEE Trans. Inf. Theory"},{"key":"10169_CR17","doi-asserted-by":"crossref","unstructured":"Carreira-Perpinan, M.A., Idelbayev, Y.: Learning-compression algorithms for neural net pruning. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8532\u20138541 (2018)","DOI":"10.1109\/CVPR.2018.00890"},{"key":"10169_CR18","unstructured":"Chao, S.K., Wang, Z., Xing, Y., Cheng, G.: Directional pruning of deep neural networks. In: NeurIPS (2020)"},{"issue":"4","key":"10169_CR19","doi-asserted-by":"publisher","first-page":"911","DOI":"10.1109\/72.392253","volume":"6","author":"T Chen","year":"1995","unstructured":"Chen, T., Chen, H.: Universal approximation to nonlinear operators by neural networks with arbitrary activation functions and its application to dynamical systems. IEEE Trans. Neural Netw. 6(4), 911\u2013917 (1995)","journal-title":"IEEE Trans. Neural Netw."},{"issue":"1","key":"10169_CR20","doi-asserted-by":"publisher","first-page":"33","DOI":"10.1137\/S1064827596304010","volume":"20","author":"SS Chen","year":"1999","unstructured":"Chen, S.S., Donoho, D.L., Saunders, M.A.: Atomic decomposition by basis pursuit. SIAM J. Sci. Comput. 20(1), 33\u201361 (1999)","journal-title":"SIAM J. Sci. Comput."},{"issue":"12","key":"10169_CR21","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2107151119","volume":"119","author":"MJ Colbrook","year":"2022","unstructured":"Colbrook, M.J., Antun, V., Hansen, A.C.: The difficulty of computing stable and accurate neural networks: on the barriers of deep learning and Smale\u2019s 18th problem. Proc. Natl. Acad. Sci. 119(12), e2107151119 (2022)","journal-title":"Proc. Natl. Acad. Sci."},{"key":"10169_CR22","unstructured":"Collins, M.D., Kohli, P.: Memory bounded deep convolutional networks. arXiv:1412.1442 (2014)"},{"key":"10169_CR23","doi-asserted-by":"crossref","unstructured":"Curci, S., Mocanu, D.C., Pechenizkiyi, M.: Truly sparse neural networks at scale. arXiv:2102.01732 (2021)","DOI":"10.21203\/rs.3.rs-133395\/v1"},{"issue":"4","key":"10169_CR24","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/BF02551274","volume":"2","author":"G Cybenko","year":"1989","unstructured":"Cybenko, G.: Approximation by superpositions of a sigmoidal function. Math. Control Signals Syst. (MCSS) 2(4), 303\u2013314 (1989)","journal-title":"Math. Control Signals Syst. (MCSS)"},{"issue":"3","key":"10169_CR25","doi-asserted-by":"publisher","first-page":"530","DOI":"10.1080\/10618600.2020.1869026","volume":"30","author":"P Descloux","year":"2021","unstructured":"Descloux, P., Sardy, S.: Model selection with lasso-zero: adding straw in the haystack to better find needles. J. Comput. Graph. Stat. 30(3), 530\u2013543 (2021)","journal-title":"J. Comput. Graph. Stat."},{"key":"10169_CR26","doi-asserted-by":"publisher","first-page":"1289","DOI":"10.1109\/TIT.2006.871582","volume":"52","author":"DL Donoho","year":"2006","unstructured":"Donoho, D.L.: Compressed sensing. IEEE Trans. Inf. Theory 52, 1289\u20131306 (2006)","journal-title":"IEEE Trans. Inf. Theory"},{"issue":"3","key":"10169_CR27","doi-asserted-by":"publisher","first-page":"425","DOI":"10.1093\/biomet\/81.3.425","volume":"81","author":"DL Donoho","year":"1994","unstructured":"Donoho, D.L., Johnstone, I.M.: Ideal spatial adaptation by wavelet shrinkage. Biometrika 81(3), 425\u2013455 (1994)","journal-title":"Biometrika"},{"issue":"6","key":"10169_CR28","doi-asserted-by":"publisher","first-page":"913","DOI":"10.1109\/JPROC.2010.2045630","volume":"98","author":"DL Donoho","year":"2010","unstructured":"Donoho, D.L., Tanner, J.: Precise undersampling theorems. Proc. IEEE 98(6), 913\u2013924 (2010)","journal-title":"Proc. IEEE"},{"issue":"2","key":"10169_CR29","doi-asserted-by":"crossref","first-page":"301","DOI":"10.1111\/j.2517-6161.1995.tb02032.x","volume":"57","author":"DL Donoho","year":"1995","unstructured":"Donoho, D.L., Johnstone, I.M., Kerkyacharian, G., Picard, D.: Wavelet shrinkage: asymptopia? J. R. Stat. Soc. B 57(2), 301\u2013369 (1995)","journal-title":"J. R. Stat. Soc. B"},{"issue":"10","key":"10169_CR30","doi-asserted-by":"publisher","first-page":"6920","DOI":"10.1109\/TIT.2011.2165823","volume":"57","author":"DL Donoho","year":"2011","unstructured":"Donoho, D.L., Maleki, A., Montanari, A.: The noise-sensitivity phase transition in compressed sensing. IEEE Trans. Inf. Theory 57(10), 6920\u20136941 (2011)","journal-title":"IEEE Trans. Inf. Theory"},{"key":"10169_CR31","unstructured":"Evci, U., Pedregosa, F., Gomez, A.N., Elsen, E.: The difficulty of training sparse neural networks. arXiv:1906.10732 (2019)"},{"key":"10169_CR32","unstructured":"Feng, J., Simon, N.: Sparse-input neural networks for high-dimensional nonparametric regression and classification. arXiv:1711.07592 (2019)"},{"key":"10169_CR33","doi-asserted-by":"publisher","first-page":"817","DOI":"10.1080\/01621459.1981.10477729","volume":"76","author":"JH Friedman","year":"1981","unstructured":"Friedman, J.H., Stuetzle, W.: Projection pursuit regression. J. Am. Stat. Assoc. 76, 817\u2013823 (1981)","journal-title":"J. Am. Stat. Assoc."},{"issue":"1","key":"10169_CR34","doi-asserted-by":"publisher","first-page":"1","DOI":"10.18637\/jss.v033.i01","volume":"33","author":"JH Friedman","year":"2010","unstructured":"Friedman, J.H., Hastie, T., Tibshirani, R.: Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 33(1), 1\u201322 (2010)","journal-title":"J. Stat. Softw."},{"key":"10169_CR35","doi-asserted-by":"crossref","unstructured":"Geiger, M., Jacot, A., Spigler, S., Gabriel, F., Sagun, L., d\u2019Ascoli, S., Biroli, G., Hongler, C., Wyart, M.: Scaling description of generalization with number of parameters in deep learning. arXiv:1901.01608 (2019)","DOI":"10.1088\/1742-5468\/ab633c"},{"issue":"2","key":"10169_CR36","doi-asserted-by":"publisher","first-page":"4701","DOI":"10.1214\/17-EJS1366","volume":"11","author":"C Giacobino","year":"2017","unstructured":"Giacobino, C., Sardy, S., Diaz Rodriguez, J., Hengardner, N.: Quantile universal threshold. Electron. J. Stat. 11(2), 4701\u20134722 (2017)","journal-title":"Electron. J. Stat."},{"key":"10169_CR37","unstructured":"Grohs, P., Perekrestenko, D., Elbr\u00e4chter, D., B\u00f6lcskei, H.: Deep neural network approximation theory. arXiv:1901.02220 (2019)"},{"key":"10169_CR38","unstructured":"Hastie, T., Montanari, A., Rosset, S., Tibshirani, R.J.: Surprises in high-dimensional ridgeless least squares interpolation. arXiv:1903.08560 (2019)"},{"issue":"1","key":"10169_CR39","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1007\/s10589-020-00202-1","volume":"77","author":"J He","year":"2020","unstructured":"He, J., Jia, X., Xu, J., Zhang, L., Zhao, L.: Make $$\\ell _1$$ regularization effective in training sparse CNN. Comput. Optim. Appl. 77(1), 163\u2013182 (2020)","journal-title":"Comput. Optim. Appl."},{"key":"10169_CR40","unstructured":"Hinton, G., Srivastava, N., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Improving neural networks by preventing co-adaptation of feature detectors. arXiv:1207.0580 (2012)"},{"issue":"1","key":"10169_CR41","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1080\/00401706.1970.10488634","volume":"12","author":"AE Hoerl","year":"1970","unstructured":"Hoerl, A.E., Kennard, R.W.: Ridge regression: biased estimation for nonorthogonal problems. Technometrics 12(1), 55\u201367 (1970)","journal-title":"Technometrics"},{"key":"10169_CR42","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1214\/aos\/1193342384","volume":"1","author":"PW Holland","year":"1973","unstructured":"Holland, P.W.: Covariance stabilizing transformations. Ann. Stat. 1, 84\u201392 (1973)","journal-title":"Ann. Stat."},{"key":"10169_CR43","doi-asserted-by":"publisher","first-page":"1594","DOI":"10.1214\/009053604000000030","volume":"32","author":"IM Johnstone","year":"2004","unstructured":"Johnstone, I.M., Silverman, B.: Needles and straw in haystacks: empirical Bayes estimates of possibly sparse sequences. Ann. Stat. 32, 1594\u20131649 (2004)","journal-title":"Ann. Stat."},{"key":"10169_CR44","doi-asserted-by":"crossref","unstructured":"Kostadinov, D., Voloshynovskiy, S., Ferdowsi, S.: Learning overcomplete and sparsifying transform with approximate and exact closed form solutions. In: 2018 7th European Workshop on Visual Information Processing (EUVIP), pp. 1\u20136 (2018)","DOI":"10.1109\/EUVIP.2018.8611650"},{"key":"10169_CR45","doi-asserted-by":"crossref","unstructured":"Lee, H., Battle, A., Raina, R., Ng, A.Y.: Efficient sparse coding algorithms. In: Proceedings of the 19th International Conference on Neural Information Processing Systems, NIPS\u201906, pages 801\u2013808. MIT Press (2006)","DOI":"10.7551\/mitpress\/7503.003.0105"},{"key":"10169_CR46","doi-asserted-by":"publisher","first-page":"322","DOI":"10.1089\/cmb.2015.0189","volume":"23","author":"Y Li","year":"2016","unstructured":"Li, Y., Chen, C.Y., Wasserman, W.W.: Deep feature selection: theory and application to identify enhancers and promoters. J. Comput. Biol. 23, 322\u2013336 (2016)","journal-title":"J. Comput. Biol."},{"key":"10169_CR47","unstructured":"Ma, R., Miao, J., Niu, L., Zhang, P.: Transformed $$\\ell _1$$ regularization for learning sparse deep neural networks. arXiv:1901.01021 (2019)"},{"key":"10169_CR48","doi-asserted-by":"publisher","DOI":"10.1002\/cpa.22008","author":"S Mei","year":"2021","unstructured":"Mei, S., Montanari, A.: The generalization error of random features regression: precise asymptotics and double descent curve. Commun. Pure Appl. Math. (2021). https:\/\/doi.org\/10.1002\/cpa.22008","journal-title":"Commun. Pure Appl. Math."},{"key":"10169_CR49","unstructured":"Mollaysa, A., Strasser, P., Kalousis, A.: Regularising non-linear models using feature side-information. In: Proceedings of the 34th International Conference on Machine Learning, volume\u00a070 of Proceedings of Machine Learning Research, pp. 2508\u20132517, Sydney (2017)"},{"issue":"5","key":"10169_CR50","doi-asserted-by":"publisher","first-page":"503","DOI":"10.1007\/s11633-017-1054-2","volume":"14","author":"T Poggio","year":"2017","unstructured":"Poggio, T., Mhaskar, H., Rosasco, L., Miranda, B., Liao, Q.: Why and when can deep-but not shallow-networks avoid the curse of dimensionality: a review. Int. J. Autom. Comput. 14(5), 503\u2013519 (2017)","journal-title":"Int. J. Autom. Comput."},{"key":"10169_CR51","unstructured":"Ranzato, M.A., Boureau, Y.L., LeCun, Y.: Sparse feature learning for deep belief networks. In: Proceedings of the 20th International Conference on Neural Information Processing Systems, NIPS\u201907, pp. 1185\u20131192. Curran Associates Inc. (2007)"},{"key":"10169_CR52","doi-asserted-by":"crossref","unstructured":"Ravishankar, S., Wen, B., Bresler, Y.: Online sparsifying transform learning-part I: algorithms. IEEE J. Sel. Top. Signal Process. 9(4), 625\u2013636 (2015)","DOI":"10.1109\/JSTSP.2015.2417131"},{"key":"10169_CR53","doi-asserted-by":"publisher","first-page":"533","DOI":"10.1038\/323533a0","volume":"323","author":"DE Rumelhart","year":"1986","unstructured":"Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations by back-propagating errors. Nature 323, 533\u2013536 (1986)","journal-title":"Nature"},{"issue":"1","key":"10169_CR54","first-page":"1929","volume":"15","author":"N Srivastava","year":"2014","unstructured":"Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929\u20131958 (2014)","journal-title":"J. Mach. Learn. Res."},{"key":"10169_CR55","doi-asserted-by":"publisher","DOI":"10.1080\/01621459.2021.1895175","author":"Y Sun","year":"2021","unstructured":"Sun, Y., Song, Q., Liang, F.: Consistent sparse deep learning: theory and computation. J. Am. Stat. Assoc. (2021). https:\/\/doi.org\/10.1080\/01621459.2021.1895175","journal-title":"J. Am. Stat. Assoc."},{"issue":"1","key":"10169_CR56","doi-asserted-by":"crossref","first-page":"267","DOI":"10.1111\/j.2517-6161.1996.tb02080.x","volume":"58","author":"R Tibshirani","year":"1996","unstructured":"Tibshirani, R.: Regression shrinkage and selection via the lasso. J. R. Stat. Soc. B 58(1), 267\u2013288 (1996)","journal-title":"J. R. Stat. Soc. B"},{"key":"10169_CR57","unstructured":"Ye, M., Sun, Y.: Variable selection via penalized neural network: a drop-out-one loss approach. In: Proceedings of the 35th International Conference on Machine Learning, vol.\u00a080, pp. 5620\u20135629 (2018)"},{"issue":"1","key":"10169_CR58","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1111\/j.1467-9868.2005.00532.x","volume":"68","author":"M Yuan","year":"2006","unstructured":"Yuan, M., Lin, Y.: Model selection and estimation in regression with grouped variables. J. R. Stat. Soc. B 68(1), 49\u201367 (2006)","journal-title":"J. R. Stat. Soc. B"}],"container-title":["Statistics and Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11222-022-10169-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11222-022-10169-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11222-022-10169-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T10:01:57Z","timestamp":1728208917000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11222-022-10169-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10,22]]},"references-count":58,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2022,12]]}},"alternative-id":["10169"],"URL":"https:\/\/doi.org\/10.1007\/s11222-022-10169-0","relation":{},"ISSN":["0960-3174","1573-1375"],"issn-type":[{"type":"print","value":"0960-3174"},{"type":"electronic","value":"1573-1375"}],"subject":[],"published":{"date-parts":[[2022,10,22]]},"assertion":[{"value":"28 February 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 October 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 October 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"99"}}