{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2022,4,4]],"date-time":"2022-04-04T12:49:40Z","timestamp":1649076580028},"publisher-location":"New York, NY, USA","reference-count":32,"publisher":"ACM","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,10,26]]},"DOI":"10.1145\/3459637.3482142","type":"proceedings-article","created":{"date-parts":[[2021,10,30]],"date-time":"2021-10-30T18:34:11Z","timestamp":1635618851000},"source":"Crossref","is-referenced-by-count":0,"title":["Mitigating Deep Double Descent by Concatenating Inputs"],"prefix":"10.1145","author":[{"given":"John","family":"Chen","sequence":"first","affiliation":[{"name":"Rice University, Houston, TX, USA"}]},{"given":"Qihan","family":"Wang","sequence":"additional","affiliation":[{"name":"Rice University, Houston, TX, USA"}]},{"given":"Anastasios","family":"Kyrillidis","sequence":"additional","affiliation":[{"name":"Rice University, Houston, TX, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,10,30]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"High-dimensional dynamics of generalization error in neural networks. arXiv preprint arXiv:1710.03667","author":"Advani Madhu S","year":"2017"},{"key":"e_1_3_2_1_2_1","unstructured":"Jimmy Ba Murat Erdogdu Taiji Suzuki Denny Wu and Tianzong Zhang. 2020. Generalization of Two-layer Neural Networks: An Asymptotic Viewpoint. In ICLR. Jimmy Ba Murat Erdogdu Taiji Suzuki Denny Wu and Tianzong Zhang. 2020. Generalization of Two-layer Neural Networks: An Asymptotic Viewpoint. In ICLR."},{"key":"e_1_3_2_1_3_1","volume-title":"Benign overfitting in linear regression. arXiv preprint arXiv:1906.11300","author":"Bartlett Peter L","year":"2019"},{"key":"e_1_3_2_1_4_1","volume-title":"Reconciling modern machine learning practice and the bias-variance trade-off. arXiv preprint arXiv:1812.11118","author":"Belkin Mikhail","year":"2018"},{"key":"e_1_3_2_1_5_1","volume-title":"2019 a. Two models of double descent for weak features. arXiv preprint arXiv:1903.07571","author":"Belkin Mikhail","year":"2019"},{"key":"e_1_3_2_1_6_1","volume-title":"2019 b. Two models of double descent for weak features. arxiv","author":"Belkin Mikhail","year":"1903"},{"key":"e_1_3_2_1_7_1","volume-title":"To understand deep learning we need to understand kernel learning. arxiv","author":"Belkin Mikhail","year":"1802"},{"key":"e_1_3_2_1_8_1","volume-title":"A new look at an old problem: A universal learning approach to linear regression. arXiv preprint arXiv:1905.04708","author":"Bibas Koby","year":"2019"},{"key":"e_1_3_2_1_9_1","volume-title":"A finite sample analysis of the double descent phenomenon for ridge function estimation. arxiv","author":"Caron Emmanuel","year":"2007"},{"key":"e_1_3_2_1_10_1","volume-title":"Multiple Descent: Design Your Own Generalization Curve. arxiv","author":"Chen Lin","year":"2020"},{"key":"e_1_3_2_1_11_1","volume-title":"Triple descent and the two kinds of overfitting: Where and why do they appear?arxiv","author":"Ascoli St\u00e9phane","year":"2006"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1088\/1742-5468\/ab633c"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1103\/PhysRevE.100.012115"},{"key":"e_1_3_2_1_14_1","volume-title":"Neural networks and the bias\/variance dilemma. Neural Computation","author":"Geman Stuart","year":"1992"},{"key":"e_1_3_2_1_15_1","volume-title":"Linearized two-layers neural networks in high dimension. arxiv","author":"Ghorbani Behrooz","year":"1904"},{"key":"e_1_3_2_1_16_1","volume-title":"Surprises in high dimensional ridgeless least squares interpolation. arXiv preprint arXiv:1903.08560","author":"Hastie Trevor","year":"2019"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-0-387-21606-5"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_19_1","volume-title":"Laurens Van Der Maaten, and Kilian Q. Weinberger","author":"Huang Gao","year":"2017"},{"key":"e_1_3_2_1_20_1","volume-title":"Ultra-deep neural networks without residuals. arXiv preprint arXiv:1605.07648","author":"Larsson Gustav","year":"2016"},{"key":"e_1_3_2_1_21_1","volume-title":"On the Multiple Descent of Minimum-Norm Interpolants and Restricted Lower Isometry of Kernels. arxiv","author":"Liang Tengyuan","year":"1908"},{"key":"e_1_3_2_1_22_1","volume-title":"The generalization error of random features regression: Precise asymptotics and double descent curve. arXiv preprint arXiv:1908.05355","author":"Mei Song","year":"2019"},{"key":"e_1_3_2_1_23_1","volume-title":"Understanding overfitting peaks in generalization error: Analytical risk curves for l2 and l1 penalized interpolation. arXiv preprint arXiv:1906.03667","author":"Mitra Partha P.","year":"2019"},{"key":"e_1_3_2_1_24_1","volume-title":"Harmless interpolation of noisy data in regression. arXiv preprint arXiv:1903.09139","author":"Muthukumar Vidya","year":"2019"},{"key":"e_1_3_2_1_25_1","volume-title":"More Data Can Hurt for Linear Regression: Sample-wise Double Descent. arxiv","author":"Nakkiran Preetum","year":"1912"},{"key":"e_1_3_2_1_26_1","volume-title":"Deep double descent: Where bigger models and more data hurt. arXiv preprint arXiv:1912.02292","author":"Nakkiran Preetum","year":"2019"},{"key":"e_1_3_2_1_27_1","volume-title":"Optimal Regularization Can Mitigate Double Descent. arxiv","author":"Nakkiran Preetum","year":"2003"},{"key":"e_1_3_2_1_28_1","volume-title":"A Modern Take on the Bias-Variance Tradeoff in Neural Networks. arxiv","author":"Neal Brady","year":"1810"},{"key":"e_1_3_2_1_29_1","volume-title":"Deep learning tutorial at the Simons Institute","author":"Salakhutdinov Ruslan","year":"2017"},{"key":"e_1_3_2_1_30_1","volume-title":"Rethinking Bias-Variance Trade-off for Generalization of Neural Networks. arXiv preprint arXiv:2002.11328","author":"Yang Zitong","year":"2020"},{"key":"e_1_3_2_1_31_1","volume-title":"Wide residual networks. arXiv preprint arXiv:1605.07146","author":"Zagoruyko Sergey","year":"2016"},{"key":"e_1_3_2_1_32_1","unstructured":"Chiyuan Zhang Samy Bengio Moritz Hardt Benjamin Recht and Oriol Vinyals. 2016. Understanding deep learning requires rethinking generalization. arxiv: 1611.03530 [cs.LG] Chiyuan Zhang Samy Bengio Moritz Hardt Benjamin Recht and Oriol Vinyals. 2016. Understanding deep learning requires rethinking generalization. arxiv: 1611.03530 [cs.LG]"}],"event":{"name":"CIKM '21: The 30th ACM International Conference on Information and Knowledge Management","location":"Virtual Event Queensland Australia","acronym":"CIKM '21","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web","SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 30th ACM International Conference on Information & Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3459637.3482142","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,11,15]],"date-time":"2021-11-15T15:53:43Z","timestamp":1636991623000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3459637.3482142"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,26]]},"references-count":32,"alternative-id":["10.1145\/3459637.3482142","10.1145\/3459637"],"URL":"http:\/\/dx.doi.org\/10.1145\/3459637.3482142","relation":{},"published":{"date-parts":[[2021,10,26]]}}}