{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,16]],"date-time":"2025-10-16T07:02:26Z","timestamp":1760598146019,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":58,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,4,25]],"date-time":"2022-04-25T00:00:00Z","timestamp":1650844800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["IIS-2128019, IIS-1553568"],"award-info":[{"award-number":["IIS-2128019, IIS-1553568"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100006785","name":"Google","doi-asserted-by":"publisher","award":["Faculty Research Award"],"award-info":[{"award-number":["Faculty Research Award"]}],"id":[{"id":"10.13039\/100006785","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,4,25]]},"DOI":"10.1145\/3485447.3512250","type":"proceedings-article","created":{"date-parts":[[2022,4,25]],"date-time":"2022-04-25T05:13:07Z","timestamp":1650863587000},"page":"431-441","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":5,"title":["Learning Neural Ranking Models Online from Implicit User Feedback"],"prefix":"10.1145","author":[{"given":"Yiling","family":"Jia","sequence":"first","affiliation":[{"name":"University of Virginia, USA"}]},{"given":"Hongning","family":"Wang","sequence":"additional","affiliation":[{"name":"University of Virginia, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,4,25]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Yasin Abbasi-Yadkori D\u00e1vid P\u00e1l and Csaba Szepesv\u00e1ri. 2011. Improved algorithms for linear stochastic bandits. In Advances in Neural Information Processing Systems. 2312\u20132320."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/1148170.1148177"},{"key":"e_1_3_2_1_3_1","volume-title":"International Conference on Machine Learning. PMLR, 242\u2013252","author":"Allen-Zhu Zeyuan","year":"2019","unstructured":"Zeyuan Allen-Zhu, Yuanzhi Li, and Zhao Song. 2019. A convergence theory for deep learning via over-parameterization. In International Conference on Machine Learning. PMLR, 242\u2013252."},{"key":"e_1_3_2_1_4_1","unstructured":"Sanjeev Arora Simon\u00a0S Du Wei Hu Zhiyuan Li Ruslan Salakhutdinov and Ruosong Wang. 2019. On exact computation with an infinitely wide neural net. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_1_5_1","volume-title":"Linear discriminant analysis-a brief tutorial","author":"Balakrishnama Suresh","year":"1998","unstructured":"Suresh Balakrishnama and Aravind Ganapathiraju. 1998. Linear discriminant analysis-a brief tutorial. Institute for Signal and information Processing 18, 1998(1998), 1\u20138."},{"key":"e_1_3_2_1_6_1","first-page":"23","article-title":"From ranknet to lambdarank to lambdamart: An overview","volume":"11","author":"Burges JC","year":"2010","unstructured":"Christopher\u00a0JC Burges. 2010. From ranknet to lambdarank to lambdamart: An overview. Learning 11, 23-581 (2010), 81.","journal-title":"Learning"},{"key":"e_1_3_2_1_7_1","unstructured":"Yuan Cao and Quanquan Gu. 2019. Generalization Bounds of Stochastic Gradient Descent for Wide and Deep Neural Networks. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_1_8_1","volume-title":"Generalization Error Bounds of Gradient Descent for Learning Over-parameterized Deep ReLU Networks. In the Thirty-Fourth AAAI Conference on Artificial Intelligence.","author":"Cao Yuan","year":"2020","unstructured":"Yuan Cao and Quanquan Gu. 2020. Generalization Error Bounds of Gradient Descent for Learning Over-parameterized Deep ReLU Networks. In the Thirty-Fourth AAAI Conference on Artificial Intelligence."},{"key":"e_1_3_2_1_9_1","volume-title":"Proceedings of the Learning to Rank Challenge. 1\u201324","author":"Chapelle Olivier","year":"2011","unstructured":"Olivier Chapelle and Yi Chang. 2011. Yahoo! learning to rank challenge overview. In Proceedings of the Learning to Rank Challenge. 1\u201324."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2094072.2094078"},{"key":"e_1_3_2_1_11_1","unstructured":"Zixiang Chen Yuan Cao Difan Zou and Quanquan Gu. 2019. How Much Over-parameterization Is Sufficient to Learn Deep ReLU Networks?arXiv preprint arXiv:1911.12360(2019)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/1341531.1341545"},{"key":"e_1_3_2_1_13_1","unstructured":"Amit Daniely. 2017. SGD learns the conjugate kernel class of the network. In Advances in Neural Information Processing Systems. 2422\u20132430."},{"key":"e_1_3_2_1_14_1","volume-title":"Gradient Descent Provably Optimizes Over-parameterized Neural Networks. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=S1eK3i09YQ","author":"Du S.","year":"2019","unstructured":"Simon\u00a0S. Du, Xiyu Zhai, Barnabas Poczos, and Aarti Singh. 2019. Gradient Descent Provably Optimizes Over-parameterized Neural Networks. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=S1eK3i09YQ"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/1008992.1009004"},{"key":"e_1_3_2_1_16_1","unstructured":"Sarah Filippi Olivier Cappe Aur\u00e9lien Garivier and Csaba Szepesv\u00e1ri. 2010. Parametric bandits: The generalized linear case. In Advances in Neural Information Processing Systems. 586\u2013594."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/1526709.1526712"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/1498759.1498818"},{"key":"e_1_3_2_1_19_1","unstructured":"Boris Hanin and Mark Sellke. 2017. Approximating Continuous Functions by ReLU Nets of Minimal Width. arXiv preprint arXiv:1710.11278(2017)."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/2396761.2398516"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10791-012-9197-9"},{"key":"e_1_3_2_1_22_1","unstructured":"Arthur Jacot Franck Gabriel and Cl\u00e9ment Hongler. 2018. Neural tangent kernel: Convergence and generalization in neural networks. In Advances in neural information processing systems. 8571\u20138580."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3449972"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/1076034.1076063"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/1229179.1229181"},{"key":"e_1_3_2_1_26_1","unstructured":"Branislav Kveton Csaba Szepesvari Zheng Wen and Azin Ashkan. 2015. Cascading bandits: Learning to rank in the cascade model. In ICML. 767\u2013776."},{"key":"e_1_3_2_1_27_1","unstructured":"Branislav Kveton Zheng Wen Azin Ashkan and Csaba Szepesvari. 2015. Combinatorial cascading bandits. In NIPS. 1450\u20131458."},{"key":"e_1_3_2_1_28_1","unstructured":"Branislav Kveton Zheng Wen Azin Ashkan and Csaba Szepesvari. 2015. Tight regret bounds for stochastic combinatorial semi-bandits. In Artificial Intelligence and Statistics. 535\u2013543."},{"key":"e_1_3_2_1_29_1","volume-title":"Toprank: A practical algorithm for online stochastic ranking. In NIPS. 3945\u20133954.","author":"Lattimore Tor","year":"2018","unstructured":"Tor Lattimore, Branislav Kveton, Shuai Li, and Csaba Szepesvari. 2018. Toprank: A practical algorithm for online stochastic ranking. In NIPS. 3945\u20133954."},{"key":"e_1_3_2_1_30_1","unstructured":"Shuai Li Tor Lattimore and Csaba Szepesv\u00e1ri. 2018. Online learning to rank with features. arXiv preprint arXiv:1810.02567(2018)."},{"key":"e_1_3_2_1_31_1","unstructured":"Shuai Li Baoxiang Wang Shengyu Zhang and Wei Chen. 2016. Contextual Combinatorial Cascading Bandits.. In ICML Vol.\u00a016. 1245\u20131253."},{"key":"e_1_3_2_1_32_1","unstructured":"Shiyu Liang and R Srikant. 2016. Why deep neural networks for function approximation?arXiv preprint arXiv:1610.04161(2016)."},{"key":"e_1_3_2_1_33_1","unstructured":"Haihao Lu and Kenji Kawaguchi. 2017. Depth Creates No Bad Local Minima. arXiv preprint arXiv:1702.08580(2017)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3132847.3132896"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3269206.3271686"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-30671-1_50"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330677"},{"key":"e_1_3_2_1_38_1","unstructured":"Tao Qin and Tie-Yan Liu. 2013. Introducing LETOR 4.0 Datasets. arxiv:1306.2597\u00a0[cs.IR]"},{"key":"e_1_3_2_1_39_1","first-page":"193","article-title":"Learning to rank with nonsmooth cost functions","volume":"19","author":"Quoc C","year":"2007","unstructured":"C Quoc and Viet Le. 2007. Learning to rank with nonsmooth cost functions. Proceedings of the Advances in Neural Information Processing Systems 19(2007), 193\u2013200.","journal-title":"Proceedings of the Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"crossref","unstructured":"Filip Radlinski Robert Kleinberg and Thorsten Joachims. 2008. Learning diverse rankings with multi-armed bandits. In ICML. 784\u2013791.","DOI":"10.1145\/1390156.1390255"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/2835776.2835804"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661829.2661952"},{"key":"e_1_3_2_1_43_1","unstructured":"Matus Telgarsky. 2015. Representation benefits of deep feedforward networks. arXiv preprint arXiv:1509.08101(2015)."},{"key":"e_1_3_2_1_44_1","unstructured":"Matus Telgarsky. 2016. Benefits of depth in neural networks. arXiv preprint arXiv:1602.04485(2016)."},{"key":"e_1_3_2_1_45_1","unstructured":"Roman Vershynin. 2010. Introduction to the non-asymptotic analysis of random matrices. arXiv preprint arXiv:1011.3027(2010)."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3331184.3331264"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210045"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"crossref","unstructured":"Xuanhui Wang Cheng Li Nadav Golbandi Michael Bendersky and Marc Najork. 2018. The LambdaLoss Framework for Ranking Metric Optimization. In CIKM \u201918. ACM 1313\u20131322.","DOI":"10.1145\/3269206.3271784"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2017.07.002"},{"key":"e_1_3_2_1_50_1","unstructured":"Dmitry Yarotsky. 2018. Optimal approximation of continuous functions by very deep ReLU networks. arXiv preprint arXiv:1802.03620(2018)."},{"key":"e_1_3_2_1_51_1","unstructured":"Yisong Yue and Thorsten Joachims. 2009. Interactively optimizing information retrieval systems as a dueling bandits problem. In ICML. 1201\u20131208."},{"key":"e_1_3_2_1_52_1","unstructured":"Weitong Zhang Dongruo Zhou Lihong Li and Quanquan Gu. 2020. Neural Thompson Sampling. arXiv preprint arXiv:2010.00827(2020)."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/2983323.2983774"},{"key":"e_1_3_2_1_54_1","volume-title":"International Conference on Machine Learning. PMLR, 11492\u201311502","author":"Zhou Dongruo","year":"2020","unstructured":"Dongruo Zhou, Lihong Li, and Quanquan Gu. 2020. Neural contextual bandits with UCB-based exploration. In International Conference on Machine Learning. PMLR, 11492\u201311502."},{"key":"e_1_3_2_1_55_1","volume-title":"ICML","author":"Zoghi Masrour","year":"2017","unstructured":"Masrour Zoghi, Tomas Tunys, Mohammad Ghavamzadeh, Branislav Kveton, Csaba Szepesvari, and Zheng Wen. 2017. Online learning to rank in stochastic click models. In ICML 2017. 4199\u20134208."},{"key":"e_1_3_2_1_56_1","volume-title":"Stochastic gradient descent optimizes over-parameterized deep ReLU networks. Machine Learning","author":"Zou Difan","year":"2019","unstructured":"Difan Zou, Yuan Cao, Dongruo Zhou, and Quanquan Gu. 2019. Stochastic gradient descent optimizes over-parameterized deep ReLU networks. Machine Learning (2019)."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-019-05839-6"},{"key":"e_1_3_2_1_58_1","unstructured":"Difan Zou and Quanquan Gu. 2019. An Improved Analysis of Training Over-parameterized Deep Neural Networks. In Advances in Neural Information Processing Systems."}],"event":{"name":"WWW '22: The ACM Web Conference 2022","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Virtual Event, Lyon France","acronym":"WWW '22"},"container-title":["Proceedings of the ACM Web Conference 2022"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3485447.3512250","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3485447.3512250","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3485447.3512250","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:30:14Z","timestamp":1750188614000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3485447.3512250"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,25]]},"references-count":58,"alternative-id":["10.1145\/3485447.3512250","10.1145\/3485447"],"URL":"https:\/\/doi.org\/10.1145\/3485447.3512250","relation":{},"subject":[],"published":{"date-parts":[[2022,4,25]]},"assertion":[{"value":"2022-04-25","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}