{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,22]],"date-time":"2025-11-22T18:29:23Z","timestamp":1763836163205,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":34,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,9,13]],"date-time":"2022-09-13T00:00:00Z","timestamp":1663027200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,9,18]]},"DOI":"10.1145\/3523227.3546766","type":"proceedings-article","created":{"date-parts":[[2022,9,13]],"date-time":"2022-09-13T14:13:46Z","timestamp":1663078426000},"page":"134-144","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":7,"title":["Identifying New Podcasts with High General Appeal Using a Pure Exploration Infinitely-Armed Bandit Strategy"],"prefix":"10.1145","author":[{"given":"Maryam","family":"Aziz","sequence":"first","affiliation":[{"name":"Spotify, United States"}]},{"given":"Jesse","family":"Anderton","sequence":"additional","affiliation":[{"name":"Spotify, United States"}]},{"given":"Kevin","family":"Jamieson","sequence":"additional","affiliation":[{"name":"University of Washington, United States"}]},{"given":"Alice","family":"Wang","sequence":"additional","affiliation":[{"name":"Spotify, United States"}]},{"given":"Hugues","family":"Bouchard","sequence":"additional","affiliation":[{"name":"Spotify, Spain"}]},{"given":"Javed","family":"Aslam","sequence":"additional","affiliation":[{"name":"Northeastern University, United States"}]}],"member":"320","published-online":{"date-parts":[[2022,9,13]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"COLT-23th Conference on Learning Theory-2010","author":"Audibert Jean-Yves","year":"2010","unstructured":"Jean-Yves Audibert and S\u00e9bastien Bubeck. 2010. Best arm identification in multi-armed bandits. In COLT-23th Conference on Learning Theory-2010. 13\u2013p."},{"key":"e_1_3_2_2_2_1","volume-title":"Pure Exploration in Infinitely-Armed Bandit Models with Fixed-Confidence. In ALT 2018-Algorithmic Learning Theory.","author":"Aziz Maryam","year":"2018","unstructured":"Maryam Aziz, Jesse Anderton, Emilie Kaufmann, and Javed Aslam. 2018. Pure Exploration in Infinitely-Armed Bandit Models with Fixed-Confidence. In ALT 2018-Algorithmic Learning Theory."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1177\/1354856517736979"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1214\/aos\/1069362389"},{"key":"e_1_3_2_2_5_1","unstructured":"Thomas Bonald and Alexandre Proutiere. 2013. Two-target algorithms for infinite-armed bandits with Bernoulli rewards. In Advances in Neural Information Processing Systems. 2184\u20132192."},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"crossref","unstructured":"S\u00e9bastien Bubeck Nicolo Cesa-Bianchi 2012. Regret analysis of stochastic and nonstochastic multi-armed bandit problems. Foundations and Trends\u00ae in Machine Learning 5 1(2012) 1\u2013122.","DOI":"10.1561\/2200000024"},{"key":"e_1_3_2_2_7_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-04414-4_7"},{"key":"e_1_3_2_2_8_1","unstructured":"Giuseppe Burtini Jason Loeppky and Ramon Lawrence. 2015. A survey of online experiment design with the stochastic multi-armed bandit. arXiv preprint arXiv:1510.00757(2015)."},{"key":"e_1_3_2_2_9_1","unstructured":"Alexandra Carpentier and Michal Valko. 2015. Simple regret for infinitely many armed bandits. CoRR abs\/1505.04627(2015)."},{"key":"e_1_3_2_2_10_1","unstructured":"Hock\u00a0Peng Chan and Shouri Hu. 2018. Infinite Arms Bandit: Optimality via Confidence Bounds. CoRR abs\/1805.11793(2018)."},{"key":"e_1_3_2_2_11_1","volume-title":"Conference on Learning Theory. 394\u2013407","author":"Chandrasekaran Karthekeyan","year":"2014","unstructured":"Karthekeyan Chandrasekaran and Richard Karp. 2014. Finding a most biased coin with fewest flips. In Conference on Learning Theory. 394\u2013407."},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"crossref","unstructured":"Arghya\u00a0Roy Chaudhuri and Shivaram Kalyanakrishnan. 2017. PAC Identification of a Bandit Arm Relative to a Reward Quantile.. In AAAI. 1777\u20131783.","DOI":"10.1609\/aaai.v31i1.10802"},{"key":"e_1_3_2_2_13_1","unstructured":"Arghya\u00a0Roy Chaudhuri and Shivaram Kalyanakrishnan. 2018. Quantile-Regret Minimisation in Infinitely Many-Armed Bandits.. In UAI. 425\u2013434."},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-662-44848-9_20"},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-23528-8_29"},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/1935826.1935925"},{"key":"e_1_3_2_2_17_1","volume-title":"Reconciling the Quality vs Popularity Dichotomy in Online Cultural Markets. ACM Transactions on Information Systems (TOIS)","author":"Gaeta Rossano","year":"2022","unstructured":"Rossano Gaeta, Michele Garetto, Giancarlo Ruffo, and Alessandro Flammini. 2022. Reconciling the Quality vs Popularity Dichotomy in Online Cultural Markets. ACM Transactions on Information Systems (TOIS) (2022)."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"crossref","unstructured":"F\u00a0Maxwell Harper and Joseph\u00a0A Konstan. 2015. The movielens datasets: History and context. Acm transactions on interactive intelligent systems (tiis) 5 4(2015) 1\u201319.","DOI":"10.1145\/2827872"},{"key":"e_1_3_2_2_19_1","unstructured":"Kevin\u00a0G Jamieson Daniel Haas and Benjamin Recht. 2016. The power of adaptivity in identifying statistical alternatives. In Advances in Neural Information Processing Systems. 775\u2013783."},{"key":"e_1_3_2_2_20_1","volume-title":"Next: A system for real-world development, evaluation, and application of active learning. Advances in neural information processing systems 28","author":"Jamieson G","year":"2015","unstructured":"Kevin\u00a0G Jamieson, Lalit Jain, Chris Fernandez, Nicholas\u00a0J Glattard, and Rob Nowak. 2015. Next: A system for real-world development, evaluation, and application of active learning. Advances in neural information processing systems 28 (2015)."},{"key":"e_1_3_2_2_21_1","unstructured":"Kevin\u00a0G. Jamieson Matthew Malloy Robert\u00a0D. Nowak and S\u00e9bastien Bubeck. 2014. lil\u2019 UCB : An Optimal Exploration Algorithm for Multi-Armed Bandits. In COLT."},{"key":"e_1_3_2_2_22_1","volume-title":"Proceedings of the 30th International Conference on Machine Learning (ICML-13)","author":"Karnin Zohar","year":"2013","unstructured":"Zohar Karnin, Tomer Koren, and Oren Somekh. 2013. Almost Optimal Exploration in Multi-Armed Bandits. In Proceedings of the 30th International Conference on Machine Learning (ICML-13), Vol.\u00a028. 1238\u20131246."},{"key":"e_1_3_2_2_23_1","volume-title":"Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization.Journal of Machine Learning Research 18","author":"Li Lisha","year":"2017","unstructured":"Lisha Li, Kevin\u00a0G Jamieson, Giulia DeSalvo, Afshin Rostamizadeh, and Ameet Talwalkar. 2017. Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization.Journal of Machine Learning Research 18 (2017), 185\u20131."},{"key":"e_1_3_2_2_24_1","volume-title":"Article arXiv:1902.08593 (Feb.","author":"Liu Larkin","year":"2019","unstructured":"Larkin Liu, Richard Downe, and Joshua Reid. 2019. Multi-Armed Bandit Strategies for Non-Stationary Reward Distributions and Delayed Feedback Processes. arXiv e-prints, Article arXiv:1902.08593 (Feb. 2019), arXiv:1902.08593\u00a0pages. arxiv:1902.08593\u00a0[cs.LG]"},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2016.07.043"},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CISS.2012.6310773"},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.5555\/1953048.2078195"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAI.2021.3117743"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2433396.2433443"},{"key":"e_1_3_2_2_30_1","unstructured":"Wenbo Ren Jia Liu and Ness Shroff. 2018. Exploring k out of Top \u03c1 Fraction of Arms in Stochastic Bandits. arXiv preprint arXiv:1810.11857(2018)."},{"key":"e_1_3_2_2_31_1","volume-title":"Experimental study of inequality and unpredictability in an artificial cultural market. science 311, 5762","author":"Salganik J","year":"2006","unstructured":"Matthew\u00a0J Salganik, Peter\u00a0Sheridan Dodds, and Duncan\u00a0J Watts. 2006. Experimental study of inequality and unpredictability in an artificial cultural market. science 311, 5762 (2006), 854\u2013856."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/1787234.1787254"},{"key":"e_1_3_2_2_33_1","unstructured":"Olivier Teytaud Sylvain Gelly and Michele Sebag. 2007. Anytime many-armed bandits. In CAP07."},{"key":"e_1_3_2_2_34_1","volume-title":"Algorithms for infinitely many-armed bandits. Advances in Neural Information Processing Systems 21","author":"Wang Yizao","year":"2008","unstructured":"Yizao Wang, Jean-Yves Audibert, and R\u00e9mi Munos. 2008. Algorithms for infinitely many-armed bandits. Advances in Neural Information Processing Systems 21 (2008)."}],"event":{"name":"RecSys '22: Sixteenth ACM Conference on Recommender Systems","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web","SIGAI ACM Special Interest Group on Artificial Intelligence","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data","SIGIR ACM Special Interest Group on Information Retrieval","SIGCHI ACM Special Interest Group on Computer-Human Interaction"],"location":"Seattle WA USA","acronym":"RecSys '22"},"container-title":["Proceedings of the 16th ACM Conference on Recommender Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3523227.3546766","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3523227.3546766","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:30:45Z","timestamp":1750188645000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3523227.3546766"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,9,13]]},"references-count":34,"alternative-id":["10.1145\/3523227.3546766","10.1145\/3523227"],"URL":"https:\/\/doi.org\/10.1145\/3523227.3546766","relation":{},"subject":[],"published":{"date-parts":[[2022,9,13]]},"assertion":[{"value":"2022-09-13","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}