{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,27]],"date-time":"2025-11-27T13:58:15Z","timestamp":1764251895229,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":50,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,8,2]],"date-time":"2024-08-02T00:00:00Z","timestamp":1722556800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"NSF","award":["IIS-2225942"],"award-info":[{"award-number":["IIS-2225942"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,8,2]]},"DOI":"10.1145\/3664190.3672522","type":"proceedings-article","created":{"date-parts":[[2024,8,5]],"date-time":"2024-08-05T12:39:41Z","timestamp":1722861581000},"page":"81-90","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["On Adaptive Knowledge Distillation with Generalized KL-Divergence Loss for Ranking Model Refinement"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6454-5796","authenticated-orcid":false,"given":"Yingrui","family":"Yang","sequence":"first","affiliation":[{"name":"Department of Computer Science, University of California, Santa Barbara, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-8581-6733","authenticated-orcid":false,"given":"Shanxiu","family":"He","sequence":"additional","affiliation":[{"name":"Department of Computer Science, University of California, Santa Barbara, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1902-3387","authenticated-orcid":false,"given":"Tao","family":"Yang","sequence":"additional","affiliation":[{"name":"Department of Computer Science, University of California, Santa Barbara, CA, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,8,5]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"23","article-title":"From ranknet to lambdarank to lambdamart: An overview","volume":"11","author":"Burges Christopher JC","year":"2010","unstructured":"Christopher JC Burges. 2010. From ranknet to lambdarank to lambdamart: An overview. Learning, Vol. 11, 23--581 (2010), 81.","journal-title":"Learning"},{"key":"e_1_3_2_1_2_1","volume-title":"MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. ArXiv","author":"Campos Daniel Fernando","year":"2016","unstructured":"Daniel Fernando Campos, Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, Li Deng, and Bhaskar Mitra. 2016. MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. ArXiv, Vol. abs\/1611.09268 (2016)."},{"key":"e_1_3_2_1_3_1","unstructured":"SimLM Checkpoint. [n. d.]. https:\/\/huggingface.co\/intfloat\/simlm-base-msmarco-finetuned. ( [n. d.])."},{"key":"e_1_3_2_1_4_1","unstructured":"Co-Condenser. 2021. https:\/\/huggingface.co\/Luyu\/co-condenser-marco. (2021)."},{"key":"e_1_3_2_1_5_1","volume-title":"Reciprocal rank fusion outperforms condorcet and individual rank learning methods. SIGIR","author":"Cormack Gordon V.","year":"2009","unstructured":"Gordon V. Cormack, Charles L. A. Clarke, and Stefan B\u00fcttcher. 2009. Reciprocal rank fusion outperforms condorcet and individual rank learning methods. SIGIR (2009)."},{"key":"e_1_3_2_1_6_1","volume-title":"Overview of the TREC 2020 Deep Learning Track. ArXiv","volume":"2102","author":"Craswell Nick","year":"2020","unstructured":"Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Fernando Campos, and Ellen M. Voorhees. 2020. Overview of the TREC 2020 Deep Learning Track. ArXiv, Vol. abs\/2102.07662 (2020)."},{"key":"e_1_3_2_1_7_1","volume-title":"Context-Aware Term Weighting For First Stage Passage Retrieval. SIGIR","author":"Dai Zhuyun","year":"2020","unstructured":"Zhuyun Dai and Jamie Callan. 2020. Context-Aware Term Weighting For First Stage Passage Retrieval. SIGIR (2020)."},{"key":"e_1_3_2_1_8_1","volume-title":"SPLADE v2: Sparse Lexical and Expansion Model for Information Retrieval. ArXiv","author":"Formal Thibault","year":"2021","unstructured":"Thibault Formal, C. Lassance, Benjamin Piwowarski, and St\u00e9phane Clinchant. 2021. SPLADE v2: Sparse Lexical and Expansion Model for Information Retrieval. ArXiv, Vol. abs\/2109.10086 (2021)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531857"},{"key":"e_1_3_2_1_10_1","volume-title":"SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking. SIGIR","author":"Formal Thibault","year":"2021","unstructured":"Thibault Formal, Benjamin Piwowarski, and St\u00e9phane Clinchant. 2021. SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking. SIGIR (2021)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.75"},{"key":"e_1_3_2_1_12_1","volume-title":"Proceedings of SIGIR","author":"Gao Luyu","year":"2020","unstructured":"Luyu Gao, Zhuyun Dai, and J. Callan. 2020. Understanding BERT Rankers Under Distillation. Proceedings of SIGIR (2020)."},{"key":"e_1_3_2_1_13_1","volume-title":"COIL: Revisit Exact Lexical Match in Information Retrieval with Contextualized Inverted List. NAACL","author":"Gao Luyu","year":"2021","unstructured":"Luyu Gao, Zhuyun Dai, and Jamie Callan. 2021. COIL: Revisit Exact Lexical Match in Information Retrieval with Contextualized Inverted List. NAACL (2021)."},{"key":"e_1_3_2_1_14_1","unstructured":"MS-MARCO Hard-Negatives. 2022. https:\/\/huggingface.co\/datasets\/sentence-transformers\/msmarco-hard-negatives. (2022)."},{"key":"e_1_3_2_1_15_1","volume-title":"Distilling the Knowledge in a Neural Network. ArXiv","author":"Hinton Geoffrey E.","year":"2015","unstructured":"Geoffrey E. Hinton, Oriol Vinyals, and Jeffrey Dean. 2015. Distilling the Knowledge in a Neural Network. ArXiv, Vol. abs\/1503.02531 (2015)."},{"key":"e_1_3_2_1_16_1","volume-title":"Improving Efficient Neural Ranking Models with Cross-Architecture Knowledge Distillation. ArXiv","author":"Hofst\u00e4tter Sebastian","year":"2020","unstructured":"Sebastian Hofst\u00e4tter, Sophia Althammer, Michael Schr\u00f6der, Mete Sertkan, and Allan Hanbury. 2020. Improving Efficient Neural Ranking Models with Cross-Architecture Knowledge Distillation. ArXiv, Vol. abs\/2010.02666 (2020)."},{"key":"e_1_3_2_1_17_1","volume-title":"Efficiently Teaching an Effective Dense Retriever with Balanced Topic Aware Sampling. SIGIR","author":"Hofst\u00e4tter Sebastian","year":"2021","unstructured":"Sebastian Hofst\u00e4tter, Sheng-Chieh Lin, Jheng-Hong Yang, Jimmy J. Lin, and Allan Hanbury. 2021. Efficiently Teaching an Effective Dense Retriever with Balanced Topic Aware Sampling. SIGIR (2021)."},{"key":"e_1_3_2_1_18_1","volume-title":"Zaharia","author":"Khattab O.","year":"2020","unstructured":"O. Khattab and Matei A. Zaharia. 2020. ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT. SIGIR (2020)."},{"key":"e_1_3_2_1_19_1","volume-title":"An Efficiency Study for SPLADE Models. SIGIR","author":"Lassance Carlos","year":"2022","unstructured":"Carlos Lassance and St\u00e9phane Clinchant. 2022. An Efficiency Study for SPLADE Models. SIGIR (2022)."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3592071"},{"key":"e_1_3_2_1_21_1","volume-title":"Interpolate: PRF, Dense and Sparse Retrievers. SIGIR","author":"Li Hang","year":"2022","unstructured":"Hang Li, Shuai Wang, Shengyao Zhuang, Ahmed Mourad, Xueguang Ma, Jimmy Lin, and G. Zuccon. 2022. To Interpolate or not to Interpolate: PRF, Dense and Sparse Retrievers. SIGIR (2022)."},{"key":"e_1_3_2_1_22_1","volume-title":"SLIM: Sparsified Late Interaction for Multi-Vector Retrieval with Inverted Indexes. In SIGIR. showeprint[arxiv and sigir 2023]2302.06587 [cs.IR]","author":"Li Minghan","year":"2023","unstructured":"Minghan Li, Sheng-Chieh Lin, Xueguang Ma, and Jimmy Lin. 2023. SLIM: Sparsified Late Interaction for Multi-Vector Retrieval with Inverted Indexes. In SIGIR. showeprint[arxiv and sigir 2023]2302.06587 [cs.IR]"},{"key":"e_1_3_2_1_23_1","volume-title":"Wen tau Yih, and Xilun Chen","author":"Li Minghan","year":"2022","unstructured":"Minghan Li, Sheng-Chieh Lin, Barlas Oguz, Asish Ghoshal, Jimmy Lin, Yashar Mehdad, Wen tau Yih, and Xilun Chen. 2022. CITADEL: Conditional Token Interaction via Dynamic Lexical Routing for Efficient and Effective Multi-Vector Retrieval. arXiv 2211.01267 (2022). arxiv: 2211.10411 [cs.IR]"},{"key":"e_1_3_2_1_24_1","volume-title":"Lin and Xueguang Ma","author":"Jimmy","year":"2021","unstructured":"Jimmy J. Lin and Xueguang Ma. 2021. A Few Brief Notes on DeepImpact, COIL, and a Conceptual Framework for Information Retrieval Techniques. ArXiv, Vol. abs\/2106.14807 (2021)."},{"key":"e_1_3_2_1_25_1","volume-title":"Lin","author":"Lin Sheng-Chieh","year":"2021","unstructured":"Sheng-Chieh Lin, Jheng-Hong Yang, and Jimmy J. Lin. 2021. In-Batch Negatives for Knowledge Distillation with Tightly-Coupled Teachers for Dense Retrieval. In REPL4NLP."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2858826"},{"key":"e_1_3_2_1_27_1","volume-title":"Foundations and Trends\u00ae in Information Retrieval","volume":"3","author":"Liu Tie-Yan","year":"2009","unstructured":"Tie-Yan Liu. 2009. Learning to Rank for Information Retrieval. Foundations and Trends\u00ae in Information Retrieval, Vol. 3, 3 (2009), 225--331."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Yi Luan Jacob Eisenstein Kristina Toutanova and M. Collins. 2021. Sparse Dense and Attentional Representations for Text Retrieval. TACL (2021).","DOI":"10.1162\/tacl_a_00369"},{"key":"e_1_3_2_1_29_1","volume-title":"Learning Passage Impacts for Inverted Indexes. SIGIR","author":"Mallia Antonio","year":"2021","unstructured":"Antonio Mallia, O. Khattab, Nicola Tonellotto, and Torsten Suel. 2021. Learning Passage Impacts for Inverted Indexes. SIGIR (2021)."},{"key":"e_1_3_2_1_30_1","volume-title":"Proceedings of the Open-Source IR Replicability Challenge","author":"Mallia Antonio","year":"2019","unstructured":"Antonio Mallia, Michal Siedlaczek, Joel Mackenzie, and Torsten Suel. 2019. PISA: Performant indexes and search for academia. Proceedings of the Open-Source IR Replicability Challenge (2019)."},{"key":"e_1_3_2_1_31_1","unstructured":"MiniLM-L-6-v2. 2022. https:\/\/huggingface.co\/cross-encoder\/ms-marco-MiniLM-L-6-v2. (2022)."},{"key":"e_1_3_2_1_32_1","volume-title":"Zhuyun Dai, Siddhartha Brahma, Iftekhar Naim, Tao Lei, and Vincent Y. Zhao.","author":"Qian Yujie","year":"2022","unstructured":"Yujie Qian, Jinhyuk Lee, Sai Meher Karthik Duddu, Zhuyun Dai, Siddhartha Brahma, Iftekhar Naim, Tao Lei, and Vincent Y. Zhao. 2022. Multi-Vector Retrieval as Sparse Alignment. arXiv 2211.01267 (2022). arxiv: 2211.01267 [cs.CL]"},{"key":"e_1_3_2_1_33_1","volume-title":"Representation Sparsification with Hybrid Thresholding for Fast SPLADE-based Document Retrieval. ACM SIGIR'23","author":"Qiao Yifan","year":"2023","unstructured":"Yifan Qiao, Yingrui Yang,, Shanxiu He, and Tao Yang. 2023. Representation Sparsification with Hybrid Thresholding for Fast SPLADE-based Document Retrieval. ACM SIGIR'23 (2023)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.224"},{"key":"e_1_3_2_1_35_1","volume-title":"NAACL'22","volume":"2112","author":"Santhanam Keshav","year":"2022","unstructured":"Keshav Santhanam, O. Khattab, Jon Saad-Falcon, Christopher Potts, and Matei A. Zaharia. 2022. ColBERTv2: Effective and Efficient Retrieval via Lightweight Late Interaction. NAACL'22, Vol. ArXiv abs\/2112.01488 (2022)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.496"},{"key":"e_1_3_2_1_37_1","unstructured":"Sentence-Transformers. 2022. https:\/\/www.sbert.net\/. (2022)."},{"key":"e_1_3_2_1_38_1","volume-title":"LexMAE: Lexicon-Bottlenecked Pretraining for Large-Scale Retrieval. In The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=PfpEtB3-csK","author":"Shen Tao","year":"2023","unstructured":"Tao Shen, Xiubo Geng, Chongyang Tao, Can Xu, Xiaolong Huang, Binxing Jiao, Linjun Yang, and Daxin Jiang. 2023. LexMAE: Lexicon-Bottlenecked Pretraining for Large-Scale Retrieval. In The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=PfpEtB3-csK"},{"key":"e_1_3_2_1_39_1","volume-title":"UnifieR: A Unified Retriever for Large-Scale Retrieval. ArXiv","author":"Shen Tao","year":"2022","unstructured":"Tao Shen, Xiubo Geng, Chongyang Tao, Can Xu, Kai Zhang, and Daxin Jiang. 2022. UnifieR: A Unified Retriever for Large-Scale Retrieval. ArXiv, Vol. abs\/2205.11194 (2022)."},{"key":"e_1_3_2_1_40_1","unstructured":"SPLADE. 2022. https:\/\/github.com\/naver\/splade. (2022)."},{"key":"e_1_3_2_1_41_1","volume-title":"BEIR: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models. In NeurIPS.","author":"Thakur Nandan","year":"2021","unstructured":"Nandan Thakur, Nils Reimers, Andreas R\u00fcckl\u00e9, Abhishek Srivastava, and Iryna Gurevych. 2021. BEIR: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models. In NeurIPS."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i8.20826"},{"key":"e_1_3_2_1_43_1","volume-title":"SimLM: Pre-training with Representation Bottleneck for Dense Passage Retrieval. ACL","author":"Wang Liang","year":"2023","unstructured":"Liang Wang, Nan Yang, Xiaolong Huang, Binxing Jiao, Linjun Yang, Daxin Jiang, Rangan Majumder, and Furu Wei. 2023. SimLM: Pre-training with Representation Bottleneck for Dense Passage Retrieval. ACL (2023)."},{"key":"e_1_3_2_1_44_1","volume-title":"RetroMAE: Pre-training Retrieval-oriented Transformers via Masked Auto-Encoder. EMNLP","author":"Xiao Shitao","year":"2022","unstructured":"Shitao Xiao, Zheng Liu, Yingxia Shao, and Zhao Cao. 2022. RetroMAE: Pre-training Retrieval-oriented Transformers via Masked Auto-Encoder. EMNLP (2022)."},{"key":"e_1_3_2_1_45_1","unstructured":"Lee Xiong Chenyan Xiong Ye Li Kwok-Fung Tang Jialin Liu Paul N. Bennett Junaid Ahmed and Arnold Overwijk. 2021. Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval. In ICLR."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3578337.3605120"},{"key":"e_1_3_2_1_47_1","volume-title":"Weighted KL-Divergence for Document Ranking Model Refinement. ArXiv and SIGIR","author":"Yang Yingrui","year":"2024","unstructured":"Yingrui Yang, Yifan Qiao, Shanxiu He, and Tao Yang. 2024. Weighted KL-Divergence for Document Ranking Model Refinement. ArXiv and SIGIR 2024. arxiv: 2406.05977"},{"key":"e_1_3_2_1_48_1","volume-title":"Curriculum Learning for Dense Retrieval Distillation. SIGIR","author":"Zeng Hansi","year":"2022","unstructured":"Hansi Zeng, Hamed Zamani, and Vishwa Vinay. 2022. Curriculum Learning for Dense Retrieval Distillation. SIGIR (2022)."},{"key":"e_1_3_2_1_49_1","volume-title":"Optimizing Dense Retrieval Model Training with Hard Negatives. CoRR","author":"Zhan Jingtao","year":"2021","unstructured":"Jingtao Zhan, Jiaxin Mao, Yiqun Liu, Jiafeng Guo, Min Zhang, and Shaoping Ma. 2021. Optimizing Dense Retrieval Model Training with Hard Negatives. CoRR, Vol. abs\/2104.08051 (2021). https:\/\/arxiv.org\/abs\/2104.08051"},{"key":"e_1_3_2_1_50_1","volume-title":"Adversarial Retriever-Ranker for dense text retrieval. ICLR","author":"Zhang Hang","year":"2022","unstructured":"Hang Zhang, Yeyun Gong, Yelong Shen, Jiancheng Lv, Nan Duan, and Weizhu Chen. 2022. Adversarial Retriever-Ranker for dense text retrieval. ICLR (2022)."}],"event":{"name":"ICTIR '24: The 2024 ACM SIGIR International Conference on the Theory of Information Retrieval","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"],"location":"Washington DC USA","acronym":"ICTIR '24"},"container-title":["Proceedings of the 2024 ACM SIGIR International Conference on Theory of Information Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664190.3672522","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3664190.3672522","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T23:59:23Z","timestamp":1755907163000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664190.3672522"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,2]]},"references-count":50,"alternative-id":["10.1145\/3664190.3672522","10.1145\/3664190"],"URL":"https:\/\/doi.org\/10.1145\/3664190.3672522","relation":{},"subject":[],"published":{"date-parts":[[2024,8,2]]},"assertion":[{"value":"2024-08-05","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}