{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T19:33:14Z","timestamp":1775763194923,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":83,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,7,18]],"date-time":"2023-07-18T00:00:00Z","timestamp":1689638400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100011039","name":"Intelligence Advanced Research Projects Activity","doi-asserted-by":"publisher","award":["2019-19051600007"],"award-info":[{"award-number":["2019-19051600007"]}],"id":[{"id":"10.13039\/100011039","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,7,19]]},"DOI":"10.1145\/3539618.3591769","type":"proceedings-article","created":{"date-parts":[[2023,7,19]],"date-time":"2023-07-19T00:22:59Z","timestamp":1689726179000},"page":"1208-1218","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["Soft Prompt Decoding for Multilingual Dense Retrieval"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2939-1936","authenticated-orcid":false,"given":"Zhiqi","family":"Huang","sequence":"first","affiliation":[{"name":"University of Massachusetts Amherst, Amherst, MA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-2699-8460","authenticated-orcid":false,"given":"Hansi","family":"Zeng","sequence":"additional","affiliation":[{"name":"University of Massachusetts Amherst, Amherst, MA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0800-3340","authenticated-orcid":false,"given":"Hamed","family":"Zamani","sequence":"additional","affiliation":[{"name":"University of Massachusetts Amherst, Amherst, MA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0132-5694","authenticated-orcid":false,"given":"James","family":"Allan","sequence":"additional","affiliation":[{"name":"University of Massachusetts Amherst, Amherst, MA, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,7,18]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"Muppet: Massive Multi-task Representations with Pre-Finetuning. ArXiv","author":"Aghajanyan Armen","year":"2021","unstructured":"Armen Aghajanyan, Anchit Gupta, Akshat Shrivastava, Xilun Chen, Luke Zettlemoyer, and Sonal Gupta. 2021. Muppet: Massive Multi-task Representations with Pre-Finetuning. ArXiv, Vol. abs\/2101.11038 (2021)."},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1042"},{"key":"e_1_3_2_2_3_1","volume-title":"On the cross-lingual transferability of monolingual representations. arXiv preprint arXiv:1910.11856","author":"Artetxe Mikel","year":"2019","unstructured":"Mikel Artetxe, Sebastian Ruder, and Dani Yogatama. 2019. On the cross-lingual transferability of monolingual representations. arXiv preprint arXiv:1910.11856 (2019)."},{"key":"e_1_3_2_2_4_1","volume-title":"ATTEMPT: Parameter-Efficient Multi-task Tuning via Attentional Mixtures of Soft Prompts.","author":"Asai Akari","year":"2022","unstructured":"Akari Asai, Mohammadreza Salehi, Matthew E. Peters, and Hannaneh Hajishirzi. 2022. ATTEMPT: Parameter-Efficient Multi-task Tuning via Attentional Mixtures of Soft Prompts."},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401035"},{"key":"e_1_3_2_2_6_1","volume-title":"Israel Campiotti, Marzieh Fadaee, Roberto Lotufo, and Rodrigo Nogueira.","author":"Bonifacio Luiz","year":"2021","unstructured":"Luiz Bonifacio, Vitor Jeronymo, Hugo Queiroz Abonizio, Israel Campiotti, Marzieh Fadaee, Roberto Lotufo, and Rodrigo Nogueira. 2021. mmarco: A multilingual version of the ms marco passage ranking dataset. arXiv preprint arXiv:2108.13897 (2021)."},{"key":"e_1_3_2_2_7_1","volume-title":"CLEF 2002-Overview of results. In Workshop of the Cross-Language Evaluation Forum for European Languages. Springer, 9--27","author":"Braschler Martin","year":"2002","unstructured":"Martin Braschler. 2002. CLEF 2002-Overview of results. In Workshop of the Cross-Language Evaluation Forum for European Languages. Springer, 9--27."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICETET.2009.132"},{"key":"e_1_3_2_2_9_1","volume-title":"Unsupervised cross-lingual representation learning at scale. arXiv preprint arXiv:1911.02116","author":"Conneau Alexis","year":"2019","unstructured":"Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzm\u00e1n, Edouard Grave, Myle Ott, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Unsupervised cross-lingual representation learning at scale. arXiv preprint arXiv:1911.02116 (2019)."},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.747"},{"key":"e_1_3_2_2_11_1","volume-title":"Overview of the TREC 2019 deep learning track. arXiv preprint arXiv:2003","author":"Craswell Nick","year":"2020","unstructured":"Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, and Ellen M Voorhees. 2020. Overview of the TREC 2019 deep learning track. arXiv preprint arXiv:2003.07820 (2020)."},{"key":"e_1_3_2_2_12_1","volume-title":"Universal transformers. arXiv preprint arXiv:1807.03819","author":"Dehghani Mostafa","year":"2018","unstructured":"Mostafa Dehghani, Stephan Gouws, Oriol Vinyals, Jakob Uszkoreit, and \u0141ukasz Kaiser. 2018. Universal transformers. arXiv preprint arXiv:1807.03819 (2018)."},{"key":"e_1_3_2_2_13_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_2_14_1","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.480"},{"key":"e_1_3_2_2_16_1","volume-title":"Emiel Van Miltenburg, and Emiel Krahmer","author":"Ferreira Thiago Castro","year":"2019","unstructured":"Thiago Castro Ferreira, Chris van der Lee, Emiel Van Miltenburg, and Emiel Krahmer. 2019. Neural data-to-text generation: A comparison between pipeline and end-to-end architectures. arXiv preprint arXiv:1908.09022 (2019)."},{"key":"e_1_3_2_2_17_1","volume-title":"Cross-language information retrieval. arXiv preprint arXiv:2111.05988","author":"\u00e1kov\u00e1 Petra Galuvs","year":"2021","unstructured":"Petra Galuvs vc \u00e1kov\u00e1, Douglas W Oard, and Suraj Nair. 2021. Cross-language information retrieval. arXiv preprint arXiv:2111.05988 (2021)."},{"key":"e_1_3_2_2_18_1","volume-title":"Unsupervised corpus aware language model pre-training for dense passage retrieval. arXiv preprint arXiv:2108.05540","author":"Gao Luyu","year":"2021","unstructured":"Luyu Gao and Jamie Callan. 2021. Unsupervised corpus aware language model pre-training for dense passage retrieval. arXiv preprint arXiv:2108.05540 (2021)."},{"key":"e_1_3_2_2_19_1","volume-title":"Asian conference on machine learning. PMLR, 17--32","author":"Glasmachers Tobias","year":"2017","unstructured":"Tobias Glasmachers. 2017. Limits of end-to-end learning. In Asian conference on machine learning. PMLR, 17--32."},{"key":"e_1_3_2_2_20_1","volume-title":"Proceedings of the thirteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings, 249--256","author":"Glorot Xavier","year":"2010","unstructured":"Xavier Glorot and Yoshua Bengio. 2010. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings, 249--256."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-021-01453-z"},{"key":"e_1_3_2_2_22_1","volume-title":"Onur cC elebi, and Holger Schwenk","author":"Heffernan Kevin","year":"2022","unstructured":"Kevin Heffernan, Onur cC elebi, and Holger Schwenk. 2022. Bitext mining using distilled sentence representations for low-resource languages. arXiv preprint arXiv:2205.12654 (2022)."},{"key":"e_1_3_2_2_23_1","unstructured":"Geoffrey Hinton Oriol Vinyals Jeff Dean et al. 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 Vol. 2 7 (2015)."},{"key":"e_1_3_2_2_24_1","volume-title":"Improving efficient neural ranking models with cross-architecture knowledge distillation. arXiv preprint arXiv:2010.02666","author":"Hofst\u00e4tter Sebastian","year":"2020","unstructured":"Sebastian Hofst\u00e4tter, Sophia Althammer, Michael Schr\u00f6der, Mete Sertkan, and Allan Hanbury. 2020. Improving efficient neural ranking models with cross-architecture knowledge distillation. arXiv preprint arXiv:2010.02666 (2020)."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3482452"},{"key":"e_1_3_2_2_26_1","volume-title":"The 16th ACM International Conferenceon Web Search and Data Mining (WSDM)","author":"Huang Zhiqi","year":"2022","unstructured":"Zhiqi Huang, Puxuan Yu, and James Allan. 2022. Improving Cross-lingual Information Retrieval on Low-Resource Languages via Optimal Transport Distillation. In The 16th ACM International Conferenceon Web Search and Data Mining (WSDM), 2023."},{"key":"e_1_3_2_2_27_1","volume-title":"International conference on machine learning. PMLR, 4651--4664","author":"Jaegle Andrew","year":"2021","unstructured":"Andrew Jaegle, Felix Gimeno, Andy Brock, Oriol Vinyals, Andrew Zisserman, and Joao Carreira. 2021. Perceiver: General perception with iterative attention. In International conference on machine learning. PMLR, 4651--4664."},{"key":"e_1_3_2_2_28_1","volume-title":"Sewon Min, Patrick Lewis, Ledell Wu, Sergey Edunov, Danqi Chen, and Wen-tau Yih.","author":"Karpukhin Vladimir","year":"2020","unstructured":"Vladimir Karpukhin, Barlas Oug uz, Sewon Min, Patrick Lewis, Ledell Wu, Sergey Edunov, Danqi Chen, and Wen-tau Yih. 2020. Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906 (2020)."},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401075"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2004.06.007"},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-99736-6_24"},{"key":"e_1_3_2_2_32_1","volume-title":"Latent retrieval for weakly supervised open domain question answering. arXiv preprint arXiv:1906.00300","author":"Lee Kenton","year":"2019","unstructured":"Kenton Lee, Ming-Wei Chang, and Kristina Toutanova. 2019. Latent retrieval for weakly supervised open domain question answering. arXiv preprint arXiv:1906.00300 (2019)."},{"key":"e_1_3_2_2_33_1","volume-title":"The Power of Scale for Parameter-Efficient Prompt Tuning. ArXiv","author":"Lester Brian","year":"2021","unstructured":"Brian Lester, Rami Al-Rfou, and Noah Constant. 2021. The Power of Scale for Parameter-Efficient Prompt Tuning. ArXiv, Vol. abs\/2104.08691 (2021)."},{"key":"e_1_3_2_2_34_1","volume-title":"Ruty Rinott, Sebastian Riedel, and Holger Schwenk.","author":"Lewis Patrick","year":"2019","unstructured":"Patrick Lewis, Barlas Oug uz, Ruty Rinott, Sebastian Riedel, and Holger Schwenk. 2019. MLQA: Evaluating cross-lingual extractive question answering. arXiv preprint arXiv:1910.07475 (2019)."},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.naacl-main.329"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01064"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.repl4nlp-1.17"},{"key":"e_1_3_2_2_38_1","unstructured":"Wen-Cheng Lin and Hsin-Hsi Chen. 2002. Description of NTU Approach to NTCIR3 Multilingual Information Retrieval.. In NTCIR. Citeseer."},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210157"},{"key":"e_1_3_2_2_40_1","volume-title":"SIGIR'19","author":"Litschko Robert","year":"2018","unstructured":"Robert Litschko, Goran Glavavs, Simone Paolo Ponzetto, and Ivan Vuli\u0107. 2018b. Unsupervised cross-lingual information retrieval using monolingual data only. In SIGIR'19. 1253--1256."},{"key":"e_1_3_2_2_41_1","volume-title":"Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning. ArXiv","author":"Liu Haokun","year":"2022","unstructured":"Haokun Liu, Derek Tam, Mohammed Muqeeth, Jay Mohta, Tenghao Huang, Mohit Bansal, and Colin Raffel. 2022. Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning. ArXiv, Vol. abs\/2205.05638 (2022)."},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3412747"},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00369"},{"key":"e_1_3_2_2_44_1","volume-title":"Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining.","author":"Ma Rongrong","unstructured":"Rongrong Ma, Guansong Pang, Ling Chen, and Anton van den Hengel. 2022. Deep Graph-level Anomaly Detection by Glocal Knowledge Distillation. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining."},{"key":"e_1_3_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-45442-5_31"},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-99736-6_26"},{"key":"e_1_3_2_2_47_1","volume-title":"MS MARCO: A human generated machine reading comprehension dataset. choice","author":"Nguyen Tri","year":"2016","unstructured":"Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and Li Deng. 2016. MS MARCO: A human generated machine reading comprehension dataset. choice, Vol. 2640 (2016), 660."},{"key":"e_1_3_2_2_48_1","volume-title":"Passage Re-ranking with BERT. arXiv preprint arXiv:1901.04085","author":"Nogueira Rodrigo","year":"2019","unstructured":"Rodrigo Nogueira and Kyunghyun Cho. 2019. Passage Re-ranking with BERT. arXiv preprint arXiv:1901.04085 (2019)."},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1162\/089120103321337421"},{"key":"e_1_3_2_2_50_1","volume-title":"Multilingual information retrieval: From research to practice","author":"Peters Carol","unstructured":"Carol Peters, Martin Braschler, and Paul Clough. 2012. Multilingual information retrieval: From research to practice. Springer."},{"key":"e_1_3_2_2_51_1","volume-title":"Liu","author":"Raffel Colin","year":"2019","unstructured":"Colin Raffel, Noam M. Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2019. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. ArXiv, Vol. abs\/1910.10683 (2019)."},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10791-020-09372-2"},{"key":"e_1_3_2_2_53_1","volume-title":"Making monolingual sentence embeddings multilingual using knowledge distillation. arXiv preprint arXiv:2004.09813","author":"Reimers Nils","year":"2020","unstructured":"Nils Reimers and Iryna Gurevych. 2020. Making monolingual sentence embeddings multilingual using knowledge distillation. arXiv preprint arXiv:2004.09813 (2020)."},{"key":"e_1_3_2_2_54_1","volume-title":"LAReQA: Language-agnostic answer retrieval from a multilingual pool. arXiv preprint arXiv:2004.05484","author":"Roy Uma","year":"2020","unstructured":"Uma Roy, Noah Constant, Rami Al-Rfou, Aditya Barua, Aaron Phillips, and Yinfei Yang. 2020. LAReQA: Language-agnostic answer retrieval from a multilingual pool. arXiv preprint arXiv:2004.05484 (2020)."},{"key":"e_1_3_2_2_55_1","volume-title":"An Overview of Multi-Task Learning in Deep Neural Networks. ArXiv","author":"Ruder Sebastian","year":"2017","unstructured":"Sebastian Ruder. 2017. An Overview of Multi-Task Learning in Deep Neural Networks. ArXiv, Vol. abs\/1706.05098 (2017)."},{"key":"e_1_3_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.613"},{"key":"e_1_3_2_2_57_1","volume-title":"a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108","author":"Sanh Victor","year":"2019","unstructured":"Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019)."},{"key":"e_1_3_2_2_58_1","unstructured":"Victor Sanh Albert Webson Colin Raffel Stephen H. Bach Lintang Sutawika Zaid Alyafeai Antoine Chaffin Arnaud Stiegler Teven Le Scao Arun Raja Manan Dey M Saiful Bari Canwen Xu Urmish Thakker Shanya Sharma Eliza Szczechla Taewoon Kim Gunjan Chhablani Nihal V. Nayak Debajyoti Datta Jonathan Chang Mike Tian-Jian Jiang Han Wang Matteo Manica Sheng Shen Zheng Xin Yong Harshit Pandey Rachel Bawden Thomas Wang Trishala Neeraj Jos Rozen Abheesht Sharma Andrea Santilli Thibault F\u00e9vry Jason Alan Fries Ryan Teehan Stella Rose Biderman Leo Gao Tali Bers Thomas Wolf and Alexander M. Rush. 2021. Multitask Prompted Training Enables Zero-Shot Task Generalization. ArXiv Vol. abs\/2110.08207 (2021)."},{"key":"e_1_3_2_2_59_1","volume-title":"Zaharia","author":"Santhanam Keshav","year":"2021","unstructured":"Keshav Santhanam, O. Khattab, Jon Saad-Falcon, Christopher Potts, and Matei A. Zaharia. 2021. ColBERTv2: Effective and Efficient Retrieval via Lightweight Late Interaction. In North American Chapter of the Association for Computational Linguistics."},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1639"},{"key":"e_1_3_2_2_61_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-2073"},{"key":"e_1_3_2_2_62_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.naacl-main.290"},{"key":"e_1_3_2_2_63_1","volume-title":"Bitext Mining for Low-Resource Languages via Contrastive Learning. arXiv preprint arXiv:2208.11194","author":"Tan Weiting","year":"2022","unstructured":"Weiting Tan and Philipp Koehn. 2022. Bitext Mining for Low-Resource Languages via Contrastive Learning. arXiv preprint arXiv:2208.11194 (2022)."},{"key":"e_1_3_2_2_64_1","volume-title":"Spot: Better frozen model adaptation through soft prompt transfer. arXiv preprint arXiv:2110.07904","author":"Vu Tu","year":"2021","unstructured":"Tu Vu, Brian Lester, Noah Constant, Rami Al-Rfou, and Daniel Cer. 2021a. Spot: Better frozen model adaptation through soft prompt transfer. arXiv preprint arXiv:2110.07904 (2021)."},{"key":"e_1_3_2_2_65_1","volume-title":"Annual Meeting of the Association for Computational Linguistics.","author":"Vu Tu","year":"2021","unstructured":"Tu Vu, Brian Lester, Noah Constant, Rami Al-Rfou, and Daniel Matthew Cer. 2021b. SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer. In Annual Meeting of the Association for Computational Linguistics."},{"key":"e_1_3_2_2_66_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.635"},{"key":"e_1_3_2_2_67_1","volume-title":"Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval. 363--372","author":"Vuli\u0107 Ivan","year":"2015","unstructured":"Ivan Vuli\u0107 and Marie-Francine Moens. 2015. Monolingual and cross-lingual information retrieval models based on (bilingual) word embeddings. In Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval. 363--372."},{"key":"e_1_3_2_2_68_1","unstructured":"Yizhong Wang Swaroop Mishra Pegah Alipoormolabashi Yeganeh Kordi Amirreza Mirzaei Anjana Arunkumar Arjun Ashok Arut Selvan Dhanasekaran Atharva Naik David Stap Eshaan Pathak Giannis Karamanolakis Haizhi Gary Lai Ishan Purohit Ishani Mondal Jacob Anderson Kirby Kuznia Krima Doshi Maitreya Patel Kuntal Kumar Pal M. Moradshahi Mihir Parmar Mirali Purohit Neeraj Varshney Phani Rohitha Kaza Pulkit Verma Ravsehaj Singh Puri Rushang Karia Shailaja Keyur Sampat Savan Doshi Siddharth Deepak Mishra Sujan Reddy Sumanta Patro Tanay Dixit Xudong Shen Chitta Baral Yejin Choi Hannaneh Hajishirzi Noah A. Smith and Daniel Khashabi. 2022. Benchmarking Generalization via In-Context Instructions on 1 600 Language Tasks. ArXiv Vol. abs\/2204.07705 (2022)."},{"key":"e_1_3_2_2_69_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.240"},{"key":"e_1_3_2_2_70_1","volume-title":"Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning. In The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Nk2pDtuhTq","author":"Wang Zhen","year":"2023","unstructured":"Zhen Wang, Rameswar Panda, Leonid Karlinsky, Rogerio Feris, Huan Sun, and Yoon Kim. 2023. Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning. In The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Nk2pDtuhTq"},{"key":"e_1_3_2_2_71_1","volume-title":"Are all languages created equal in multilingual BERT? arXiv preprint arXiv:2005.09093","author":"Wu Shijie","year":"2020","unstructured":"Shijie Wu and Mark Dredze. 2020. Are all languages created equal in multilingual BERT? arXiv preprint arXiv:2005.09093 (2020)."},{"key":"e_1_3_2_2_72_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01070"},{"key":"e_1_3_2_2_73_1","volume-title":"Approximate nearest neighbor negative contrastive learning for dense text retrieval. arXiv preprint arXiv:2007.00808","author":"Xiong Lee","year":"2020","unstructured":"Lee Xiong, Chenyan Xiong, Ye Li, Kwok-Fung Tang, Jialin Liu, Paul Bennett, Junaid Ahmed, and Arnold Overwijk. 2020. Approximate nearest neighbor negative contrastive learning for dense text retrieval. arXiv preprint arXiv:2007.00808 (2020)."},{"key":"e_1_3_2_2_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/383952.383968"},{"key":"e_1_3_2_2_75_1","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531886"},{"key":"e_1_3_2_2_76_1","volume-title":"Parameter-efficient Zero-shot Transfer for Cross-Language Dense Retrieval with Adapters. arXiv preprint arXiv:2212.10448","author":"Yang Eugene","year":"2022","unstructured":"Eugene Yang, Suraj Nair, Dawn Lawrie, James Mayfield, and Douglas W Oard. 2022b. Parameter-efficient Zero-shot Transfer for Cross-Language Dense Retrieval with Adapters. arXiv preprint arXiv:2212.10448 (2022)."},{"key":"e_1_3_2_2_77_1","first-page":"19","volume-title":"Proceedings of Machine Translation Summit XVII","volume":"20","author":"Yarmohammadi Mahsa","year":"2019","unstructured":"Mahsa Yarmohammadi, Xutai Ma, Sorami Hisamoto, Muhammad Rahman, Yiming Wang, Hainan Xu, Daniel Povey, Philipp Koehn, and Kevin Duh. 2019. Robust Document Representations for Cross-Lingual Information Retrieval in Low-Resource Settings. In Proceedings of Machine Translation Summit XVII Volume 1: Research Track. European Association for Machine Translation, Dublin, Ireland, 12--20. https:\/\/www.aclweb.org\/anthology\/W19-6602"},{"key":"e_1_3_2_2_78_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401322"},{"key":"e_1_3_2_2_79_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3449830"},{"key":"e_1_3_2_2_80_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2019.2951463"},{"key":"e_1_3_2_2_81_1","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531791"},{"key":"e_1_3_2_2_82_1","volume-title":"TyDi: A multi-lingual benchmark for dense retrieval. arXiv preprint arXiv:2108.08787","author":"Zhang Xinyu","year":"2021","unstructured":"Xinyu Zhang, Xueguang Ma, Peng Shi, and Jimmy Lin. 2021. Mr. TyDi: A multi-lingual benchmark for dense retrieval. arXiv preprint arXiv:2108.08787 (2021)."},{"key":"e_1_3_2_2_83_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-emnlp.244"}],"event":{"name":"SIGIR '23: The 46th International ACM SIGIR Conference on Research and Development in Information Retrieval","location":"Taipei Taiwan","acronym":"SIGIR '23","sponsor":["SIGIR ACM Special Interest Group on Information Retrieval"]},"container-title":["Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3539618.3591769","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/abs\/10.1145\/3539618.3591769","content-type":"text\/html","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3539618.3591769","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3539618.3591769","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:47:02Z","timestamp":1750178822000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3539618.3591769"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,18]]},"references-count":83,"alternative-id":["10.1145\/3539618.3591769","10.1145\/3539618"],"URL":"https:\/\/doi.org\/10.1145\/3539618.3591769","relation":{},"subject":[],"published":{"date-parts":[[2023,7,18]]},"assertion":[{"value":"2023-07-18","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}