{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T22:59:34Z","timestamp":1772319574244,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":48,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,8,14]],"date-time":"2021-08-14T00:00:00Z","timestamp":1628899200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,8,14]]},"DOI":"10.1145\/3447548.3469053","type":"proceedings-article","created":{"date-parts":[[2021,8,12]],"date-time":"2021-08-12T06:12:05Z","timestamp":1628748725000},"page":"3717-3725","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["Domain-Specific Pretraining for Vertical Search: Case Study on Biomedical Literature"],"prefix":"10.1145","author":[{"given":"Yu","family":"Wang","sequence":"first","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Jinchao","family":"Li","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Tristan","family":"Naumann","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Chenyan","family":"Xiong","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Hao","family":"Cheng","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Robert","family":"Tinn","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Cliff","family":"Wong","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Naoto","family":"Usuyama","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Richard","family":"Rogahn","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Zhihong","family":"Shen","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Yang","family":"Qin","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Eric","family":"Horvitz","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Paul N.","family":"Bennett","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Jianfeng","family":"Gao","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]},{"given":"Hoifung","family":"Poon","sequence":"additional","affiliation":[{"name":"Microsoft Research, Redmond, WA, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,8,14]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W19-1909"},{"key":"e_1_3_2_1_2_1","unstructured":"Ricardo Baeza-Yates Berthier Ribeiro-Neto etal 2011. Modern information retrieval. Addison Wesley. Ricardo Baeza-Yates Berthier Ribeiro-Neto et al. 2011. Modern information retrieval. Addison Wesley."},{"key":"e_1_3_2_1_3_1","volume-title":"Proc. 2019 EMNLP-IJCNLP","author":"Beltagy Iz","year":"1865","unstructured":"Iz Beltagy , Kyle Lo , and Arman Cohan . 2019. SciBERT: A Pretrained Language Model for Scientific Text . In Proc. 2019 EMNLP-IJCNLP . Association for Computational Linguistics , Hong Kong , China, 3615--3620. https:\/\/doi.org\/10. 1865 3\/v1\/D19--1371 10.18653\/v1 Iz Beltagy, Kyle Lo, and Arman Cohan. 2019. SciBERT: A Pretrained Language Model for Scientific Text. In Proc. 2019 EMNLP-IJCNLP. Association for Computational Linguistics, Hong Kong, China, 3615--3620. https:\/\/doi.org\/10.18653\/v1\/D19--1371"},{"key":"e_1_3_2_1_4_1","unstructured":"Denis Bilenko. [n.d.]. gevent. http:\/\/www.gevent.org\/ Denis Bilenko. [n.d.]. gevent. http:\/\/www.gevent.org\/"},{"key":"e_1_3_2_1_5_1","volume-title":"The unified medical language system (UMLS): integrating biomedical terminology. Nucleic acids research32, suppl_1","author":"Bodenreider Olivier","year":"2004","unstructured":"Olivier Bodenreider . 2004. The unified medical language system (UMLS): integrating biomedical terminology. Nucleic acids research32, suppl_1 ( 2004 ), D267--D270. Olivier Bodenreider. 2004. The unified medical language system (UMLS): integrating biomedical terminology. Nucleic acids research32, suppl_1 (2004), D267--D270."},{"key":"e_1_3_2_1_6_1","unstructured":"Tom B Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell etal 2020. Language models are few-shot learners. arXiv preprintarXiv:2005.14165(2020). Tom B Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language models are few-shot learners. arXiv preprintarXiv:2005.14165(2020)."},{"key":"e_1_3_2_1_7_1","unstructured":"Hao Cheng Yelong Shen Xiaodong Liu Pengcheng He Weizhu Chen and Jianfeng Gao. [n.d.]. UnitedQA: A Hybrid Approach for Open Domain Question Answering. arXiv preprint arXiv:2101.00178([n. d.]). Hao Cheng Yelong Shen Xiaodong Liu Pengcheng He Weizhu Chen and Jianfeng Gao. [n.d.]. UnitedQA: A Hybrid Approach for Open Domain Question Answering. arXiv preprint arXiv:2101.00178([n. d.])."},{"key":"e_1_3_2_1_8_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2018 . Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv : 1810.04805(2018). Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv: 1810.04805(2018)."},{"key":"e_1_3_2_1_9_1","unstructured":"Li Dong Nan Yang Wenhui Wang Furu Wei Xiaodong Liu Yu Wang Jianfeng Gao Ming Zhou and Hsiao-Wuen Hon. 2019. Unified language model pre-training for natural language understanding and generation. arXiv preprintarXiv:1905.03197(2019). Li Dong Nan Yang Wenhui Wang Furu Wei Xiaodong Liu Yu Wang Jianfeng Gao Ming Zhou and Hsiao-Wuen Hon. 2019. Unified language model pre-training for natural language understanding and generation. arXiv preprintarXiv:1905.03197(2019)."},{"key":"e_1_3_2_1_10_1","volume-title":"Elasticsearch: the definitive guide: a distributed real-time search and analytics engine. \" O'Reilly Media","author":"Gormley Clinton","unstructured":"Clinton Gormley and Zachary Tong . 2015. Elasticsearch: the definitive guide: a distributed real-time search and analytics engine. \" O'Reilly Media , Inc .\". Clinton Gormley and Zachary Tong. 2015. Elasticsearch: the definitive guide: a distributed real-time search and analytics engine. \" O'Reilly Media, Inc.\"."},{"key":"e_1_3_2_1_11_1","unstructured":"Yu Gu Robert Tinn Hao Cheng Michael Lucas Naoto Usuyama Xiaodong Liu Tristan Naumann Jianfeng Gao and Hoifung Poon. 2020. Domain-specific language model pretraining for biomedical natural language processing. arXivpreprint arXiv:2007.15779(2020). Yu Gu Robert Tinn Hao Cheng Michael Lucas Naoto Usuyama Xiaodong Liu Tristan Naumann Jianfeng Gao and Hoifung Poon. 2020. Domain-specific language model pretraining for biomedical natural language processing. arXivpreprint arXiv:2007.15779(2020)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"Suchin Gururangan Ana Marasovic Swabha Swayamdipta Kyle Lo Iz Beltagy Doug Downey and Noah A Smith. 2020. Don't Stop Pretraining: Adapt Language Models to Domains and Tasks. arXiv preprint arXiv:2004.10964(2020). Suchin Gururangan Ana Marasovic Swabha Swayamdipta Kyle Lo Iz Beltagy Doug Downey and Noah A Smith. 2020. Don't Stop Pretraining: Adapt Language Models to Domains and Tasks. arXiv preprint arXiv:2004.10964(2020).","DOI":"10.18653\/v1\/2020.acl-main.740"},{"key":"e_1_3_2_1_13_1","volume-title":"Long short-term memory. Neural computation 9, 8","author":"Hochreiter Sepp","year":"1997","unstructured":"Sepp Hochreiter and J\u00fcrgen Schmidhuber . 1997. Long short-term memory. Neural computation 9, 8 ( 1997 ), 1735--1780. Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural computation 9, 8 (1997), 1735--1780."},{"key":"e_1_3_2_1_14_1","unstructured":"Lars Holmberg and Jonatan Heyman. [n.d.].locust. https:\/\/locust.io\/ Lars Holmberg and Jonatan Heyman. [n.d.].locust. https:\/\/locust.io\/"},{"key":"e_1_3_2_1_15_1","unstructured":"Kexin Huang Jaan Altosaar and Rajesh Ranganath. 2019. Clinical BERT: Modeling clinical notes and predicting hospital readmission. arXiv preprint arXiv:1904.05342(2019). Kexin Huang Jaan Altosaar and Rajesh Ranganath. 2019. Clinical BERT: Modeling clinical notes and predicting hospital readmission. arXiv preprint arXiv:1904.05342(2019)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/2505515.2505665"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Vladimir Karpukhin Barlas O?uz Sewon Min Ledell Wu Sergey Edunov Danqi Chen and Wen-tau Yih. 2020. Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906(2020). Vladimir Karpukhin Barlas O?uz Sewon Min Ledell Wu Sergey Edunov Danqi Chen and Wen-tau Yih. 2020. Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906(2020).","DOI":"10.18653\/v1\/2020.emnlp-main.550"},{"key":"e_1_3_2_1_18_1","volume-title":"Proc. 2018 EMNLP: System Demonstrations. Association for Computational Linguistics","author":"Kudo Taku","year":"2018","unstructured":"Taku Kudo and John Richardson . 2018 . Sentence Piece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing . In Proc. 2018 EMNLP: System Demonstrations. Association for Computational Linguistics , Brussels, Belgium, 66--71. https:\/\/doi.org\/10. 18653\/v1\/D18--2012 10.18653\/v1 Taku Kudo and John Richardson. 2018. Sentence Piece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing. In Proc. 2018 EMNLP: System Demonstrations. Association for Computational Linguistics, Brussels, Belgium, 66--71. https:\/\/doi.org\/10.18653\/v1\/D18--2012"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00276"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"crossref","first-page":"1234","DOI":"10.1093\/bioinformatics\/btz682","article-title":"BioBERT: a pre-trained biomedical language representation model for biomedical text mining","volume":"36","author":"Lee Jinhyuk","year":"2020","unstructured":"Jinhyuk Lee , Wonjin Yoon , Sungdong Kim , Donghyeon Kim , Sunkyu Kim , Chan Ho So , and Jaewoo Kang . 2020 . BioBERT: a pre-trained biomedical language representation model for biomedical text mining . Bioinformatics 36 , 4 (2020), 1234 -- 1240 . Jinhyuk Lee, Wonjin Yoon, Sungdong Kim, Donghyeon Kim, Sunkyu Kim, Chan Ho So, and Jaewoo Kang. 2020. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36, 4 (2020), 1234--1240.","journal-title":"Bioinformatics"},{"key":"e_1_3_2_1_21_1","unstructured":"Jimmy Lin Rodrigo Nogueira and Andrew Yates. 2020. Pretrained transformers for text ranking: Bert and beyond. arXiv preprint arXiv:2010.06467(2020). Jimmy Lin Rodrigo Nogueira and Andrew Yates. 2020. Pretrained transformers for text ranking: Bert and beyond. arXiv preprint arXiv:2010.06467(2020)."},{"key":"e_1_3_2_1_22_1","first-page":"265","article-title":"Medical subject headings (MeSH)","volume":"88","author":"Lipscomb Carolyn E","year":"2000","unstructured":"Carolyn E Lipscomb . 2000 . Medical subject headings (MeSH) . Bulletin of the Medical Library Association 88 , 3 (2000), 265 . Carolyn E Lipscomb. 2000. Medical subject headings (MeSH). Bulletin of the Medical Library Association 88, 3 (2000), 265.","journal-title":"Bulletin of the Medical Library Association"},{"key":"e_1_3_2_1_23_1","volume-title":"Roberta: Arobustly optimized bert pretraining approach.arXiv preprint arXiv:1907.11692(2019).","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu , Myle Ott , Naman Goyal , Jingfei Du , Mandar Joshi , Danqi Chen , Omer Levy , Mike Lewis , Luke Zettlemoyer , and Veselin Stoyanov . 2019 . Roberta: Arobustly optimized bert pretraining approach.arXiv preprint arXiv:1907.11692(2019). Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: Arobustly optimized bert pretraining approach.arXiv preprint arXiv:1907.11692(2019)."},{"key":"e_1_3_2_1_24_1","volume-title":"SLEDGE: A Simple Yet Effective Baseline for Coronavirus Scientific Knowledge Search. arXiv preprint arXiv:2005.02365(2020).","author":"MacAvaney Sean","year":"2020","unstructured":"Sean MacAvaney , Arman Cohan , and Nazli Goharian . 2020 . SLEDGE: A Simple Yet Effective Baseline for Coronavirus Scientific Knowledge Search. arXiv preprint arXiv:2005.02365(2020). Sean MacAvaney, Arman Cohan, and Nazli Goharian. 2020. SLEDGE: A Simple Yet Effective Baseline for Coronavirus Scientific Knowledge Search. arXiv preprint arXiv:2005.02365(2020)."},{"key":"e_1_3_2_1_25_1","volume-title":"MS MARCO: A human generated machine reading comprehension dataset. In CoCo@ NIPS.","author":"Nguyen Tri","year":"2016","unstructured":"Tri Nguyen , Mir Rosenberg , Xia Song , Jianfeng Gao , Saurabh Tiwary , Rangan Majumder , and Li Deng . 2016 . MS MARCO: A human generated machine reading comprehension dataset. In CoCo@ NIPS. Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and Li Deng. 2016. MS MARCO: A human generated machine reading comprehension dataset. In CoCo@ NIPS."},{"key":"e_1_3_2_1_26_1","unstructured":"Rodrigo Nogueira and Kyunghyun Cho. 2019. Passage Reranking with BERT. (2019). http:\/\/arxiv.org\/abs\/1901.04085 Rodrigo Nogueira and Kyunghyun Cho. 2019. Passage Reranking with BERT. (2019). http:\/\/arxiv.org\/abs\/1901.04085"},{"key":"e_1_3_2_1_27_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog 1, 8","author":"Radford Alec","year":"2019","unstructured":"Alec Radford , Jeffrey Wu , Rewon Child , David Luan , Dario Amodei , and Ilya Sutskever . 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 ( 2019 ), 9. Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9."},{"key":"e_1_3_2_1_28_1","volume-title":"Liu","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel , Noam Shazeer , Adam Roberts , Katherine Lee , Sharan Narang , Michael Matena , Yanqi Zhou , Wei Li , and Peter J . Liu . 2020 . Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer.Journal of Machine Learning Research 21, 140 (2020), 1--67. http:\/\/jmlr.org\/papers\/v21\/20-074.html Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer.Journal of Machine Learning Research21, 140 (2020), 1--67. http:\/\/jmlr.org\/papers\/v21\/20-074.html"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocaa091"},{"key":"e_1_3_2_1_30_1","first-page":"129","article-title":"Relevance weighting of search terms","volume":"27","author":"Robertson Stephen E","year":"1976","unstructured":"Stephen E Robertson and K Sparck Jones . 1976 . Relevance weighting of search terms . J. Assoc. Inf. Sci. Technol. 27 , 3 (1976), 129 -- 146 . Stephen E Robertson and K Sparck Jones. 1976. Relevance weighting of search terms. J. Assoc. Inf. Sci. Technol. 27, 3 (1976), 129--146.","journal-title":"J. Assoc. Inf. Sci. Technol."},{"key":"e_1_3_2_1_31_1","unstructured":"Armin Ronacher. [n.d.]. Flask. https:\/\/palletsprojects.com\/p\/flask\/ Armin Ronacher. [n.d.]. Flask. https:\/\/palletsprojects.com\/p\/flask\/"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P16-1162"},{"key":"e_1_3_2_1_33_1","unstructured":"Yuqi Si Jingqi Wang Hua Xu and Kirk Roberts. 2019. Enhancing clinical concept extraction with contextual embeddings. J. Am. Med. Inform. Assoc.(2019). Yuqi Si Jingqi Wang Hua Xu and Kirk Roberts. 2019. Enhancing clinical concept extraction with contextual embeddings. J. Am. Med. Inform. Assoc.(2019)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/2740908.2742839"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocaa271"},{"key":"e_1_3_2_1_36_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez Lukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998--6008. Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez Lukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998--6008."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"crossref","unstructured":"Ellen Voorhees Tasmeer Alam Steven Bedrick Dina Demner-Fushman William R Hersh Kyle Lo Kirk Roberts Ian Soboroff and Lucy Lu Wang. 2020. TREC-COVID: constructing a pandemic information retrieval test collection. arXiv preprint arXiv:2005.04474(2020). Ellen Voorhees Tasmeer Alam Steven Bedrick Dina Demner-Fushman William R Hersh Kyle Lo Kirk Roberts Ian Soboroff and Lucy Lu Wang. 2020. TREC-COVID: constructing a pandemic information retrieval test collection. arXiv preprint arXiv:2005.04474(2020).","DOI":"10.1145\/3451964.3451965"},{"key":"e_1_3_2_1_38_1","volume-title":"Superglue: A stickier benchmark for general-purpose language understanding systems. In Advances in Neural Information Processing Systems. 3266--3280.","author":"Wang Alex","year":"2019","unstructured":"Alex Wang , Yada Pruksachatkun , Nikita Nangia , Amanpreet Singh , Julian Michael , Felix Hill , Omer Levy , and Samuel Bowman . 2019 . Superglue: A stickier benchmark for general-purpose language understanding systems. In Advances in Neural Information Processing Systems. 3266--3280. Alex Wang, Yada Pruksachatkun, Nikita Nangia, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel Bowman. 2019. Superglue: A stickier benchmark for general-purpose language understanding systems. In Advances in Neural Information Processing Systems. 3266--3280."},{"key":"e_1_3_2_1_39_1","volume-title":"Bowman","author":"Wang Alex","year":"2019","unstructured":"Alex Wang , Amanpreet Singh , Julian Michael , Felix Hill , Omer Levy , and Samuel R . Bowman . 2019 . GLUE : A MULTI-TASK BENCHMARK AND ANALYSIS PLAT-FORM FOR NATURAL LANGUAGE UNDERSTANDING. In ICLR. Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R.Bowman. 2019. GLUE: A MULTI-TASK BENCHMARK AND ANALYSIS PLAT-FORM FOR NATURAL LANGUAGE UNDERSTANDING. In ICLR."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1162\/qss_a_00021"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.3389\/fdata.2019.00045"},{"key":"e_1_3_2_1_42_1","unstructured":"Lucy Lu Wang Kyle Lo Yoganand Chandrasekhar Russell Reas Jiangjiang Yang Darrin Eide Kathryn Funk Rodney Kinney Ziyang Liu William Merrill etal2020. Cord-19: The covid-19 open research dataset.ArXiv(2020). Lucy Lu Wang Kyle Lo Yoganand Chandrasekhar Russell Reas Jiangjiang Yang Darrin Eide Kathryn Funk Rodney Kinney Ziyang Liu William Merrill et al.2020. Cord-19: The covid-19 open research dataset.ArXiv(2020)."},{"key":"e_1_3_2_1_43_1","unstructured":"Chenyan Xiong Zhenghao Liu Si Sun Zhuyun Dai Kaitao Zhang Shi Yu Zhiyuan Liu Hoifung Poon Jianfeng Gao and Paul Bennett. 2020. CMT in TREC-COVID Round 2: Mitigating the Generalization Gaps from Web to Special Domain Search. arXiv preprint arXiv:2011.01580(2020). Chenyan Xiong Zhenghao Liu Si Sun Zhuyun Dai Kaitao Zhang Shi Yu Zhiyuan Liu Hoifung Poon Jianfeng Gao and Paul Bennett. 2020. CMT in TREC-COVID Round 2: Mitigating the Generalization Gaps from Web to Special Domain Search. arXiv preprint arXiv:2011.01580(2020)."},{"key":"e_1_3_2_1_44_1","volume-title":"Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=zeFrfgyZln","author":"Xiong Lee","year":"2021","unstructured":"Lee Xiong , Chenyan Xiong , Ye Li , Kwok-Fung Tang , Jialin Liu , Paul N. Bennett , Junaid Ahmed , and Arnold Overwikj . 2021 . Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=zeFrfgyZln Lee Xiong, Chenyan Xiong, Ye Li, Kwok-Fung Tang, Jialin Liu, Paul N. Bennett, Junaid Ahmed, and Arnold Overwikj. 2021. Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=zeFrfgyZln"},{"key":"e_1_3_2_1_45_1","unstructured":"Wei Yang Haotian Zhang and Jimmy Lin. 2019. Simple applications of BERT for ad hoc document retrieval.arXiv preprint arXiv:1903.10972(2019). Wei Yang Haotian Zhang and Jimmy Lin. 2019. Simple applications of BERT for ad hoc document retrieval.arXiv preprint arXiv:1903.10972(2019)."},{"key":"e_1_3_2_1_46_1","volume-title":"Proc. 2019 EMNLP-IJCNLP: System Demonstrations. 19--24","author":"Yilmaz Zeynep Akkalyoncu","year":"2019","unstructured":"Zeynep Akkalyoncu Yilmaz , Shengjin Wang , Wei Yang , Haotian Zhang , and Jimmy Lin . 2019 . Applying BERT to document retrieval with birch . In Proc. 2019 EMNLP-IJCNLP: System Demonstrations. 19--24 . Zeynep Akkalyoncu Yilmaz, Shengjin Wang, Wei Yang, Haotian Zhang, and Jimmy Lin. 2019. Applying BERT to document retrieval with birch. In Proc. 2019 EMNLP-IJCNLP: System Demonstrations. 19--24."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380131"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"crossref","unstructured":"Yukun Zhu Ryan Kiros Rich Zemel Ruslan Salakhutdinov Raquel Urtasun Antonio Torralba and Sanja Fidler. 2015. Aligning books and movies: Towards story-like visual explanations by watching movies and reading books. In ICCV. Yukun Zhu Ryan Kiros Rich Zemel Ruslan Salakhutdinov Raquel Urtasun Antonio Torralba and Sanja Fidler. 2015. Aligning books and movies: Towards story-like visual explanations by watching movies and reading books. In ICCV.","DOI":"10.1109\/ICCV.2015.11"}],"event":{"name":"KDD '21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Virtual Event Singapore","acronym":"KDD '21","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3447548.3469053","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3447548.3469053","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:37Z","timestamp":1750191517000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3447548.3469053"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,8,14]]},"references-count":48,"alternative-id":["10.1145\/3447548.3469053","10.1145\/3447548"],"URL":"https:\/\/doi.org\/10.1145\/3447548.3469053","relation":{},"subject":[],"published":{"date-parts":[[2021,8,14]]},"assertion":[{"value":"2021-08-14","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}