{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,3]],"date-time":"2025-05-03T00:28:21Z","timestamp":1746232101022,"version":"3.40.3"},"publisher-location":"Cham","reference-count":26,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031247545"},{"type":"electronic","value":"9783031247552"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-24755-2_3","type":"book-chapter","created":{"date-parts":[[2023,2,2]],"date-time":"2023-02-02T17:02:20Z","timestamp":1675357340000},"page":"28-39","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Enhance Performance of Ad-hoc Search via Prompt Learning"],"prefix":"10.1007","author":[{"given":"Shenghao","family":"Yang","sequence":"first","affiliation":[]},{"given":"Yiqun","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Xiaohui","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Min","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Shaoping","family":"Ma","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,3]]},"reference":[{"key":"3_CR1","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"3_CR2","unstructured":"Nogueira, R., Cho, K.: Passage re-ranking with BERT. arXiv preprint arXiv:1901.04085 (2019)"},{"key":"3_CR3","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., Neubig, G.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. arXiv preprint arXiv:2107.13586 (2021)"},{"key":"3_CR4","doi-asserted-by":"crossref","unstructured":"Zhang, X., Yates, A., Lin, J.: A little bit is worse than none: ranking with limited training data. In: Proceedings of SustaiNLP: Workshop on Simple and Efficient Natural Language Processing, pp. 107\u2013112 (2020)","DOI":"10.18653\/v1\/2020.sustainlp-1.14"},{"key":"3_CR5","unstructured":"Petroni, F., et al.: Language models as knowledge bases? arXiv preprint arXiv:1909.01066 (2019)"},{"key":"3_CR6","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3_CR7","unstructured":"Liu, X., et al.: GPT understands, too. arXiv preprint arXiv:2103.10385 (2021)"},{"key":"3_CR8","doi-asserted-by":"crossref","unstructured":"Han, X., Zhao, W., Ding, N., Liu, Z., Sun, M.: PTR: prompt tuning with rules for text classification. arXiv preprint arXiv:2105.11259 (2021)","DOI":"10.1016\/j.aiopen.2022.11.003"},{"key":"3_CR9","doi-asserted-by":"crossref","unstructured":"Nogueira, R., Jiang, Z., Lin, J.: Document ranking with a pretrained sequence-to-sequence model. arXiv preprint arXiv:2003.06713 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.63"},{"issue":"140","key":"3_CR10","first-page":"1","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(140), 1\u201367 (2020)","journal-title":"J. Mach. Learn. Res."},{"key":"3_CR11","doi-asserted-by":"crossref","unstructured":"Hu, X., Yu, S., Xiong, C., Liu, Z., Liu, Z., Yu, G.: P $$^{3}$$ ranker: mitigating the gaps between pre-training and ranking fine-tuning with prompt-based learning and pre-finetuning. arXiv preprint arXiv:2205.01886 (2022)","DOI":"10.1145\/3477495.3531786"},{"key":"3_CR12","doi-asserted-by":"crossref","unstructured":"Williams, A., Nangia, N., Bowman, S.R.: A broad-coverage challenge corpus for sentence understanding through inference. arXiv preprint arXiv:1704.05426 (2017)","DOI":"10.18653\/v1\/N18-1101"},{"key":"3_CR13","unstructured":"Liu, Y., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"3_CR14","unstructured":"Nguyen, T., et al.: MS MARCO: a human generated machine reading comprehension dataset. In: CoCo@ NIPS (2016)"},{"key":"3_CR15","first-page":"109","volume":"109","author":"SE Robertson","year":"1995","unstructured":"Robertson, S.E., Walker, S., Jones, S., Hancock-Beaulieu, M.M., Gatford, M., et al.: Okapi at TREC-3. NIST Spec. Publ. 109, 109 (1995)","journal-title":"NIST Spec. Publ."},{"key":"3_CR16","doi-asserted-by":"crossref","unstructured":"Karpukhin, V., et al.: Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.550"},{"key":"3_CR17","unstructured":"Xiong, L., et al.: Approximate nearest neighbor negative contrastive learning for dense text retrieval. arXiv preprint arXiv:2007.00808 (2020)"},{"key":"3_CR18","doi-asserted-by":"crossref","unstructured":"Qu, Y., et al.: RocketQA: an optimized training approach to dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2010.08191 (2020)","DOI":"10.18653\/v1\/2021.naacl-main.466"},{"key":"3_CR19","unstructured":"Li, C., Yates, A., MacAvaney, S., He, B., Sun, Y.: PARADE: passage representation aggregation for document reranking. arXiv preprint arXiv:2008.09093 (2020)"},{"key":"3_CR20","doi-asserted-by":"crossref","unstructured":"Dai, Z., Callan, J.: Deeper text understanding for IR with contextual neural language modeling. In: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 985\u2013988 (2019)","DOI":"10.1145\/3331184.3331303"},{"key":"3_CR21","unstructured":"Nogueira, R., Yang, W., Cho, K., Lin, J.: Multi-stage document ranking with BERT. arXiv preprint arXiv:1910.14424 (2019)"},{"key":"3_CR22","doi-asserted-by":"crossref","unstructured":"Shin, T., Razeghi, Y., Logan IV, R.L., Wallace, E., Singh, S.: Autoprompt: eliciting knowledge from language models with automatically generated prompts. arXiv preprint arXiv:2010.15980 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.346"},{"key":"3_CR23","doi-asserted-by":"crossref","unstructured":"Li, X.L., Liang, P.: Prefix-tuning: optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190 (2021)","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"3_CR24","doi-asserted-by":"crossref","unstructured":"Hambardzumyan, K., Khachatrian, H., May, J.: WARP: word-level adversarial reprogramming. arXiv preprint arXiv:2101.00121 (2021)","DOI":"10.18653\/v1\/2021.acl-long.381"},{"key":"3_CR25","doi-asserted-by":"crossref","unstructured":"Ding, N., et al.: OpenPrompt: an open-source framework for prompt-learning. arXiv preprint arXiv:2111.01998 (2021)","DOI":"10.18653\/v1\/2022.acl-demo.10"},{"key":"3_CR26","unstructured":"Wolf, T., et al.: HuggingFace\u2019s transformers: state-of-the-art natural language processing. arXiv preprint arXiv:1910.03771 (2019)"}],"container-title":["Lecture Notes in Computer Science","Information Retrieval"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-24755-2_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,2]],"date-time":"2023-02-02T17:02:59Z","timestamp":1675357379000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-24755-2_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031247545","9783031247552"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-24755-2_3","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"3 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CCIR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China Conference on Information Retrieval","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chongqing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccir2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}