{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T13:04:20Z","timestamp":1760101460423,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":22,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819981779"},{"type":"electronic","value":"9789819981786"}],"license":[{"start":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T00:00:00Z","timestamp":1701302400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T00:00:00Z","timestamp":1701302400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8178-6_1","type":"book-chapter","created":{"date-parts":[[2023,11,29]],"date-time":"2023-11-29T10:02:54Z","timestamp":1701252174000},"page":"3-15","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Improve Conversational Search with\u00a0Multi-document Information"],"prefix":"10.1007","author":[{"given":"Shiyulong","family":"He","sequence":"first","affiliation":[]},{"given":"Sai","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xiaowang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Zhiyong","family":"Feng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,30]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Bansal, S., et al.: R3: refined retriever-reader pipeline for multidoc2dial. In: Proceedings of the Second DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering, DialDoc@ACL 2022, Dublin, Ireland, May 26, 2022, pp. 148\u2013154 (2022)","key":"1_CR1","DOI":"10.18653\/v1\/2022.dialdoc-1.17"},{"doi-asserted-by":"crossref","unstructured":"Bonifacio, L., Abonizio, H., Fadaee, M., Nogueira, R.: Inpars: data augmentation for information retrieval using large language models (2022)","key":"1_CR2","DOI":"10.1145\/3477495.3531863"},{"doi-asserted-by":"crossref","unstructured":"Feng, S., Patel, S.S., Wan, H., Joshi, S.: Multidoc2dial: modeling dialogues grounded in multiple documents, pp. 6162\u20136176 (2021)","key":"1_CR3","DOI":"10.18653\/v1\/2021.emnlp-main.498"},{"doi-asserted-by":"crossref","unstructured":"Han, J., Hong, T., Kim, B., Ko, Y., Seo, J.: Fine-grained post-training for improving retrieval-based dialogue systems. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1549\u20131558 (2021)","key":"1_CR4","DOI":"10.18653\/v1\/2021.naacl-main.122"},{"doi-asserted-by":"crossref","unstructured":"He, S., Zhang, S., Zhang, X., Feng, Z.: Conversational search based on utterance-mask-passage post-training. In: China Conference on Knowledge Graph and Semantic Computing (2023)","key":"1_CR5","DOI":"10.1007\/978-981-99-7224-1_12"},{"doi-asserted-by":"crossref","unstructured":"Ishii, E., Xu, Y., Cahyawijaya, S., Wilie, B.: Can question rewriting help conversational question answering? pp. 94\u201399 (2022)","key":"1_CR6","DOI":"10.18653\/v1\/2022.insights-1.13"},{"doi-asserted-by":"crossref","unstructured":"Jang, Y., et al.: Improving multiple documents grounded goal-oriented dialog systems via diverse knowledge enhanced pretrained language model. In: Proceedings of the Second DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering, pp. 136\u2013141 (2022)","key":"1_CR7","DOI":"10.18653\/v1\/2022.dialdoc-1.15"},{"unstructured":"Jeronymo, V., et al.: Inpars-v2: large language models as efficient dataset generators for information retrieval. arXiv preprint arXiv:2301.01820 (2023)","key":"1_CR8"},{"doi-asserted-by":"crossref","unstructured":"Jiang, Y., Hadifar, A., Deleu, J., Demeester, T., Develder, C.: Ugent-t2k at the 2nd dialdoc shared task: a retrieval-focused dialog system grounded in multiple documents. In: Proceedings of the Second DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering, pp. 115\u2013122 (2022)","key":"1_CR9","DOI":"10.18653\/v1\/2022.dialdoc-1.12"},{"doi-asserted-by":"crossref","unstructured":"Karpukhin, V., et al.: Dense passage retrieval for open-domain question answering, pp. 6769\u20136781 (2020)","key":"1_CR10","DOI":"10.18653\/v1\/2020.emnlp-main.550"},{"doi-asserted-by":"crossref","unstructured":"Kim, S., Kim, G.: Saving dense retriever from shortcut dependency in conversational search, pp. 10278\u201310287 (2022)","key":"1_CR11","DOI":"10.18653\/v1\/2022.emnlp-main.701"},{"doi-asserted-by":"crossref","unstructured":"Li, K., et al.: Grounded dialogue generation with cross-encoding re-ranker, grounding span prediction, and passage dropout. In: Proceedings of the Second DialDoc Workshop on Document-Grounded Dialogue and Conversational Question Answering, pp. 123\u2013129 (2022)","key":"1_CR12","DOI":"10.18653\/v1\/2022.dialdoc-1.13"},{"doi-asserted-by":"crossref","unstructured":"Liu, S., Johns, E., Davison, A.J.: End-to-end multi-task learning with attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1871\u20131880 (2019)","key":"1_CR13","DOI":"10.1109\/CVPR.2019.00197"},{"doi-asserted-by":"crossref","unstructured":"Mao, K., et al.: Learning denoised and interpretable session representation for conversational search. In: Proceedings of the ACM Web Conference, pp. 3193\u20133202 (2023)","key":"1_CR14","DOI":"10.1145\/3543507.3583265"},{"doi-asserted-by":"crossref","unstructured":"Shen, T., Long, G., Geng, X., Tao, C., Zhou, T., Jiang, D.: Large language models are strong zero-shot retriever. arXiv preprint arXiv:2304.14233 (2023)","key":"1_CR15","DOI":"10.18653\/v1\/2024.findings-acl.943"},{"doi-asserted-by":"crossref","unstructured":"Tran, N., Litman, D.: Getting better dialogue context for knowledge identification by leveraging document-level topic shift. In: Proceedings of the 23rd Annual Meeting of the Special Interest Group on Discourse and Dialogue, pp. 368\u2013375 (2022)","key":"1_CR16","DOI":"10.18653\/v1\/2022.sigdial-1.36"},{"doi-asserted-by":"crossref","unstructured":"Wang, Z., Tu, Y., Rosset, C., Craswell, N., Wu, M., Ai, Q.: Zero-shot clarifying question generation for conversational search. In: Proceedings of the ACM Web Conference 2023, pp. 3288\u20133298 (2023)","key":"1_CR17","DOI":"10.1145\/3543507.3583420"},{"doi-asserted-by":"crossref","unstructured":"Wu, Z., et al.: CONQRR: conversational query rewriting for retrieval with reinforcement learning. arXiv preprint arXiv:2112.08558 (2021)","key":"1_CR18","DOI":"10.18653\/v1\/2022.emnlp-main.679"},{"doi-asserted-by":"crossref","unstructured":"Xu, Y., Zhao, H., Zhang, Z.: Topic-aware multi-turn dialogue modeling. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 14176\u201314184 (2021)","key":"1_CR19","DOI":"10.1609\/aaai.v35i16.17668"},{"unstructured":"Yu, W., et al.: Generate rather than retrieve: large language models are strong context generators. arXiv preprint arXiv:2209.10063 (2022)","key":"1_CR20"},{"doi-asserted-by":"crossref","unstructured":"Zhang, S., Du, Y., Liu, G., Yan, Z., Cao, Y.: G4: grounding-guided goal-oriented dialogues generation with multiple documents. In: Proceedings of the Second DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering, pp. 108\u2013114 (2022)","key":"1_CR21","DOI":"10.18653\/v1\/2022.dialdoc-1.11"},{"doi-asserted-by":"crossref","unstructured":"Zhang, T., Ladhak, F., Durmus, E., Liang, P., McKeown, K., Hashimoto, T.B.: Benchmarking large language models for news summarization. arXiv preprint arXiv:2301.13848 (2023)","key":"1_CR22","DOI":"10.1162\/tacl_a_00632"}],"container-title":["Communications in Computer and Information Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8178-6_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,4]],"date-time":"2024-11-04T07:34:31Z","timestamp":1730705671000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8178-6_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,30]]},"ISBN":["9789819981779","9789819981786"],"references-count":22,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8178-6_1","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2023,11,30]]},"assertion":[{"value":"30 November 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Changsha","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 November 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/iconip2023.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1274","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"650","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"51% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4.14","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.46","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}