{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:10:02Z","timestamp":1755821402720,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":33,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,4,30]],"date-time":"2023-04-30T00:00:00Z","timestamp":1682812800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,4,30]]},"DOI":"10.1145\/3543873.3587540","type":"proceedings-article","created":{"date-parts":[[2023,4,28]],"date-time":"2023-04-28T11:36:14Z","timestamp":1682681774000},"page":"586-590","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Multi-turn mediated solutions for Conversational Artificial Intelligent systems leveraging graph-based techniques"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-3826-4443","authenticated-orcid":false,"given":"Riya","family":"Naik","sequence":"first","affiliation":[{"name":"Computer Science &amp; Information Systems, Birla Institute Of Technology And Science, Pilani, India"}]}],"member":"320","published-online":{"date-parts":[[2023,4,30]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Proceedings of the","author":"Bobrow G","year":"1964","unstructured":"Daniel\u00a0G Bobrow. 1964. A question-answering system for high school algebra word problems. In Proceedings of the October 27-29, 1964, fall joint computer conference, part I. 591\u2013614."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2015.12.005"},{"key":"e_1_3_2_1_3_1","volume-title":"Language models are few-shot learners. Advances in neural information processing systems 33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared\u00a0D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, 2020. Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877\u20131901."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12045-014-0029-7"},{"key":"e_1_3_2_1_5_1","volume-title":"Improving commonsense question answering by graph-based iterative retrieval over multiple knowledge sources. arXiv preprint arXiv:2011.02705","author":"Chen Qianglong","year":"2020","unstructured":"Qianglong Chen, Feng Ji, Haiqing Chen, and Yin Zhang. 2020. Improving commonsense question answering by graph-based iterative retrieval over multiple knowledge sources. arXiv preprint arXiv:2011.02705 (2020)."},{"key":"e_1_3_2_1_6_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_7_1","volume-title":"Building Watson: An overview of the DeepQA project. AI magazine 31, 3","author":"Ferrucci David","year":"2010","unstructured":"David Ferrucci, Eric Brown, Jennifer Chu-Carroll, James Fan, David Gondek, Aditya\u00a0A Kalyanpur, Adam Lally, J\u00a0William Murdock, Eric Nyberg, John Prager, 2010. Building Watson: An overview of the DeepQA project. AI magazine 31, 3 (2010), 59\u201379."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1610"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/1460690.1460714"},{"key":"e_1_3_2_1_10_1","first-page":"12837","article-title":"Convbert: Improving bert with span-based dynamic convolution","volume":"33","author":"Jiang Zi-Hang","year":"2020","unstructured":"Zi-Hang Jiang, Weihao Yu, Daquan Zhou, Yunpeng Chen, Jiashi Feng, and Shuicheng Yan. 2020. Convbert: Improving bert with span-based dynamic convolution. Advances in Neural Information Processing Systems 33 (2020), 12837\u201312848.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00276"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.5555\/1624435.1624467"},{"key":"e_1_3_2_1_13_1","first-page":"64","article-title":"Recurrent neural networks","volume":"5","author":"Medsker R","year":"2001","unstructured":"Larry\u00a0R Medsker and LC Jain. 2001. Recurrent neural networks. Design and Applications 5 (2001), 64\u201367.","journal-title":"Design and Applications"},{"volume-title":"International semantic web conference","author":"P\u00e9rez Jorge","key":"e_1_3_2_1_14_1","unstructured":"Jorge P\u00e9rez, Marcelo Arenas, and Claudio Gutierrez. 2006. Semantics and Complexity of SPARQL. In International semantic web conference. Springer, 30\u201343."},{"key":"e_1_3_2_1_15_1","volume-title":"100,000+ questions for machine comprehension of text. arXiv preprint arXiv:1606.05250","author":"Rajpurkar Pranav","year":"2016","unstructured":"Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. 2016. Squad: 100,000+ questions for machine comprehension of text. arXiv preprint arXiv:1606.05250 (2016)."},{"key":"e_1_3_2_1_16_1","unstructured":"Adam Roberts and Colin Raffel. 2020. Exploring transfer learning with t5: the text-to-text transfer transformer. Accessed on (2020) 23\u201307."},{"key":"e_1_3_2_1_17_1","volume-title":"a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108","author":"Sanh Victor","year":"2019","unstructured":"Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019)."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-tutorials.7"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2014.2303296"},{"key":"e_1_3_2_1_20_1","unstructured":"J Schulman B Zoph C Kim J Hilton J Menick J Weng JFC Uribe L Fedus L Metz M Pokorny 2022. ChatGPT: Optimizing language models for dialogue."},{"key":"e_1_3_2_1_21_1","volume-title":"Bidirectional attention flow for machine comprehension. arXiv preprint arXiv:1611.01603","author":"Seo Minjoon","year":"2016","unstructured":"Minjoon Seo, Aniruddha Kembhavi, Ali Farhadi, and Hannaneh Hajishirzi. 2016. Bidirectional attention flow for machine comprehension. arXiv preprint arXiv:1611.01603 (2016)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.physd.2019.132306"},{"key":"e_1_3_2_1_23_1","volume-title":"Commonsenseqa: A question answering challenge targeting commonsense knowledge. arXiv preprint arXiv:1811.00937","author":"Talmor Alon","year":"2018","unstructured":"Alon Talmor, Jonathan Herzig, Nicholas Lourie, and Jonathan Berant. 2018. Commonsenseqa: A question answering challenge targeting commonsense knowledge. arXiv preprint arXiv:1811.00937 (2018)."},{"key":"e_1_3_2_1_24_1","volume-title":"Lamda: Language models for dialog applications. arXiv preprint arXiv:2201.08239","author":"Thoppilan Romal","year":"2022","unstructured":"Romal Thoppilan, Daniel De\u00a0Freitas, Jamie Hall, Noam Shazeer, Apoorv Kulshreshtha, Heng-Tze Cheng, Alicia Jin, Taylor Bos, Leslie Baker, Yu Du, 2022. Lamda: Language models for dialog applications. arXiv preprint arXiv:2201.08239 (2022)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/2187836.2187923"},{"key":"e_1_3_2_1_26_1","volume-title":"Attention is all you need. Advances in neural information processing systems 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan\u00a0N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_27_1","volume-title":"Machine comprehension using match-lstm and answer pointer. arXiv preprint arXiv:1608.07905","author":"Wang Shuohang","year":"2016","unstructured":"Shuohang Wang and Jing Jiang. 2016. Machine comprehension using match-lstm and answer pointer. arXiv preprint arXiv:1608.07905 (2016)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/HPEC.2016.7761624"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/1499586.1499695"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.aiopen.2021.05.001"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/EMC2-NIPS53020.2019.00016"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-022-01744-y"},{"key":"e_1_3_2_1_33_1","volume-title":"Build a Robust QA System with Transformer-based Mixture of Experts. arXiv preprint arXiv:2204.09598","author":"Zhou Yu\u00a0Qing","year":"2022","unstructured":"Yu\u00a0Qing Zhou, Xixuan\u00a0Julie Liu, and Yuanzhe Dong. 2022. Build a Robust QA System with Transformer-based Mixture of Experts. arXiv preprint arXiv:2204.09598 (2022)."}],"event":{"name":"WWW '23: The ACM Web Conference 2023","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Austin TX USA","acronym":"WWW '23"},"container-title":["Companion Proceedings of the ACM Web Conference 2023"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543873.3587540","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3543873.3587540","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T23:36:28Z","timestamp":1755819388000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543873.3587540"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,30]]},"references-count":33,"alternative-id":["10.1145\/3543873.3587540","10.1145\/3543873"],"URL":"https:\/\/doi.org\/10.1145\/3543873.3587540","relation":{},"subject":[],"published":{"date-parts":[[2023,4,30]]},"assertion":[{"value":"2023-04-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}