{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T11:40:03Z","timestamp":1755862803469,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":27,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,2,2]],"date-time":"2024-02-02T00:00:00Z","timestamp":1706832000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,2,2]]},"DOI":"10.1145\/3651671.3651759","type":"proceedings-article","created":{"date-parts":[[2024,6,7]],"date-time":"2024-06-07T18:55:50Z","timestamp":1717786550000},"page":"609-615","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["QuDial: A Quadruple-driven Dialogue System for Real Estate Consulting Services"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-2384-6693","authenticated-orcid":false,"given":"Shuling","family":"Yang","sequence":"first","affiliation":[{"name":"School of Data Science, University of Science and Technology of China, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-4314-2599","authenticated-orcid":false,"given":"Hanzhu","family":"Chen","sequence":"additional","affiliation":[{"name":"School of Data Science, University of Science and Technology of China, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-5179-8250","authenticated-orcid":false,"given":"Binbin","family":"Fang","sequence":"additional","affiliation":[{"name":"School of Data Science, University of Science and Technology of China, China"}]}],"member":"320","published-online":{"date-parts":[[2024,6,7]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.108318"},{"key":"e_1_3_2_1_2_1","volume-title":"Building a role specified open-domain dialogue system leveraging large-scale language models. arXiv preprint arXiv:2205.00176","author":"Bae Sanghwan","year":"2022","unstructured":"Sanghwan Bae, Donghyun Kwak, Sungdong Kim, Donghoon Ham, Soyoung Kang, Sang-Woo Lee, and Woomyoung Park. 2022. Building a role specified open-domain dialogue system leveraging large-scale language models. arXiv preprint arXiv:2205.00176 (2022)."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.sigdial-1.25"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1547"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2021-618"},{"key":"e_1_3_2_1_6_1","volume-title":"A Research to Determine the Problems of Real Estate Consultants in the Housing Marketing Process. \u00c7ank\u0131r\u0131 Karatekin \u00dcniversitesi \u0130ktisadi ve \u0130dari Bilimler Fak\u00fcltesi Dergisi 12, 1","author":"Demirta\u015f Mehmet\u00a0Can","year":"2022","unstructured":"Mehmet\u00a0Can Demirta\u015f. 2022. A Research to Determine the Problems of Real Estate Consultants in the Housing Marketing Process. \u00c7ank\u0131r\u0131 Karatekin \u00dcniversitesi \u0130ktisadi ve \u0130dari Bilimler Fak\u00fcltesi Dergisi 12, 1 (2022), 104\u2013126."},{"key":"e_1_3_2_1_7_1","volume-title":"J urgen Schmidhuber, and Corso Elvezia","author":"Hochreiter Sepp","year":"1997","unstructured":"Sepp Hochreiter, J urgen Schmidhuber, and Corso Elvezia. 1997. Long short-term memory. Neural computation 9, 8 (1997), 1735\u20131780. https:\/\/ieeexplore.ieee.org\/abstract\/document\/6795963"},{"key":"e_1_3_2_1_8_1","volume-title":"Advances in Neural Information Processing Systems, H.\u00a0Larochelle, M.\u00a0Ranzato, R.\u00a0Hadsell, M.F. Balcan, and H.\u00a0Lin (Eds.). Vol.\u00a033. Curran Associates","author":"Hosseini-Asl Ehsan","year":"2017","unstructured":"Ehsan Hosseini-Asl, Bryan McCann, Chien-Sheng Wu, Semih Yavuz, and Richard Socher. 2020. A Simple Language Model for Task-Oriented Dialogue. In Advances in Neural Information Processing Systems, H.\u00a0Larochelle, M.\u00a0Ranzato, R.\u00a0Hadsell, M.F. Balcan, and H.\u00a0Lin (Eds.). Vol.\u00a033. Curran Associates, Inc., 20179\u201320191. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2020\/file\/e946209592563be0f01c844ab2170f0c-Paper.pdf"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2021.101310"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3571730"},{"key":"e_1_3_2_1_11_1","volume-title":"Proceedings of naacL-HLT, Vol.\u00a01. 2.","author":"Ming-Wei\u00a0Chang Jacob Devlin","year":"2019","unstructured":"Jacob Devlin Ming-Wei\u00a0Chang Kenton and Lee\u00a0Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of naacL-HLT, Vol.\u00a01. 2."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.53"},{"key":"e_1_3_2_1_13_1","volume-title":"5th International Conference on Learning Representations (ICLR)","author":"Kipf N","year":"2017","unstructured":"Thomas\u00a0N Kipf and Max Welling. 2017. Semi-supervised classification with graph convolutional networks. In 5th International Conference on Learning Representations (ICLR) (2017). https:\/\/arxiv.org\/abs\/1609.02907"},{"key":"e_1_3_2_1_14_1","volume-title":"International Conference on Learning Representations, ICLR 2020","author":"Lan Zhenzhong","year":"2019","unstructured":"Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. Albert: A lite bert for self-supervised learning of language representations. In International Conference on Learning Representations, ICLR 2020 (2019). https:\/\/openreview.net\/forum?id=H1eA7AEtvS"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P18-1133"},{"key":"e_1_3_2_1_16_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019). https:\/\/arxiv.org\/abs\/1907.11692"},{"key":"e_1_3_2_1_17_1","volume-title":"7th International Conference on Learning Representations, ICLR 2019","author":"Loshchilov Ilya","year":"2019","unstructured":"Ilya Loshchilov and Frank Hutter. 2019. Decoupled weight decay regularization. In 7th International Conference on Learning Representations, ICLR 2019 (2019). https:\/\/openreview.net\/forum?id=Bkg6RiCqY7"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3605943"},{"key":"e_1_3_2_1_19_1","volume-title":"Recent advances in deep learning based dialogue systems: A systematic survey. Artificial intelligence review 56, 4","author":"Ni Jinjie","year":"2023","unstructured":"Jinjie Ni, Tom Young, Vlad Pandelea, Fuzhao Xue, and Erik Cambria. 2023. Recent advances in deep learning based dialogue systems: A systematic survey. Artificial intelligence review 56, 4 (2023), 3055\u20133155."},{"key":"e_1_3_2_1_20_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog 1, 8","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9. https:\/\/cdn.openai.com\/better-language-models\/language_models_are_unsupervised_multitask_learners.pdf"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.59"},{"key":"e_1_3_2_1_22_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.458"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.66"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v31i1.11182"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.sigdial-1.34"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3617680"}],"event":{"name":"ICMLC 2024: 2024 16th International Conference on Machine Learning and Computing","acronym":"ICMLC 2024","location":"Shenzhen China"},"container-title":["Proceedings of the 2024 16th International Conference on Machine Learning and Computing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3651671.3651759","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3651671.3651759","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T11:18:44Z","timestamp":1755861524000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3651671.3651759"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2,2]]},"references-count":27,"alternative-id":["10.1145\/3651671.3651759","10.1145\/3651671"],"URL":"https:\/\/doi.org\/10.1145\/3651671.3651759","relation":{},"subject":[],"published":{"date-parts":[[2024,2,2]]},"assertion":[{"value":"2024-06-07","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}