{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T07:06:14Z","timestamp":1742972774692,"version":"3.40.3"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031159183"},{"type":"electronic","value":"9783031159190"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-15919-0_35","type":"book-chapter","created":{"date-parts":[[2022,9,6]],"date-time":"2022-09-06T04:04:59Z","timestamp":1662437099000},"page":"416-427","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Coarse-to-Fine Training Paradigm for\u00a0Dialogue Summarization"],"prefix":"10.1007","author":[{"given":"Zhiyue","family":"Liu","sequence":"first","affiliation":[]},{"given":"Zhaoyang","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Jiahai","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,7]]},"reference":[{"key":"35_CR1","doi-asserted-by":"crossref","unstructured":"Chen, J., Yang, D.: Multi-view sequence-to-sequence models with conversational structure for abstractive dialogue summarization. In: Proceedings of EMNLP, pp. 4106\u20134118. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.336"},{"key":"35_CR2","doi-asserted-by":"crossref","unstructured":"Chen, Y., Liu, Y., Chen, L., Zhang, Y.: DialogSum: a real-life scenario dialogue summarization dataset. In: Findings of ACL, pp. 5062\u20135074. Association for Computational Linguistics (2021)","DOI":"10.18653\/v1\/2021.findings-acl.449"},{"key":"35_CR3","doi-asserted-by":"crossref","unstructured":"Feng, X., Feng, X., Qin, B.: A survey on dialogue summarization: recent advances and new frontiers. arXiv preprint arXiv:2107.03175 (2021)","DOI":"10.24963\/ijcai.2022\/764"},{"key":"35_CR4","doi-asserted-by":"crossref","unstructured":"Feng, X., Feng, X., Qin, L., Qin, B., Liu, T.: Language model as an annotator: exploring DialoGPT for dialogue summarization. In: Proceedings of ACL-IJCNLP, pp. 1479\u20131491. Association for Computational Linguistics (2021)","DOI":"10.18653\/v1\/2021.acl-long.117"},{"key":"35_CR5","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of ICML. Proceedings of Machine Learning Research, vol. 70, pp. 1126\u20131135. PMLR (2017)"},{"key":"35_CR6","doi-asserted-by":"crossref","unstructured":"Gliwa, B., Mochol, I., Biesek, M., Wawer, A.: SAMSum corpus: a human-annotated dialogue dataset for abstractive summarization. In: Proceedings of the 2nd Workshop on New Frontiers in Summarization, pp. 70\u201379. Association for Computational Linguistics, Hong Kong, China (2019)","DOI":"10.18653\/v1\/D19-5409"},{"key":"35_CR7","doi-asserted-by":"crossref","unstructured":"Hardy, H., Vlachos, A.: Guided neural language generation for abstractive summarization using Abstract Meaning Representation. In: Proceedings of EMNLP, pp. 768\u2013773. Association for Computational Linguistics, Brussels, Belgium (2018)","DOI":"10.18653\/v1\/D18-1086"},{"key":"35_CR8","unstructured":"Hermann, K.M., et al.: Teaching machines to read and comprehend. In: Proceedings of NeurIPS, vol. 28. Curran Associates, Inc. (2015)"},{"key":"35_CR9","doi-asserted-by":"crossref","unstructured":"Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of ACL, pp. 7871\u20137880. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"35_CR10","doi-asserted-by":"crossref","unstructured":"Li, M., Zhang, L., Ji, H., Radke, R.J.: Keep meeting summaries on topic: Abstractive multi-modal meeting summarization. In: Proceedings of ACL, pp. 2190\u20132196. Association for Computational Linguistics, Florence, Italy (2019)","DOI":"10.18653\/v1\/P19-1210"},{"key":"35_CR11","unstructured":"Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74\u201381. Association for Computational Linguistics, Barcelona, Spain (2004)"},{"key":"35_CR12","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., Neubig, G.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. arXiv preprint arXiv:2107.13586 (2021)"},{"key":"35_CR13","doi-asserted-by":"crossref","unstructured":"Liu, Y., Liu, P.: SimCLS: a simple framework for contrastive learning of abstractive summarization. In: Proceedings of ACL-IJCNLP, pp. 1065\u20131072. Association for Computational Linguistics (2021)","DOI":"10.18653\/v1\/2021.acl-short.135"},{"key":"35_CR14","doi-asserted-by":"crossref","unstructured":"Liu, Z., Chen, N.: Reading turn by turn: hierarchical attention architecture for spoken dialogue comprehension. In: Proceedings of ACL, pp. 5460\u20135466. Association for Computational Linguistics, Florence, Italy (2019)","DOI":"10.18653\/v1\/P19-1543"},{"key":"35_CR15","doi-asserted-by":"crossref","unstructured":"Maynez, J., Narayan, S., Bohnet, B., McDonald, R.: On faithfulness and factuality in abstractive summarization. In: Proceedings of ACL, pp. 1906\u20131919. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.acl-main.173"},{"key":"35_CR16","unstructured":"Mihalcea, R., Tarau, P.: TextRank: bringing order into text. In: Proceedings of EMNLP, pp. 404\u2013411. Association for Computational Linguistics, Barcelona, Spain (2004)"},{"key":"35_CR17","doi-asserted-by":"crossref","unstructured":"Murray, G., Renals, S., Carletta, J., Moore, J.: Incorporating speaker and discourse features into speech summarization. In: Proceedings of NAACL, pp. 367\u2013374. Association for Computational Linguistics, New York City, USA (2006)","DOI":"10.3115\/1220835.1220882"},{"key":"35_CR18","doi-asserted-by":"crossref","unstructured":"Narayan, S., Cohen, S.B., Lapata, M.: Don\u2019t give me the details, just the summary! topic-aware convolutional neural networks for extreme summarization. In: Proceedings of EMNLP, pp. 1797\u20131807. Association for Computational Linguistics, Brussels, Belgium (2018)","DOI":"10.18653\/v1\/D18-1206"},{"key":"35_CR19","doi-asserted-by":"crossref","unstructured":"Qi, P., Zhang, Y., Zhang, Y., Bolton, J., Manning, C.D.: Stanza: a python natural language processing toolkit for many human languages. In: Proceedings of ACL, pp. 101\u2013108. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.acl-demos.14"},{"key":"35_CR20","unstructured":"Schick, T., Sch\u00fctze, H.: Few-shot text generation with pattern-exploiting training. arXiv preprint arXiv:2012.11926 (2020)"},{"key":"35_CR21","doi-asserted-by":"crossref","unstructured":"See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: Proceedings of ACL, pp. 1073\u20131083. Association for Computational Linguistics, Vancouver, Canada (2017)","DOI":"10.18653\/v1\/P17-1099"},{"key":"35_CR22","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Proceedings of NeurIPS, vol. 30. Curran Associates, Inc. (2017)"},{"key":"35_CR23","unstructured":"Yuan, W., Neubig, G., Liu, P.: BARTScore: evaluating generated text as text generation. arXiv preprint arXiv:2106.11520 (2021)"},{"key":"35_CR24","unstructured":"Zhang, T., Kishore, V., Wu, F., Weinberger, K.Q., Artzi, Y.: BERTScore: Evaluating text generation with BERT. In: Proceedings of ICLR (2020)"},{"key":"35_CR25","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: DIALOGPT : large-scale generative pre-training for conversational response generation. In: Proceedings of ACL, pp. 270\u2013278. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.acl-demos.30"},{"key":"35_CR26","doi-asserted-by":"crossref","unstructured":"Zhao, L., Xu, W., Guo, J.: Improving abstractive dialogue summarization with graph structures and topic words. In: Proceedings of COLING, pp. 437\u2013449. International Committee on Computational Linguistics, Barcelona, Spain (2020)","DOI":"10.18653\/v1\/2020.coling-main.39"},{"key":"35_CR27","doi-asserted-by":"crossref","unstructured":"Zhu, C., Xu, R., Zeng, M., Huang, X.: A hierarchical network for abstractive meeting summarization with cross-domain pretraining. In: Findings of EMNLP, pp. 194\u2013203. Association for Computational Linguistics (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.19"},{"key":"35_CR28","doi-asserted-by":"crossref","unstructured":"Zou, Y., et al.: Topic-oriented spoken dialogue summarization for customer service with saliency-aware topic modeling. In: Proceedings of AAAI, pp. 14665\u201314673 (2021)","DOI":"10.1609\/aaai.v35i16.17723"},{"key":"35_CR29","doi-asserted-by":"crossref","unstructured":"Zou, Y., Zhu, B., Hu, X., Gui, T., Zhang, Q.: Low-resource dialogue summarization with domain-agnostic multi-source pretraining. arXiv preprint arXiv:2109.04080 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.7"}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-15919-0_35","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,29]],"date-time":"2023-03-29T08:21:10Z","timestamp":1680078070000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-15919-0_35"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031159183","9783031159190"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-15919-0_35","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"7 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bristol","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2022\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"561","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"255","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}