{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T14:43:44Z","timestamp":1773326624542,"version":"3.50.1"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031442155","type":"print"},{"value":"9783031442162","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-44216-2_9","type":"book-chapter","created":{"date-parts":[[2023,9,21]],"date-time":"2023-09-21T07:02:58Z","timestamp":1695279778000},"page":"99-114","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["CopiFilter: An Auxiliary Module Adapts Pre-trained Transformers for\u00a0Medical Dialogue Summarization"],"prefix":"10.1007","author":[{"given":"Jiaxin","family":"Duan","sequence":"first","affiliation":[]},{"given":"Junfei","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,22]]},"reference":[{"key":"9_CR1","doi-asserted-by":"crossref","unstructured":"Cho, K., et al.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. In: EMNLP 2014, pp. 1724\u20131734 (2014)","DOI":"10.3115\/v1\/D14-1179"},{"key":"9_CR2","unstructured":"Deaton, J., Jacobs, A., Kenealy, K., See, A.: Transformers and pointer-generator networks for abstractive summarization (2019)"},{"key":"9_CR3","unstructured":"Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: NAACL-HLT 2019, pp. 4171\u20134186 (2019)"},{"key":"9_CR4","doi-asserted-by":"crossref","unstructured":"Enarvi, S., et al.: Generating medical reports from patient-doctor conversations using sequence-to-sequence models. In: Proceedings of the First Workshop on Natural Language Processing for Medical Conversations, pp. 22\u201330 (2020)","DOI":"10.18653\/v1\/2020.nlpmc-1.4"},{"key":"9_CR5","doi-asserted-by":"crossref","unstructured":"Feng, X., Feng, X., Qin, B.: A survey on dialogue summarization: recent advances and new frontiers. In: IJCAI 2022, pp. 5453\u20135460 (2022)","DOI":"10.24963\/ijcai.2022\/764"},{"key":"9_CR6","doi-asserted-by":"crossref","unstructured":"Joshi, A., Katariya, N., Amatriain, X., Kannan, A.: Dr.summarize: global summarization of medical dialogue by exploiting local structures. In: EMNLP 2020, pp. 3755\u20133763 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.335"},{"key":"9_CR7","doi-asserted-by":"crossref","unstructured":"Kazi, N., Kahanda, I.: Automatically generating psychiatric case notes from digital transcripts of doctor-patient conversations. In: Proceedings of the 2nd Clinical Natural Language Processing Workshop, pp. 140\u2013148 (2019)","DOI":"10.18653\/v1\/W19-1918"},{"key":"9_CR8","doi-asserted-by":"crossref","unstructured":"Krishna, K., Khosla, S., Bigham, J.P., Lipton, Z.C.: Generating SOAP notes from doctor-patient conversations using modular summarization techniques. In: ACL\/IJCNLP 2021, pp. 4958\u20134972 (2021)","DOI":"10.18653\/v1\/2021.acl-long.384"},{"key":"9_CR9","doi-asserted-by":"crossref","unstructured":"Lewis, M., Liu, Y., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: ACL 2020, pp. 7871\u20137880 (2020)","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"9_CR10","unstructured":"Lin, C.Y.: Rouge: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74\u201381 (2004)"},{"key":"9_CR11","doi-asserted-by":"crossref","unstructured":"Liu, Y., Lapata, M.: Text summarization with pretrained encoders. In: EMNLP-IJCNLP 2019, pp. 3728\u20133738 (2019)","DOI":"10.18653\/v1\/D19-1387"},{"key":"9_CR12","doi-asserted-by":"crossref","unstructured":"Liu, Z., Ng, A., Guang, S.L.S., Aw, A.T., Chen, N.F.: Topic-aware pointer-generator networks for summarizing spoken conversations. In: ASRU 2019, pp. 814\u2013821 (2019)","DOI":"10.1109\/ASRU46091.2019.9003764"},{"issue":"29","key":"9_CR13","doi-asserted-by":"publisher","first-page":"861","DOI":"10.21105\/joss.00861","volume":"3","author":"L McInnes","year":"2018","unstructured":"McInnes, L., Healy, J., Saul, N., Grossberger, L.: UMAP: uniform manifold approximation and projection. J. Open Source Softw. 3(29), 861 (2018)","journal-title":"J. Open Source Softw."},{"key":"9_CR14","doi-asserted-by":"crossref","unstructured":"Qi, M., Liu, H., Fu, Y., Liu, T.: Improving abstractive dialogue summarization with hierarchical pretraining and topic segment. In: EMNLP 2021, pp. 1121\u20131130 (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.97"},{"key":"9_CR15","doi-asserted-by":"crossref","unstructured":"Qi, W., et al.: ProphetNet: predicting future n-gram for sequence-to-sequence pre-training. In: EMNLP 2020, pp. 2401\u20132410 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.217"},{"key":"9_CR16","unstructured":"Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. CoRR abs\/1910.10683 (2019)"},{"key":"9_CR17","doi-asserted-by":"crossref","unstructured":"See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: ACL 2017, pp. 1073\u20131083 (2017)","DOI":"10.18653\/v1\/P17-1099"},{"key":"9_CR18","unstructured":"Shao, Y., et al.: CPT: a pre-trained unbalanced transformer for both Chinese language understanding and generation. arXiv preprint arXiv:2109.05729 (2021)"},{"key":"9_CR19","unstructured":"Song, K., Tan, X., Qin, T., Lu, J., Liu, T.: MASS: masked sequence to sequence pre-training for language generation. In: ICML 2019, pp. 5926\u20135936 (2019)"},{"key":"9_CR20","doi-asserted-by":"crossref","unstructured":"Song, Y., Tian, Y., Wang, N., Xia, F.: Summarizing medical conversations via identifying important utterances. In: COLING 2020, pp. 717\u2013729 (2020)","DOI":"10.18653\/v1\/2020.coling-main.63"},{"key":"9_CR21","unstructured":"Souza, F., Nogueira, R., Lotufo, R.: Portuguese named entity recognition using BERT-CRF. arXiv preprint arXiv:1909.10649 (2019)"},{"key":"9_CR22","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NeurIPS 2017, pp. 5998\u20136008 (2017)"},{"key":"9_CR23","unstructured":"Wang, J., et al.: Fengshenbang 1.0: being the foundation of Chinese cognitive intelligence. CoRR abs\/2209.02970 (2022)"},{"key":"9_CR24","unstructured":"Yuan, L., Yu, Z.: Abstractive dialog summarization with semantic scaffolds. CoRR abs\/1910.00825 (2019). http:\/\/arxiv.org\/abs\/1910.00825"},{"key":"9_CR25","unstructured":"Zhang, J., Zhao, Y., Saleh, M., Liu, P.J.: PEGASUS: pre-training with extracted gap-sentences for abstractive summarization. In: ICML 2020, pp. 11328\u201311339 (2020)"},{"key":"9_CR26","unstructured":"Zhang, N., Chen, M., Bi, Z., Liang, X., et al.: CBLUE: a Chinese biomedical language understanding evaluation benchmark. In: ACL 2022, pp. 7888\u20137915 (2022)"},{"key":"9_CR27","doi-asserted-by":"crossref","unstructured":"Zhang, N., Deng, S., Li, J., Chen, X., Zhang, W., Chen, H.: Summarizing Chinese medical answer with graph convolution networks and question-focused dual attention. In: EMNLP 2020, pp. 15\u201324 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.2"},{"key":"9_CR28","unstructured":"Zhang, Z., et al.: Mengzi: towards lightweight yet ingenious pre-trained models for Chinese (2021)"},{"key":"9_CR29","doi-asserted-by":"crossref","unstructured":"Zhong, M., Liu, Y., Xu, Y., Zhu, C., Zeng, M.: DialogLM: pre-trained model for long dialogue understanding and summarization. In: AAAI, pp. 11765\u201311773 (2022)","DOI":"10.1609\/aaai.v36i10.21432"},{"key":"9_CR30","doi-asserted-by":"crossref","unstructured":"Zhu, C., et al.: Enhancing factual consistency of abstractive summarization. In: NAACL-HLT 2021, pp. 718\u2013733 (2021)","DOI":"10.18653\/v1\/2021.naacl-main.58"}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-44216-2_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,21]],"date-time":"2023-09-21T07:03:47Z","timestamp":1695279827000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-44216-2_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031442155","9783031442162"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-44216-2_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"22 September 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Heraklion","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"32","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"easyacademia.org","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"947","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"426","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"22","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.4","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"type of other papers accepted  : 9 Abstract","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}