{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T06:12:22Z","timestamp":1743142342738,"version":"3.40.3"},"publisher-location":"Cham","reference-count":35,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031301049"},{"type":"electronic","value":"9783031301056"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-30105-6_33","type":"book-chapter","created":{"date-parts":[[2023,4,12]],"date-time":"2023-04-12T20:31:55Z","timestamp":1681331515000},"page":"393-405","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Factual Error Correction in\u00a0Summarization with\u00a0Retriever-Reader Pipeline"],"prefix":"10.1007","author":[{"given":"Weiwei","family":"Li","sequence":"first","affiliation":[]},{"given":"Junzhuo","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Hui","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,13]]},"reference":[{"key":"33_CR1","unstructured":"Lin, C.Y.: Rouge: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74\u201381 (2004)"},{"key":"33_CR2","doi-asserted-by":"crossref","unstructured":"Kry\u015bci\u0144ski, W., McCann, B., Xiong, C., Socher, R.: Evaluating the factual consistency of abstractive text summarization. In: EMNLP, pp. 9332\u20139346 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.750"},{"key":"33_CR3","doi-asserted-by":"crossref","unstructured":"Cao, Z., Wei, F., Li, W., Li, S.: Faithful to the original: fact aware neural abstractive summarization. In: AAAI (2018)","DOI":"10.1609\/aaai.v32i1.11912"},{"key":"33_CR4","doi-asserted-by":"crossref","unstructured":"Maynez, J., Narayan, S., Bohnet, B., McDonald, R.: On faithfulness and factuality in abstractive summarization. In: ACL, pp. 1906\u20131919 (2020)","DOI":"10.18653\/v1\/2020.acl-main.173"},{"key":"33_CR5","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Cohen, S.B., Webber, B.: Reducing quantity hallucinations in abstractive summarization. In: EMNLP(Findings), pp. 2237\u20132249 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.203"},{"key":"33_CR6","doi-asserted-by":"crossref","unstructured":"Aralikatte, R., Narayan, S., Maynez, J., Rothe, S., McDonald, R.: Focus attention: promoting faithfulness and diversity in summarization. In: ACL-IJCNLP, pp. 6078\u20136095 (2021)","DOI":"10.18653\/v1\/2021.acl-long.474"},{"key":"33_CR7","doi-asserted-by":"crossref","unstructured":"Nan, F., et al.: Improving factual consistency of abstractive summarization via question answering. In: ACL-IJCNLP, pp. 6881\u20136894 (2021)","DOI":"10.18653\/v1\/2021.acl-long.536"},{"key":"33_CR8","doi-asserted-by":"crossref","unstructured":"Chen, S., Zhang, F., Sone, K., Roth, D.: Improving faithfulness in abstractive summarization with contrast candidate generation and selection. In: NAACL, pp. 5935\u20135941 (2021)","DOI":"10.18653\/v1\/2021.naacl-main.475"},{"key":"33_CR9","doi-asserted-by":"crossref","unstructured":"Huang, L., Wu, L., Wang, L.: Knowledge graph-augmented abstractive summarization with semantic-driven cloze reward. In: ACL, pp. 5094\u20135107 (2020)","DOI":"10.18653\/v1\/2020.acl-main.457"},{"issue":"3","key":"33_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2020.102478","volume":"58","author":"M Zhang","year":"2021","unstructured":"Zhang, M., Zhou, G., Yu, W., Liu, W.: Far-ass: fact-aware reinforced abstractive sentence summarization. Inf. Process. Manage. 58(3), 102478 (2021)","journal-title":"Inf. Process. Manage."},{"key":"33_CR11","doi-asserted-by":"crossref","unstructured":"Dong, Y., Wang, S., Gan, Z., Cheng, Y., Cheung, J.C.K., Liu, J.: Multi-fact correction in abstractive text summarization. In: EMNLP, pp. 9320\u20139331 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.749"},{"key":"33_CR12","doi-asserted-by":"crossref","unstructured":"Cao, M., Dong, Y., Wu, J., Cheung, J.C.K.: Factual error correction for abstractive summarization models. In: EMNLP, pp. 6251\u20136258 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.506"},{"key":"33_CR13","doi-asserted-by":"crossref","unstructured":"Lewis, M., et al.: Bart: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: ACL, pp. 7871\u20137880 (2020)","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"33_CR14","doi-asserted-by":"crossref","unstructured":"Sen, P., Saffari, A.: What do models learn from question answering datasets? In: EMNLP, pp. 2429\u20132438 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.190"},{"key":"33_CR15","doi-asserted-by":"crossref","unstructured":"Ko, M., Lee, J., Kim, H., Kim, G., Kang, J.: Look at the first sentence: position bias in question answering. In: EMNLP, pp. 1109\u20131121 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.84"},{"key":"33_CR16","doi-asserted-by":"crossref","unstructured":"Falke, T., Ribeiro, L.F., Utama, P.A., Dagan, I., Gurevych, I.: Ranking generated summaries by correctness: an interesting but challenging application for natural language inference. In: ACL, pp. 2214\u20132220 (2019)","DOI":"10.18653\/v1\/P19-1213"},{"key":"33_CR17","doi-asserted-by":"crossref","unstructured":"Cao, S., Wang, L.: Cliff: contrastive learning for improving faithfulness and factuality in abstractive summarization. In: EMNLP, pp. 6633\u20136649 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.532"},{"key":"33_CR18","unstructured":"Zhang, J., Zhao, Y., Saleh, M., Liu, P.: Pegasus: pre-training with extracted gap-sentences for abstractive summarization. In: ICML, pp. 11328\u201311339. PMLR (2020)"},{"key":"33_CR19","doi-asserted-by":"crossref","unstructured":"Zhu, C., et al.: Enhancing factual consistency of abstractive summarization. In: NAACL, pp. 718\u2013733 (2021)","DOI":"10.18653\/v1\/2021.naacl-main.58"},{"key":"33_CR20","doi-asserted-by":"crossref","unstructured":"Lee, D., et al.: Capturing speaker incorrectness: speaker-focused post-correction for abstractive dialogue summarization. In: EMNLP(newsum), pp. 65\u201373 (2021)","DOI":"10.18653\/v1\/2021.newsum-1.8"},{"key":"33_CR21","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: NAACL, pp. 4171\u20134186 (2019)"},{"key":"33_CR22","doi-asserted-by":"crossref","unstructured":"Xie, S., Tu, Z.: Holistically-nested edge detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1395\u20131403 (2015)","DOI":"10.1109\/ICCV.2015.164"},{"key":"33_CR23","doi-asserted-by":"crossref","unstructured":"Papineni, K., Roukos, S., Ward, T., Zhu, W.J.: Bleu: a method for automatic evaluation of machine translation. In: ACL, pp. 311\u2013318 (2002)","DOI":"10.3115\/1073083.1073135"},{"key":"33_CR24","doi-asserted-by":"crossref","unstructured":"Nallapati, R., Zhou, B., dos Santos, C., Gul\u00e7ehre, \u00c7., Xiang, B.: Abstractive text summarization using sequence-to-sequence rnns and beyond. In: CoNLL, pp. 280\u2013290 (2016)","DOI":"10.18653\/v1\/K16-1028"},{"issue":"1","key":"33_CR25","first-page":"34","volume":"4","author":"D Graff","year":"2003","unstructured":"Graff, D., Kong, J., Chen, K., Maeda, K.: English gigaword. Linguistic Data Consortium, Philadelphia 4(1), 34 (2003)","journal-title":"Linguistic Data Consortium, Philadelphia"},{"key":"33_CR26","doi-asserted-by":"crossref","unstructured":"Gehrmann, S., Deng, Y., Rush, A.M.: Bottom-up abstractive summarization. In: EMNLP, pp. 4098\u20134109 (2018)","DOI":"10.18653\/v1\/D18-1443"},{"key":"33_CR27","doi-asserted-by":"crossref","unstructured":"See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: ACL, pp. 1073\u20131083 (2017)","DOI":"10.18653\/v1\/P17-1099"},{"key":"33_CR28","doi-asserted-by":"crossref","unstructured":"Song, K., et al.: Joint parsing and generation for abstractive summarization. In: AAAI, vol. 34, pp. 8894\u20138901 (2020)","DOI":"10.1609\/aaai.v34i05.6419"},{"key":"33_CR29","doi-asserted-by":"crossref","unstructured":"Liu, Y., Lapata, M.: Text summarization with pretrained encoders. In: EMNLP-IJCNLP, pp. 3730\u20133740 (2019)","DOI":"10.18653\/v1\/D19-1387"},{"key":"33_CR30","doi-asserted-by":"crossref","unstructured":"Laban, P., Hsi, A., Canny, J., Hearst, M.A.: The summary loop: learning to write abstractive summaries without examples. In: ACL, pp. 5135\u20135150 (2020)","DOI":"10.18653\/v1\/2020.acl-main.460"},{"key":"33_CR31","doi-asserted-by":"crossref","unstructured":"Shah, D., Schuster, T., Barzilay, R.: Automatic fact-guided sentence modification. In: AAAI. 34, pp. 8791\u20138798 (2020)","DOI":"10.1609\/aaai.v34i05.6406"},{"key":"33_CR32","unstructured":"Wolf, T., et al.: Transformers: state-of-the-art natural language processing. In: EMNLP, pp. 38\u201345 (2020)"},{"key":"33_CR33","unstructured":"Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learning library. NIPS 32 (2019)"},{"key":"33_CR34","doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: Squad: 100,000+ questions for machine comprehension of text. In: EMNLP, pp. 2383\u20132392 (2016)","DOI":"10.18653\/v1\/D16-1264"},{"key":"33_CR35","unstructured":"Loshchilov, I., Hutter, F.: Fixing weight decay regularization in adam. arxiv 2017. arXiv preprint arXiv:1711.05101"}],"container-title":["Lecture Notes in Computer Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-30105-6_33","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,12]],"date-time":"2023-04-12T20:36:39Z","timestamp":1681331799000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-30105-6_33"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031301049","9783031301056"],"references-count":35,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-30105-6_33","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"13 April 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Delhi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iconip2022.apnns.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy Chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"810","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"359","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"44% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.65","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ICONIP 2022 consists of a two-volume set, LNCS & CCIS, which includes 146 and 213 papers","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}