{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T15:29:09Z","timestamp":1772206149393,"version":"3.50.1"},"publisher-location":"Cham","reference-count":32,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031667046","type":"print"},{"value":"9783031667053","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-66705-3_14","type":"book-chapter","created":{"date-parts":[[2024,8,22]],"date-time":"2024-08-22T11:23:00Z","timestamp":1724325780000},"page":"203-217","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Investigating a\u00a0Semantic Similarity Loss Function for\u00a0the\u00a0Parallel Training of\u00a0Abstractive and\u00a0Extractive Scientific Document Summarizers"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2640-6300","authenticated-orcid":false,"given":"Sudipta","family":"Singha Roy","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0080-715X","authenticated-orcid":false,"given":"Robert E.","family":"Mercer","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,21]]},"reference":[{"key":"14_CR1","unstructured":"Aksenov, D., Schneider, J.M., Bourgonje, P., Schwarzenberg, R., Hennig, L., Rehm, G.: Abstractive text summarization based on language model conditioning and locality modeling. In: Proceedings of the 12th Language Resources and Evaluation Conference, pp. 6680\u20136689 (2020)"},{"key":"14_CR2","doi-asserted-by":"crossref","unstructured":"Aralikatte, R., Narayan, S., Maynez, J., Rothe, S., McDonald, R.: Focus attention: promoting faithfulness and diversity in summarization. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 6078\u20136095 (2021)","DOI":"10.18653\/v1\/2021.acl-long.474"},{"key":"14_CR3","unstructured":"Beltagy, I., Peters, M.E., Cohan, A.: Longformer: the long-document transformer. arXiv preprint arXiv:2004.05150 (2020)"},{"key":"14_CR4","doi-asserted-by":"crossref","unstructured":"Cheng, J., Lapata, M.: Neural summarization by extracting sentences and words. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 484\u2013494 (2016)","DOI":"10.18653\/v1\/P16-1046"},{"key":"14_CR5","doi-asserted-by":"crossref","unstructured":"Cho, S., Song, K., Wang, X., Liu, F., Yu, D.: Toward unifying text segmentation and long document summarization. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 106\u2013118 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.8"},{"key":"14_CR6","doi-asserted-by":"crossref","unstructured":"Cohan, A., et al.: A discourse-aware attention model for abstractive summarization of long documents. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pp. 615\u2013621 (2018)","DOI":"10.18653\/v1\/N18-2097"},{"key":"14_CR7","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171\u20134186 (2019)"},{"key":"14_CR8","doi-asserted-by":"crossref","unstructured":"Fonseca, M., Ziser, Y., Cohen, S.B.: Factorizing content and budget decisions in abstractive summarization of long documents. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 6341\u20136364 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.426"},{"key":"14_CR9","doi-asserted-by":"crossref","unstructured":"Fu, X., Wang, J., Zhang, J., Wei, J., Yang, Z.: Document summarization with VHTM: variational hierarchical topic-aware mechanism. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a034, pp. 7740\u20137747 (2020)","DOI":"10.1609\/aaai.v34i05.6277"},{"key":"14_CR10","doi-asserted-by":"crossref","unstructured":"Huang, D., et al.: What have we achieved on text summarization? In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pp. 446\u2013469 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.33"},{"key":"14_CR11","unstructured":"Koo, T., Globerson, A., Carreras\u00a0P\u00e9rez, X., Collins, M.: Structured prediction models via the matrix-tree theorem. In: Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, pp. 141\u2013150 (2007)"},{"key":"14_CR12","doi-asserted-by":"crossref","unstructured":"Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871\u20137880 (2020)","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"14_CR13","doi-asserted-by":"crossref","unstructured":"Liu, Y., Lapata, M.: Text summarization with pretrained encoders. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3730\u20133740 (2019)","DOI":"10.18653\/v1\/D19-1387"},{"key":"14_CR14","doi-asserted-by":"crossref","unstructured":"Liu, Y., Titov, I., Lapata, M.: Single document summarization as tree induction. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 1745\u20131755 (2019)","DOI":"10.18653\/v1\/N19-1173"},{"key":"14_CR15","doi-asserted-by":"crossref","unstructured":"Liu, Y., et al.: Leveraging locality in abstractive text summarization. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 6081\u20136093 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.408"},{"key":"14_CR16","doi-asserted-by":"crossref","unstructured":"Mao, Z., et al.: DYLE: dynamic latent extraction for abstractive long-input summarization. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1687\u20131698 (2022)","DOI":"10.18653\/v1\/2022.acl-long.118"},{"key":"14_CR17","doi-asserted-by":"crossref","unstructured":"Nallapati, R., Zhai, F., Zhou, B.: SummaRuNNer: a recurrent neural network based sequence model for extractive summarization of documents. In: Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI 2017), pp. 3075\u20133081 (2017)","DOI":"10.1609\/aaai.v31i1.10958"},{"key":"14_CR18","doi-asserted-by":"crossref","unstructured":"Narayan, S., Cohen, S.B., Lapata, M.: Ranking sentences for extractive summarization with reinforcement learning. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp. 1747\u20131759 (2018)","DOI":"10.18653\/v1\/N18-1158"},{"key":"14_CR19","doi-asserted-by":"crossref","unstructured":"Qiu, Y., Cohen, S.B.: Abstractive summarization guided by latent hierarchical document structure. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 5303\u20135317 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.355"},{"key":"14_CR20","doi-asserted-by":"crossref","unstructured":"Ruan, Q., Ostendorff, M., Rehm, G.: HiStruct+: improving extractive text summarization with hierarchical structure information. In: Findings of the Association for Computational Linguistics: ACL 2022, pp. 1292\u20131308 (2022)","DOI":"10.18653\/v1\/2022.findings-acl.102"},{"key":"14_CR21","doi-asserted-by":"crossref","unstructured":"See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1073\u20131083 (2017)","DOI":"10.18653\/v1\/P17-1099"},{"key":"14_CR22","doi-asserted-by":"crossref","unstructured":"Singha\u00a0Roy, S., Mercer, R.E.: Generating extractive and abstractive summaries in parallel from scientific articles incorporating citing statements. In: Proceedings of the 4th New Frontiers in Summarization Workshop, pp. 75\u201386 (2023)","DOI":"10.18653\/v1\/2023.newsum-1.8"},{"key":"14_CR23","doi-asserted-by":"crossref","unstructured":"Wang, D., Liu, P., Zheng, Y., Qiu, X., Huang, X.: Heterogeneous graph neural networks for extractive document summarization. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 6209\u20136219 (2020)","DOI":"10.18653\/v1\/2020.acl-main.553"},{"key":"14_CR24","doi-asserted-by":"crossref","unstructured":"Wang, F., et al.: Salience allocation as guidance for abstractive summarization. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 6094\u20136106 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.409"},{"key":"14_CR25","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al.: Friendly topic assistant for transformer based abstractive summarization. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pp. 485\u2013497 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.35"},{"key":"14_CR26","unstructured":"Xie, Q., Huang, J., Saha, T., Ananiadou, S.: GRETEL: graph contrastive topic enhanced language model for long document extractive summarization. In: Proceedings of the 29th International Conference on Computational Linguistics, pp. 6259\u20136269 (2022)"},{"key":"14_CR27","doi-asserted-by":"crossref","unstructured":"Yu, T., Su, D., Dai, W., Fung, P.: Dimsum @LaySumm 20: BART-based approach for scientific document summarization. In: Proceedings of the First Workshop on Scholarly Document Processing, pp. 303\u2013309 (2020)","DOI":"10.18653\/v1\/2020.sdp-1.35"},{"key":"14_CR28","unstructured":"Zaheer, M., et al.: Big bird: transformers for longer sequences. In: Advances in Neural Information Processing Systems, vol. 33, pp. 17283\u201317297 (2020)"},{"key":"14_CR29","doi-asserted-by":"crossref","unstructured":"Zhang, H., Liu, X., Zhang, J.: HEGEL: hypergraph transformer for long document summarization. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, pp. 10167\u201310176 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.692"},{"key":"14_CR30","doi-asserted-by":"crossref","unstructured":"Zhang, X., Wei, F., Zhou, M.: HIBERT: document level pre-training of hierarchical bidirectional transformers for document summarization. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 5059\u20135069 (2019)","DOI":"10.18653\/v1\/P19-1499"},{"key":"14_CR31","doi-asserted-by":"crossref","unstructured":"Zhong, M., Liu, P., Chen, Y., Wang, D., Qiu, X., Huang, X.J.: Extractive summarization as text matching. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 6197\u20136208 (2020)","DOI":"10.18653\/v1\/2020.acl-main.552"},{"key":"14_CR32","doi-asserted-by":"crossref","unstructured":"Zhou, Q., Yang, N., Wei, F., Huang, S., Zhou, M., Zhao, T.: Neural document summarization by jointly learning to score and select sentences. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 654\u2013663 (2018)","DOI":"10.18653\/v1\/P18-1061"}],"container-title":["Communications in Computer and Information Science","Deep Learning Theory and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-66705-3_14","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,22]],"date-time":"2024-08-22T11:51:36Z","timestamp":1724327496000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-66705-3_14"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031667046","9783031667053"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-66705-3_14","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"21 August 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DeLTA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Deep Learning Theory and Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Dijon","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"France","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10 July 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 July 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"delta2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/delta.scitevents.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}