{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,5]],"date-time":"2025-04-05T04:11:16Z","timestamp":1743826276388,"version":"3.40.3"},"publisher-location":"Cham","reference-count":26,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031887192","type":"print"},{"value":"9783031887208","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-88720-8_28","type":"book-chapter","created":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T12:08:23Z","timestamp":1743768503000},"page":"171-177","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Hierarchical Prefixes for\u00a0Long Document Representations"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-2501-8124","authenticated-orcid":false,"given":"Iskandar","family":"Boucharenc","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,3]]},"reference":[{"key":"28_CR1","doi-asserted-by":"crossref","unstructured":"Beltagy, I., Lo, K., Cohan, A.: SciBERT: a pretrained language model for scientific text (2019). https:\/\/arxiv.org\/abs\/1903.10676","DOI":"10.18653\/v1\/D19-1371"},{"key":"28_CR2","unstructured":"Chen, Y., et al.: LongloRA: efficient fine-tuning of long-context large language models. In: The Twelfth International Conference on Learning Representations (2024). https:\/\/openreview.net\/forum?id=6PmJoRfdaK"},{"key":"28_CR3","doi-asserted-by":"publisher","unstructured":"Cho, K., et al.: Learning phrase representations using RNN encoder\u2013decoder for statistical machine translation. In: Moschitti, A., Pang, B., Daelemans, W. (eds.) Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1724\u20131734. Association for Computational Linguistics, Doha, Qatar (2014). https:\/\/doi.org\/10.3115\/v1\/D14-1179, https:\/\/aclanthology.org\/D14-1179","DOI":"10.3115\/v1\/D14-1179"},{"key":"28_CR4","doi-asserted-by":"publisher","unstructured":"Czinczoll, T., H\u00f6nes, C., Schall, M., De\u00a0Melo, G.: NextLevelBERT: masked language modeling with higher-level representations for long documents. In: Ku, L.W., Martins, A., Srikumar, V. (eds.) Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 4656\u20134666. Association for Computational Linguistics, Bangkok, Thailand (2024). https:\/\/doi.org\/10.18653\/v1\/2024.acl-long.256, https:\/\/aclanthology.org\/2024.acl-long.256","DOI":"10.18653\/v1\/2024.acl-long.256"},{"key":"28_CR5","doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Burstein, J., Doran, C., Solorio, T. (eds.) Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171\u20134186. Association for Computational Linguistics, Minneapolis, Minnesota (2019). https:\/\/doi.org\/10.18653\/v1\/N19-1423, https:\/\/aclanthology.org\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"28_CR6","doi-asserted-by":"publisher","unstructured":"Guo, J., Fan, Y., Ai, Q., Croft, W.B.: A deep relevance matching model for ad-hoc retrieval. In: Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, pp. 55\u201364. CIKM \u201916, Association for Computing Machinery, New York, NY, USA (2016). https:\/\/doi.org\/10.1145\/2983323.2983769","DOI":"10.1145\/2983323.2983769"},{"key":"28_CR7","unstructured":"Hu, B., Lu, Z., Li, H., Chen, Q.: Convolutional neural network architectures for matching natural language sentences. In: Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2, pp. 2042\u20132050. NIPS\u201914, MIT Press, Cambridge, MA, USA (2014)"},{"key":"28_CR8","unstructured":"Hu, E.J., et al.: LoRA: low-rank adaptation of large language models. In: International Conference on Learning Representations (2022). https:\/\/openreview.net\/forum?id=nZeVKeeFYf9"},{"key":"28_CR9","doi-asserted-by":"publisher","unstructured":"Huang, P.S., He, X., Gao, J., Deng, L., Acero, A., Heck, L.: Learning deep structured semantic models for web search using clickthrough data. In: Proceedings of the 22nd ACM International Conference on Information & Knowledge Management, pp. 2333\u20132338. CIKM \u201913, Association for Computing Machinery, New York, NY, USA (2013). https:\/\/doi.org\/10.1145\/2505515.2505665","DOI":"10.1145\/2505515.2505665"},{"key":"28_CR10","doi-asserted-by":"publisher","unstructured":"Karpukhin, V., et al.: Dense passage retrieval for open-domain question answering. In: Webber, B., Cohn, T., He, Y., Liu, Y. (eds.) Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 6769\u20136781. Association for Computational Linguistics, Online (2020). https:\/\/doi.org\/10.18653\/v1\/2020.emnlp-main.550, https:\/\/aclanthology.org\/2020.emnlp-main.550","DOI":"10.18653\/v1\/2020.emnlp-main.550"},{"issue":"1","key":"28_CR11","doi-asserted-by":"crossref","first-page":"79","DOI":"10.1214\/aoms\/1177729694","volume":"22","author":"S Kullback","year":"1951","unstructured":"Kullback, S., Leibler, R.A.: On information and sufficiency. Ann. Math. Stat. 22(1), 79\u201386 (1951)","journal-title":"Ann. Math. Stat."},{"key":"28_CR12","doi-asserted-by":"publisher","unstructured":"Lester, B., Al-Rfou, R., Constant, N.: The power of scale for parameter-efficient prompt tuning. In: Moens, M.F., Huang, X., Specia, L., Yih, S.W.t. (eds.) Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 3045\u20133059. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic (2021). https:\/\/doi.org\/10.18653\/v1\/2021.emnlp-main.243, https:\/\/aclanthology.org\/2021.emnlp-main.243","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"key":"28_CR13","unstructured":"Li, T., Zhang, G., Do, Q.D., Yue, X., Chen, W.: Long-context LLMs struggle with long in-context learning. CoRR abs\/2404.02060 (2024)"},{"key":"28_CR14","doi-asserted-by":"publisher","unstructured":"Li, X.L., Liang, P.: Prefix-tuning: optimizing continuous prompts for generation. In: Zong, C., Xia, F., Li, W., Navigli, R. (eds.) Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 4582\u20134597. Association for Computational Linguistics, Online (2021). https:\/\/doi.org\/10.18653\/v1\/2021.acl-long.353, https:\/\/aclanthology.org\/2021.acl-long.353","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"28_CR15","doi-asserted-by":"publisher","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., Neubig, G.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9) (2023). https:\/\/doi.org\/10.1145\/3560815","DOI":"10.1145\/3560815"},{"key":"28_CR16","doi-asserted-by":"publisher","unstructured":"Liu, X., et al.: P-tuning: prompt tuning can be comparable to fine-tuning across scales and tasks. In: Muresan, S., Nakov, P., Villavicencio, A. (eds.) Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 61\u201368. Association for Computational Linguistics, Dublin, Ireland (2022). https:\/\/doi.org\/10.18653\/v1\/2022.acl-short.8, https:\/\/aclanthology.org\/2022.acl-short.8","DOI":"10.18653\/v1\/2022.acl-short.8"},{"key":"28_CR17","unstructured":"Mitchell, J., Lapata, M.: Vector-based models of semantic composition. In: Moore, J.D., Teufel, S., Allan, J., Furui, S. (eds.) Proceedings of ACL-08: HLT, pp. 236\u2013244. Association for Computational Linguistics, Columbus, Ohio (2008). https:\/\/aclanthology.org\/P08-1028"},{"key":"28_CR18","doi-asserted-by":"publisher","unstructured":"Muennighoff, N., Tazi, N., Magne, L., Reimers, N.: MTEB: massive text embedding benchmark. In: Vlachos, A., Augenstein, I. (eds.) Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, pp. 2014\u20132037. Association for Computational Linguistics, Dubrovnik, Croatia (2023). https:\/\/doi.org\/10.18653\/v1\/2023.eacl-main.148, https:\/\/aclanthology.org\/2023.eacl-main.148","DOI":"10.18653\/v1\/2023.eacl-main.148"},{"key":"28_CR19","doi-asserted-by":"publisher","unstructured":"Peters, M.E., et al.: Deep contextualized word representations. In: Walker, M., Ji, H., Stent, A. (eds.) Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp. 2227\u20132237. Association for Computational Linguistics, New Orleans, Louisiana (2018). https:\/\/doi.org\/10.18653\/v1\/N18-1202, https:\/\/aclanthology.org\/N18-1202","DOI":"10.18653\/v1\/N18-1202"},{"key":"28_CR20","doi-asserted-by":"publisher","unstructured":"Reimers, N., Gurevych, I.: Sentence-BERT: sentence embeddings using Siamese BERT-networks. In: Inui, K., Jiang, J., Ng, V., Wan, X. (eds.) Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3982\u20133992. Association for Computational Linguistics, Hong Kong, China (2019). https:\/\/doi.org\/10.18653\/v1\/D19-1410, https:\/\/aclanthology.org\/D19-1410","DOI":"10.18653\/v1\/D19-1410"},{"key":"28_CR21","unstructured":"Saad-Falcon, J., Fu, D.Y., Arora, S., Guha, N., Re, C.: Benchmarking and building long-context retrieval models with LoCo and m2-BERT. In: Salakhutdinov, R., et al. (eds.) Proceedings of the 41st International Conference on Machine Learning. Proceedings of Machine Learning Research, vol.\u00a0235, pp. 42918\u201342946. PMLR (2024). https:\/\/proceedings.mlr.press\/v235\/saad-falcon24a.html"},{"key":"28_CR22","unstructured":"Thakur, N., Reimers, N., R\u00fcckl\u00e9, A., Srivastava, A., Gurevych, I.: BEIR: a heterogeneous benchmark for zero-shot evaluation of information retrieval models. In: Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2) (2021). https:\/\/openreview.net\/forum?id=wCu6T5xFjeJ"},{"key":"28_CR23","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Guyon, I., et al. (eds.) Advances in Neural Information Processing Systems. vol.\u00a030. Curran Associates, Inc. (2017)"},{"key":"28_CR24","unstructured":"Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., Bowman, S.R.: GLUE: a multi-task benchmark and analysis platform for natural language understanding. In: International Conference on Learning Representations (2019). https:\/\/openreview.net\/forum?id=rJ4km2R5t7"},{"key":"28_CR25","doi-asserted-by":"publisher","unstructured":"Zhang, X., Wei, F., Zhou, M.: HIBERT: document level pre-training of hierarchical bidirectional transformers for document summarization. In: Korhonen, A., Traum, D., M\u00e0rquez, L. (eds.) Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 5059\u20135069. Association for Computational Linguistics, Florence, Italy (2019). https:\/\/doi.org\/10.18653\/v1\/P19-1499, https:\/\/aclanthology.org\/P19-1499","DOI":"10.18653\/v1\/P19-1499"},{"key":"28_CR26","doi-asserted-by":"publisher","unstructured":"Zhao, L., et al.: Domain-oriented prefix-tuning: towards efficient and generalizable fine-tuning for zero-shot dialogue summarization. In: Carpuat, M., de\u00a0Marneffe, M.C., Meza\u00a0Ruiz, I.V. (eds.) Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4848\u20134862. Association for Computational Linguistics, Seattle, United States (2022). https:\/\/doi.org\/10.18653\/v1\/2022.naacl-main.357, https:\/\/aclanthology.org\/2022.naacl-main.357","DOI":"10.18653\/v1\/2022.naacl-main.357"}],"container-title":["Lecture Notes in Computer Science","Advances in Information Retrieval"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-88720-8_28","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T12:08:33Z","timestamp":1743768513000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-88720-8_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031887192","9783031887208"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-88720-8_28","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"3 April 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECIR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Information Retrieval","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Lucca","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 April 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 April 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"47","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecir2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ecir2025.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}