{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,29]],"date-time":"2025-12-29T11:29:10Z","timestamp":1767007750724,"version":"3.44.0"},"publisher-location":"Cham","reference-count":14,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783032025500"},{"type":"electronic","value":"9783032025517"}],"license":[{"start":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:00:00Z","timestamp":1755820800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:00:00Z","timestamp":1755820800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-02551-7_17","type":"book-chapter","created":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T05:27:48Z","timestamp":1755754068000},"page":"190-202","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Evaluating Prompt-Based and\u00a0Fine-Tuned Approaches to\u00a0Czech Anaphora Resolution"],"prefix":"10.1007","author":[{"given":"Patrik","family":"Stano","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6348-109X","authenticated-orcid":false,"given":"Ale\u0161","family":"Hor\u00e1k","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,22]]},"reference":[{"key":"17_CR1","doi-asserted-by":"publisher","unstructured":"Agrawal, M., et\u00a0al.: Large language models are few-shot clinical information extractors. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 1998\u20132022. Association for Computational Linguistics, Abu Dhabi (2022). https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.130, https:\/\/aclanthology.org\/2022.emnlp-main.130","DOI":"10.18653\/v1\/2022.emnlp-main.130"},{"key":"17_CR2","unstructured":"Aryabumi, V., et\u00a0al.: Aya 23: Open weight releases to further multilingual progress (2024). https:\/\/arxiv.org\/abs\/2405.15032"},{"key":"17_CR3","unstructured":"Dubey, A., et\u00a0al.: The LLaMA 3 herd of models (2024). https:\/\/arxiv.org\/abs\/2407.21783"},{"key":"17_CR4","unstructured":"Haji\u010d, J., et\u00a0al.: Prague dependency treebank\u00a0\u2013\u00a0Consolid.\u00a01.0, PDT-C\u00a01.0 (2020). http:\/\/hdl.handle.net\/11234\/1-3185. LINDAT\/CLARIAH-CZ digital library at the Institute of Formal and Applied Linguistics \u00daFAL, Faculty of Mathematics and Physics, Charles University"},{"key":"17_CR5","unstructured":"Haji\u010d, J., et\u00a0al.: Prague dependency treebank 3.5 (2018). http:\/\/hdl.handle.net\/11234\/1-2621. LINDAT\/CLARIN digital library at the Institute of Formal and Applied Linguistics \u00daFAL, Faculty of Mathematics and Physics, Charles University"},{"key":"17_CR6","unstructured":"Hicke, R., Mimno, D.: [Lions: 1] and [Tigers: 2] and [Bears: 3], Oh my! Literary coreference annotation with LLMs. In: Proceedings of the 8th Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Literature (LaTeCH-CLfL 2024), pp. 270\u2013277. Association for Computational Linguistics, St. Julians (2024). https:\/\/aclanthology.org\/2024.latechclfl-1.27\/"},{"key":"17_CR7","unstructured":"Hu, E.J., et\u00a0al.: LoRA: low-rank adaptation of large language models. In: International Conference on Learning Representations (2022). https:\/\/openreview.net\/forum?id=nZeVKeeFYf9"},{"key":"17_CR8","unstructured":"Jiang, A.Q., et\u00a0al.: Mistral 7B (2023). https:\/\/arxiv.org\/abs\/2310.06825"},{"key":"17_CR9","unstructured":"Le, N.T., Ritter, A.: Are large language models robust coreference resolvers? (2023). https:\/\/arxiv.org\/abs\/2305.14489"},{"key":"17_CR10","unstructured":"Gemma Team, Google Deepmind: Gemma 2: Improving open language models at a practical size (2024). https:\/\/arxiv.org\/abs\/2408.00118"},{"key":"17_CR11","unstructured":"Mistral AI Team: Mistral large-instruct-2407 (2024). https:\/\/huggingface.co\/mistralai\/Mistral-Large-Instruct-2407"},{"issue":"140","key":"17_CR12","first-page":"1","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(140), 1\u201367 (2020)","journal-title":"J. Mach. Learn. Res."},{"key":"17_CR13","unstructured":"Straka, M.: UDPipe 2.0 prototype at CoNLL 2018 UD shared task. In: Proceedings of the CoNLL 2018 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies, pp. 197\u2013207. Association for Computational Linguistics (2018)"},{"key":"17_CR14","doi-asserted-by":"publisher","unstructured":"Xue, L., et\u00a0al.: mT5: a massively multilingual pre-trained text-to-text transformer. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 483\u2013498. Association for Computational Linguistics, Online (2021). https:\/\/doi.org\/10.18653\/v1\/2021.naacl-main.41, https:\/\/aclanthology.org\/2021.naacl-main.41\/","DOI":"10.18653\/v1\/2021.naacl-main.41"}],"container-title":["Lecture Notes in Computer Science","Text, Speech, and Dialogue"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-02551-7_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T05:27:51Z","timestamp":1755754071000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-02551-7_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,22]]},"ISBN":["9783032025500","9783032025517"],"references-count":14,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-02551-7_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025,8,22]]},"assertion":[{"value":"22 August 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"TSD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Text, Speech, and Dialogue","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Erlangen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Germany","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 August 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"tsd2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.kiv.zcu.cz\/tsd2025\/index.php","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}