{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T03:59:07Z","timestamp":1774497547986,"version":"3.50.1"},"reference-count":28,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,10,20]],"date-time":"2025-10-20T00:00:00Z","timestamp":1760918400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,10,20]],"date-time":"2025-10-20T00:00:00Z","timestamp":1760918400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,10,20]]},"DOI":"10.1109\/icpm66919.2025.11220725","type":"proceedings-article","created":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T18:34:15Z","timestamp":1762281255000},"page":"1-8","source":"Crossref","is-referenced-by-count":2,"title":["LLMs that Understand Processes: Instruction-tuning for Semantics-Aware Process Mining"],"prefix":"10.1109","author":[{"given":"Vira","family":"Pyrih","sequence":"first","affiliation":[{"name":"University of Vienna,Faculty of Computer Science,Vienna,Austria"}]},{"given":"Adrian","family":"Rebmann","sequence":"additional","affiliation":[{"name":"SAP Signavio,Berlin,Germany"}]},{"given":"Han van der","family":"Aa","sequence":"additional","affiliation":[{"name":"University of Vienna,Faculty of Computer Science,Vienna,Austria"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.is.2021.101824"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-41620-0_10"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-78666-2_4"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-70396-6_18"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/s44311-025-00019-3"},{"key":"ref6","article-title":"Finetuned language models are zero-shot learners","author":"Wei","year":"2021","journal-title":"ICLR"},{"key":"ref7","article-title":"Instruction tuning for large language models: A survey","author":"Zhang","year":"2023","journal-title":"arXiv:2308.10792"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-08848-3_2"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref10","first-page":"4171","article-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","author":"Devlin","year":"2019","journal-title":"NAACL. ACL"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.64"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.340"},{"key":"ref13","article-title":"Instruction datasets for process mining","author":"Pyrih","year":"2025"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-27815-0_33"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1162\/coli_a_00523"},{"key":"ref16","first-page":"22631","article-title":"The flan collection: Designing data and methods for effective instruction tuning","author":"Longpre","year":"2023","journal-title":"ICML"},{"issue":"140","key":"ref17","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"Journal of machine learning research"},{"key":"ref18","article-title":"Lora: Low-rank adaptation of large language models","author":"Hu","year":"2021","journal-title":"ICLR"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-99414-7"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-61007-3_3"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v39i11.33277"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICPM63005.2024.10680620"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v39i28.35153"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-50974-2_32"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-56107-8_29"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-70396-6_23"},{"key":"ref27","article-title":"PM-LLMBenchmark: Evaluating large language models on process mining tasks","author":"Berti","year":"2024","journal-title":"arXiv:2407.13244"},{"key":"ref28","article-title":"LLM evaluators recognize and favor their own generations","author":"Panickssery","year":"2024","journal-title":"preprint arXiv:2404.13076"}],"event":{"name":"2025 7th International Conference on Process Mining (ICPM)","location":"Montevideo, Uruguay","start":{"date-parts":[[2025,10,20]]},"end":{"date-parts":[[2025,10,24]]}},"container-title":["2025 7th International Conference on Process Mining (ICPM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11220591\/11220277\/11220725.pdf?arnumber=11220725","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T06:22:12Z","timestamp":1762323732000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11220725\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,20]]},"references-count":28,"URL":"https:\/\/doi.org\/10.1109\/icpm66919.2025.11220725","relation":{},"subject":[],"published":{"date-parts":[[2025,10,20]]}}}