{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T16:20:36Z","timestamp":1773246036832,"version":"3.50.1"},"reference-count":44,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,7,8]],"date-time":"2024-07-08T00:00:00Z","timestamp":1720396800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,7,8]],"date-time":"2024-07-08T00:00:00Z","timestamp":1720396800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,7,8]]},"DOI":"10.23919\/fusion59988.2024.10706385","type":"proceedings-article","created":{"date-parts":[[2024,10,11]],"date-time":"2024-10-11T17:19:50Z","timestamp":1728667190000},"page":"1-8","source":"Crossref","is-referenced-by-count":6,"title":["Decompose, Enrich, and Extract! Schema-aware Event Extraction using LLMs."],"prefix":"10.23919","author":[{"given":"Fatemeh","family":"Shiri","sequence":"first","affiliation":[{"name":"Monash University,Faculty of Information Technology,Department of Data Science and AI,Melbourne,Australia"}]},{"given":"Farhad","family":"Moghimifar","sequence":"additional","affiliation":[{"name":"Monash University,Faculty of Information Technology,Department of Data Science and AI,Melbourne,Australia"}]},{"given":"Reza","family":"Haffari","sequence":"additional","affiliation":[{"name":"Monash University,Faculty of Information Technology,Department of Data Science and AI,Melbourne,Australia"}]},{"given":"Yuan-Fang","family":"Li","sequence":"additional","affiliation":[{"name":"Monash University,Faculty of Information Technology,Department of Data Science and AI,Melbourne,Australia"}]},{"given":"Van","family":"Nguyen","sequence":"additional","affiliation":[{"name":"Information Sciences Division Defence Science and Technology Group,Adelaide,Australia"}]},{"given":"John","family":"Yoo","sequence":"additional","affiliation":[{"name":"Information Sciences Division Defence Science and Technology Group,Adelaide,Australia"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.23919\/FUSION49465.2021.9626935"},{"key":"ref2","article-title":"Gpt-4 technical report","volume-title":"arXiv preprint arXiv:2303.08774","author":"Achiam","year":"2023"},{"key":"ref3","article-title":"How large-language models can revolutionize military planning","volume":"12","author":"Jensen","year":"2023","journal-title":"War on the Rocks"},{"issue":"12","key":"ref4","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3571730","article-title":"Survey of hallucination in natural language generation","volume":"55","author":"Ji","year":"2023","journal-title":"ACM Computing Surveys"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.397"},{"key":"ref6","first-page":"24 824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume":"35","author":"Wei","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref7","first-page":"9459","article-title":"Retrieval-augmented generation for knowledge-intensive nlp tasks","volume":"33","author":"Lewis","year":"2020","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380107"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.597"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.217"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.ijcnlp-short.16"},{"key":"ref12","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref13","volume-title":"Ace 2005 multilingual training corpus","author":"Christopher Walker","year":"2006"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.713"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33016851"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/WACV.2019.00049"},{"key":"ref17","first-page":"22","article-title":"Tcg-event: Effective task conditioning for generation-based event extraction","volume-title":"Proceedings of the The 20th Annual Workshop of the Australasian Language Technology Association","author":"Shiri"},{"key":"ref18","article-title":"Structured prediction as translation between augmented natural languages","author":"Paolini","year":"2021","journal-title":"arXiv preprint arXiv:2101.05779"},{"key":"ref19","article-title":"Grit: Generative role-filler transformers for document-level event entity extraction","author":"Du","year":"2020","journal-title":"arXiv preprint arXiv:2008.09249"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/s11280-023-01216-5"},{"key":"ref21","article-title":"Emergent abilities of large language models","author":"Wei","year":"2022","journal-title":"arXiv preprint arXiv:2206.07682"},{"key":"ref22","article-title":"Dissociating language and thought in large language models: a cognitive perspective","author":"Mahowald","year":"2023","journal-title":"arXiv preprint arXiv:2301.06627"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.23919\/FUSION52260.2023.10224213"},{"key":"ref24","article-title":"Training verifiers to solve math word problems","author":"Cobbe","year":"2021","journal-title":"arXiv preprint arXiv:2110.14168"},{"key":"ref25","article-title":"Evaluating chatgpt\u2019s information extraction capabilities: An assessment of performance, explainability, calibration, and faithfulness","author":"Li","year":"2023","journal-title":"arXiv preprint arXiv:2304.11633"},{"key":"ref26","article-title":"Is information extraction solved by chatgpt? an analysis of performance, evaluation criteria, robustness and errors","author":"Han","year":"2023","journal-title":"arXiv preprint arXiv:2305.14450"},{"key":"ref27","article-title":"Characterizing information seeking events in health-related social discourse","author":"Sharif","year":"2023","journal-title":"arXiv preprint arXiv:2308.09156"},{"key":"ref28","article-title":"Ultra: Unleash llms\u2019 potential for event argument extraction through hierarchical modeling and pair-wise refinement","author":"Zhang","year":"2024","journal-title":"arXiv preprint arXiv:2401.13218"},{"key":"ref29","volume-title":"Zero-shot information extraction via chatting with chatgpt (arxiv: 2302.10205). arxiv","author":"Wei","year":"2023"},{"key":"ref30","article-title":"Benchmarking large language models with augmented instructions for fine-grained information extraction","author":"Gao","year":"2023","journal-title":"arXiv preprint arXiv:2310.05092"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TBDATA.2019.2921572"},{"key":"ref32","article-title":"Never lost in the middle: Improving large language models via attention strengthening question answering","author":"Junqing","year":"2023","journal-title":"arXiv preprint arXiv:2311.09198"},{"key":"ref33","article-title":"Retrieval meets long context large language models","author":"Xu","year":"2023","journal-title":"arXiv preprint arXiv:2310.03025"},{"key":"ref34","article-title":"Decomposed prompting: A modular approach for solving complex tasks","author":"Khot","year":"2022","journal-title":"arXiv preprint arXiv:2210.02406"},{"key":"ref35","article-title":"Compositional semantic parsing with large language models","author":"Drozdov","year":"2022","journal-title":"arXiv preprint arXiv:2209.15003"},{"key":"ref36","article-title":"Finetuned language models are zero-shot learners","author":"Wei","year":"2021","journal-title":"arXiv preprint arXiv:2109.01652"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.69"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.217"},{"issue":"1","key":"ref39","first-page":"5485","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"The Journal of Machine Learning Research"},{"key":"ref40","article-title":"Structured prompt interrogation and recursive extraction of semantics (spires): A method for populating knowledge bases using zero-shot learning","author":"Caufield","year":"2023","journal-title":"arXiv preprint arXiv:2304.02711"},{"key":"ref41","first-page":"148","article-title":"The linked data modeling language (linkml): A general-purpose data modeling framework grounded in machine-readable semantics","volume-title":"ICBO","author":"Moxon"},{"key":"ref42","article-title":"Universal sentence encoder","author":"Cer","year":"2018","journal-title":"arXiv preprint arXiv:1803.11175"},{"key":"ref43","article-title":"Roberta: A robustly optimized bert pretraining approach","author":"Liu","year":"2019","journal-title":"arXiv preprint arXiv:1907.11692"},{"key":"ref44","article-title":"Text and code embeddings by contrastive pre-training","author":"Neelakantan","year":"2022","journal-title":"arXiv preprint arXiv:2201.10005"}],"event":{"name":"2024 27th International Conference on Information Fusion (FUSION)","location":"Venice, Italy","start":{"date-parts":[[2024,7,8]]},"end":{"date-parts":[[2024,7,11]]}},"container-title":["2024 27th International Conference on Information Fusion (FUSION)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10706250\/10706251\/10706385.pdf?arnumber=10706385","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,12]],"date-time":"2024-10-12T04:34:28Z","timestamp":1728707668000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10706385\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,8]]},"references-count":44,"URL":"https:\/\/doi.org\/10.23919\/fusion59988.2024.10706385","relation":{},"subject":[],"published":{"date-parts":[[2024,7,8]]}}}