{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,24]],"date-time":"2026-01-24T19:53:49Z","timestamp":1769284429446,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":32,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,4,8]],"date-time":"2024-04-08T00:00:00Z","timestamp":1712534400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,4,8]]},"DOI":"10.1145\/3605098.3635957","type":"proceedings-article","created":{"date-parts":[[2024,5,21]],"date-time":"2024-05-21T17:59:16Z","timestamp":1716314356000},"page":"1634-1641","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":5,"title":["Process Knowledge Extraction and Knowledge Graph Construction Through Prompting: A Quantitative Analysis"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2971-1872","authenticated-orcid":false,"given":"Patrizio","family":"Bellan","sequence":"first","affiliation":[{"name":"Fondazione Bruno Kessler, Free University of Bozen-Bolzano, Povo, Trento, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0380-6571","authenticated-orcid":false,"given":"Mauro","family":"Dragoni","sequence":"additional","affiliation":[{"name":"Fondazione Bruno Kessler, Povo, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1563-4965","authenticated-orcid":false,"given":"Chiara","family":"Ghidini","sequence":"additional","affiliation":[{"name":"Fondazione Bruno Kessler, Povo, Italy"}]}],"member":"320","published-online":{"date-parts":[[2024,5,21]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Ives","author":"Auer S\u00f6ren","year":"2007","unstructured":"S\u00f6ren Auer, Christian Bizer, Georgi Kobilarov, Jens Lehmann, Richard Cyganiak, and Zachary G. Ives. 2007. DBpedia: A Nucleus for a Web of Open Data (LNCS, Vol. 4825). Springer."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P14-1023"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Patrizio Bellan Mauro Dragoni and Chiara Ghidini. 2022. Assisted Process Knowledge Graph Building Using Pre-trained Language Models (LNCS Vol. 13796). Springer.","DOI":"10.1007\/978-3-031-27181-6_5"},{"key":"e_1_3_2_1_4_1","volume-title":"PET: An Annotated Dataset for Process Extraction from Natural Language Text Tasks (Lecture Notes in Business Information Processing","author":"Bellan Patrizio","year":"2022","unstructured":"Patrizio Bellan, Han van der Aa, Mauro Dragoni, Chiara Ghidini, and Simone Paolo Ponzetto. 2022. PET: An Annotated Dataset for Process Extraction from Natural Language Text Tasks (Lecture Notes in Business Information Processing, Vol. 460). Springer."},{"key":"e_1_3_2_1_5_1","volume-title":"Annual Conf. on Neural Information Processing Systems 2020","author":"Tom","year":"2020","unstructured":"Tom B. Brown and et al. 2020. Language Models are Few-Shot Learners. In Annual Conf. on Neural Information Processing Systems 2020, NeurIPS 2020."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Nicola De Cao Wilker Aziz and Ivan Titov. 2021. Editing Factual Knowledge in Language Models. ACL.","DOI":"10.18653\/v1\/2021.emnlp-main.522"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.nlpmc-1.9"},{"key":"e_1_3_2_1_8_1","volume-title":"Ximing Lu, Yejin Choi, and Antoine Bosselut.","author":"Da Jeff","year":"2021","unstructured":"Jeff Da, Ronan Le Bras, Ximing Lu, Yejin Choi, and Antoine Bosselut. 2021. Analyzing Commonsense Emergence in Few-shot Knowledge Models."},{"key":"e_1_3_2_1_9_1","volume-title":"Proc. of NAACL-HLT","volume":"1","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proc. of NAACL-HLT 2019, Volume 1. ACL."},{"key":"e_1_3_2_1_10_1","volume-title":"Measuring and Improving Consistency in Pretrained Language Models. Trans. ACL 9","author":"Elazar Yanai","year":"2021","unstructured":"Yanai Elazar, Nora Kassner, Shauli Ravfogel, Abhilasha Ravichander, Eduard H. Hovy, Hinrich Sch\u00fctze, and Yoav Goldberg. 2021. Measuring and Improving Consistency in Pretrained Language Models. Trans. ACL 9 (2021)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.295"},{"key":"e_1_3_2_1_12_1","volume-title":"International Journal of Emerging Technology and Advanced Engineering","author":"Gupta Sachin","year":"2022","unstructured":"Sachin Gupta. 2022. Hate Speech Detection using OpenAI and GPT-3. International Journal of Emerging Technology and Advanced Engineering (2022)."},{"key":"e_1_3_2_1_13_1","volume-title":"Manning","author":"Hewitt John","year":"2019","unstructured":"John Hewitt and Christopher D. Manning. 2019. A Structural Probe for Finding Syntax in Word Representations. ACL."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1162\/COLI_a_00237"},{"key":"e_1_3_2_1_15_1","volume-title":"How Can We Know What Language Models Know. Trans. ACL 8","author":"Jiang Zhengbao","year":"2020","unstructured":"Zhengbao Jiang, Frank F. Xu, Jun Araki, and Graham Neubig. 2020. How Can We Know What Language Models Know. Trans. ACL 8 (2020)."},{"key":"e_1_3_2_1_16_1","unstructured":"Jinhao Ju Deqing Yang and Jingping Liu. 2022. Commonsense Knowledge Base Completion with Relational Graph Attention Network and Pre-trained Language Model. ACM."},{"key":"e_1_3_2_1_17_1","volume-title":"Decomposed Prompting: A Modular Approach for Solving Complex Tasks. OpenReview.net.","author":"Khot Tushar","year":"2023","unstructured":"Tushar Khot, Harsh Trivedi, Matthew Finlayson, Yao Fu, Kyle Richardson, Peter Clark, and Ashish Sabharwal. 2023. Decomposed Prompting: A Modular Approach for Solving Complex Tasks. OpenReview.net."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/219717.219745"},{"key":"e_1_3_2_1_19_1","volume-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach. ArXiv abs\/1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. ArXiv abs\/1907.11692 (2019)."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.3233\/SW-180333"},{"key":"e_1_3_2_1_21_1","volume-title":"Miller","author":"Petroni Fabio","year":"2019","unstructured":"Fabio Petroni, Tim Rockt\u00e4schel, Sebastian Riedel, Patrick S. H. Lewis, Anton Bakhtin, Yuxiang Wu, and Alexander H. Miller. 2019. Language Models as Knowledge Bases? ACL."},{"key":"e_1_3_2_1_22_1","unstructured":"Guanghui Qin and Jason Eisner. 2021. Learning How to Ask: Querying LMs with Mixtures of Soft Prompts. ACL."},{"key":"e_1_3_2_1_23_1","volume-title":"Liu","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 21 (2020)."},{"key":"e_1_3_2_1_24_1","volume-title":"Forbus","author":"Ribeiro Danilo Neves","year":"2021","unstructured":"Danilo Neves Ribeiro and Kenneth D. Forbus. 2021. Combining Analogy with Language Models for Knowledge Extraction."},{"key":"e_1_3_2_1_25_1","volume-title":"Proc. of NAACL-HLT","author":"Scao Teven Le","year":"2021","unstructured":"Teven Le Scao and Alexander M. Rush. 2021. How many data points is a prompt worth?. In Proc. of NAACL-HLT 2021. ACL."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"Taylor Shin Yasaman Razeghi Robert L. Logan IV Eric Wallace and Sameer Singh. 2020. AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts. ACL.","DOI":"10.18653\/v1\/2020.emnlp-main.346"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"Jaspreet Singh Jonas Wallat and Avishek Anand. 2020. BERTnesia: Investigating the capture and forgetting of knowledge in BERT. ACL.","DOI":"10.18653\/v1\/2020.blackboxnlp-1.17"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Fabian M. Suchanek Gjergji Kasneci and Gerhard Weikum. 2007. Yago: a core of semantic knowledge. ACM.","DOI":"10.1145\/1242572.1242667"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Ian Tenney Dipanjan Das and Ellie Pavlick. 2019. BERT Rediscovers the Classical NLP Pipeline. ACL.","DOI":"10.18653\/v1\/P19-1452"},{"key":"e_1_3_2_1_30_1","volume-title":"Wikidata: a free collaborative knowledgebase. Commun. ACM 57, 10","author":"Vrandecic Denny","year":"2014","unstructured":"Denny Vrandecic and Markus Kr\u00f6tzsch. 2014. Wikidata: a free collaborative knowledgebase. Commun. ACM 57, 10 (2014)."},{"key":"e_1_3_2_1_31_1","volume-title":"Quoc V. Le, and Denny Zhou.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed H. Chi, Quoc V. Le, and Denny Zhou. 2022. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models."},{"key":"e_1_3_2_1_32_1","volume-title":"Tree of Thoughts: Deliberate Problem Solving with Large Language Models. CoRR abs\/2305.10601","author":"Yao Shunyu","year":"2023","unstructured":"Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Thomas L. Griffiths, Yuan Cao, and Karthik Narasimhan. 2023. Tree of Thoughts: Deliberate Problem Solving with Large Language Models. CoRR abs\/2305.10601 (2023). arXiv:2305.10601"}],"event":{"name":"SAC '24: 39th ACM\/SIGAPP Symposium on Applied Computing","location":"Avila Spain","acronym":"SAC '24","sponsor":["SIGAPP ACM Special Interest Group on Applied Computing"]},"container-title":["Proceedings of the 39th ACM\/SIGAPP Symposium on Applied Computing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3605098.3635957","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3605098.3635957","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:14Z","timestamp":1750178174000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3605098.3635957"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,8]]},"references-count":32,"alternative-id":["10.1145\/3605098.3635957","10.1145\/3605098"],"URL":"https:\/\/doi.org\/10.1145\/3605098.3635957","relation":{},"subject":[],"published":{"date-parts":[[2024,4,8]]},"assertion":[{"value":"2024-05-21","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}