{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,15]],"date-time":"2026-01-15T10:23:03Z","timestamp":1768472583228,"version":"3.49.0"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783031171048","type":"print"},{"value":"9783031171055","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,9,20]],"date-time":"2022-09-20T00:00:00Z","timestamp":1663632000000},"content-version":"vor","delay-in-days":262,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Recent years have shown that deep learning models pre-trained on large text corpora using the language model objective can help solve various tasks requiring natural language understanding. However, many commonsense concepts are underrepresented in online resources because they are too obvious for most humans. To solve this problem, we propose the use of affordances \u2013 common-sense knowledge that can be injected into models to increase their ability to understand our world. We show that injecting ConceptNet knowledge into BERT-based models leads to an increase in evaluation scores measured on the PIQA dataset.<\/jats:p>","DOI":"10.1007\/978-3-031-17105-5_7","type":"book-chapter","created":{"date-parts":[[2022,9,19]],"date-time":"2022-09-19T19:05:41Z","timestamp":1663614341000},"page":"97-104","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Should We Afford Affordances? Injecting ConceptNet Knowledge into BERT-Based Models to Improve Commonsense Reasoning Ability"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7540-3259","authenticated-orcid":false,"given":"Andrzej","family":"Gretkowski","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1194-7921","authenticated-orcid":false,"given":"Dawid","family":"Wi\u015bniewski","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2442-345X","authenticated-orcid":false,"given":"Agnieszka","family":"\u0141awrynowicz","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,20]]},"reference":[{"key":"7_CR1","doi-asserted-by":"publisher","unstructured":"Be\u00dfler, et al.: A formal model of affordances for flexible robotic task execution. In: ECAI 2020, pp. 2425\u20132432 (2020). https:\/\/doi.org\/10.3233\/FAIA200374, https:\/\/ebooks.iospress.nl\/doi\/10.3233\/FAIA200374","DOI":"10.3233\/FAIA200374"},{"key":"7_CR2","doi-asserted-by":"crossref","unstructured":"Bisk, Y., et al.: PIQA: reasoning about physical commonsense in natural language. In: Proceedings of AAAI, vol. 34, pp. 7432\u20137439 (2020)","DOI":"10.1609\/aaai.v34i05.6239"},{"key":"7_CR3","unstructured":"Brown, T.B., et al.: Language models are few-shot learners. arXiv preprint arXiv:2005.14165 (2020)"},{"key":"7_CR4","doi-asserted-by":"publisher","first-page":"257","DOI":"10.1016\/j.ins.2019.09.013","volume":"509","author":"R Campos","year":"2020","unstructured":"Campos, R., et al.: YAKE! keyword extraction from single documents using multiple local features. Inf. Sci. 509, 257\u2013289 (2020)","journal-title":"Inf. Sci."},{"key":"7_CR5","unstructured":"Cui, L., et al.: Does BERT solve commonsense task via commonsense knowledge? arXiv preprint arXiv:2008.03945 (2020)"},{"key":"7_CR6","unstructured":"Devlin, J., et al.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"issue":"4","key":"7_CR7","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1007\/s11023-020-09548-1","volume":"30","author":"L Floridi","year":"2020","unstructured":"Floridi, L., et al.: GPT-3: its nature, scope, limits, and consequences. Mind. Mach. 30(4), 681\u2013694 (2020)","journal-title":"Mind. Mach."},{"key":"7_CR8","unstructured":"Forbes, M., et al.: Do neural language representations learn physical commonsense? In: Proceedings of CogSci 2019, pp. 1753\u20131759. cognitivesciencesociety.org (2019)"},{"key":"7_CR9","unstructured":"Gibson, J.J.: The Theory of Affordances, Hilldale, USA, vol. 1, no. 2, pp. 67\u201382 (1977)"},{"key":"7_CR10","unstructured":"He, P., et al.: DeBERTa: decoding-enhanced BERT with disentangled attention. arXiv preprint arXiv:2006.03654 (2020)"},{"key":"7_CR11","doi-asserted-by":"crossref","unstructured":"Ilievski, F., et al.: Dimensions of commonsense knowledge. arXiv preprint arXiv:2101.04640 (2021)","DOI":"10.1016\/j.knosys.2021.107347"},{"key":"7_CR12","unstructured":"Lan, Z., et al.: Albert: a lite BERT for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942 (2019)"},{"key":"7_CR13","doi-asserted-by":"crossref","unstructured":"Lauscher, A., et al.: Common sense or world knowledge? Investigating adapter-based knowledge injection into pretrained transformers. CoRR abs\/2005.11787 (2020). https:\/\/arxiv.org\/abs\/2005.11787","DOI":"10.18653\/v1\/2020.deelio-1.5"},{"key":"7_CR14","unstructured":"Liu, Y., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"7_CR15","doi-asserted-by":"crossref","unstructured":"Lourie, N., et al.: UNICORN on RAINBOW: a universal commonsense reasoning model on a new multitask benchmark. In: Proceedings of AAAI, pp. 13480\u201313488. AAAI Press (2021)","DOI":"10.1609\/aaai.v35i15.17590"},{"key":"7_CR16","doi-asserted-by":"crossref","unstructured":"Lv, S., et al.: Graph-based reasoning over heterogeneous external knowledge for commonsense question answering. In: Proceedings of AAAI, vol. 34, pp. 8449\u20138456 (2020)","DOI":"10.1609\/aaai.v34i05.6364"},{"key":"7_CR17","unstructured":"Ostendorff, M., et al.: Enriching BERT with knowledge graph embeddings for document classification. arXiv preprint arXiv:1909.08402 (2019)"},{"key":"7_CR18","doi-asserted-by":"publisher","unstructured":"Potoniec, J., et al.: Incorporating presuppositions of competency questions into test-driven development of ontologies. In: Proceedings of SEKE 2021, pp. 437\u2013440 (2021). https:\/\/doi.org\/10.18293\/SEKE2021-165","DOI":"10.18293\/SEKE2021-165"},{"key":"7_CR19","doi-asserted-by":"crossref","unstructured":"Rajani, N.F., et al.: Explain yourself! leveraging language models for commonsense reasoning. arXiv preprint arXiv:1906.02361 (2019)","DOI":"10.18653\/v1\/P19-1487"},{"key":"7_CR20","doi-asserted-by":"crossref","unstructured":"Sap, M., et al.: ATOMIC: an atlas of machine commonsense for if-then reasoning. In: Proceedings of AAAI, vol. 33, pp. 3027\u20133035 (2019)","DOI":"10.1609\/aaai.v33i01.33013027"},{"key":"7_CR21","doi-asserted-by":"crossref","unstructured":"Shwartz, V., et al.: Unsupervised commonsense question answering with self-talk. arXiv preprint arXiv:2004.05483 (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.373"},{"key":"7_CR22","doi-asserted-by":"crossref","unstructured":"Speer, R., et al.: ConceptNet 5.5: an open multilingual graph of general knowledge. In: Proceedings of AAAI, vol. 31 (2017)","DOI":"10.1609\/aaai.v31i1.11164"},{"key":"7_CR23","doi-asserted-by":"crossref","unstructured":"Speer, R., et al.: ConceptNet 5.5: an open multilingual graph of general knowledge. In: Proceedings of AAAI, AAAI 2017, pp. 4444\u20134451. AAAI Press (2017)","DOI":"10.1609\/aaai.v31i1.11164"},{"key":"7_CR24","unstructured":"Talmor, A., et al.: CommonsenseQA: a question answering challenge targeting commonsense knowledge. arXiv preprint arXiv:1811.00937 (2018)"},{"key":"7_CR25","unstructured":"Wales, J.: The Wikimedia community: Wiktionary (2002). https:\/\/www.wiktionary.org\/. Accessed 10 Oct 2021"},{"key":"7_CR26","unstructured":"Yang, Y., et al.: G-DAUG: generative data augmentation for commonsense reasoning. arXiv preprint arXiv:2004.11546 (2020)"},{"key":"7_CR27","doi-asserted-by":"crossref","unstructured":"Zellers, R., et al.: SWAG: a large-scale adversarial dataset for grounded commonsense inference. arXiv preprint arXiv:1808.05326 (2018)","DOI":"10.18653\/v1\/D18-1009"},{"key":"7_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"408","DOI":"10.1007\/978-3-319-10605-2_27","volume-title":"Computer Vision \u2013 ECCV 2014","author":"Y Zhu","year":"2014","unstructured":"Zhu, Y., Fathi, A., Fei-Fei, L.: Reasoning about object affordances in a knowledge base representation. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8690, pp. 408\u2013424. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10605-2_27"}],"container-title":["Lecture Notes in Computer Science","Knowledge Engineering and Knowledge Management"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-17105-5_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T14:13:13Z","timestamp":1710252793000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-17105-5_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031171048","9783031171055"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-17105-5_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"20 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"EKAW","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Knowledge Engineering and Knowledge Management","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bolzano","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ekaw2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ekaw2022.inf.unibz.it\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"57","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"11","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"19% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,47","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,25","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}