{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T12:57:34Z","timestamp":1763643454424,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":28,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,12,14]],"date-time":"2021-12-14T00:00:00Z","timestamp":1639440000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"FCT (Funda\u00e7\u00e3o para a Ci\u00eancia e a Tecnologia)","award":["POCI\/01\/0145\/FEDER\/03146"],"award-info":[{"award-number":["POCI\/01\/0145\/FEDER\/03146"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,12,14]]},"DOI":"10.1145\/3486622.3493930","type":"proceedings-article","created":{"date-parts":[[2022,4,14]],"date-time":"2022-04-14T01:18:53Z","timestamp":1649899133000},"page":"277-284","source":"Crossref","is-referenced-by-count":1,"title":["Towards Better Evidence Extraction Methods for Fact-Checking Systems"],"prefix":"10.1145","author":[{"given":"Pedro","family":"Azevedo","sequence":"first","affiliation":[{"name":"Faculdade de Engenharia da Universidade do Porto, Portugal"}]},{"given":"Gil","family":"Rocha","sequence":"additional","affiliation":[{"name":"Laborat\u00f3rio de Intelig\u00eancia Artificial e Ci\u00eancia de Computadores (LIACC), Portugal"}]},{"given":"Diego","family":"Esteves","sequence":"additional","affiliation":[{"name":"SDA Research, Germany"}]},{"given":"Henrique Lopes","family":"Cardoso","sequence":"additional","affiliation":[{"name":"Laborat\u00f3rio de Intelig\u00eancia Artificial e Ci\u00eancia de Computadores (LIACC), Portugal"}]}],"member":"320","published-online":{"date-parts":[[2022,4,13]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Samuel\u00a0R. Bowman Gabor Angeli Christopher Potts and Christopher\u00a0D. Manning. 2015. A large annotated corpus for learning natural language inference. CoRR abs\/1508.05326(2015). arxiv:1508.05326http:\/\/arxiv.org\/abs\/1508.05326  Samuel\u00a0R. Bowman Gabor Angeli Christopher Potts and Christopher\u00a0D. Manning. 2015. A large annotated corpus for learning natural language inference. CoRR abs\/1508.05326(2015). arxiv:1508.05326http:\/\/arxiv.org\/abs\/1508.05326"},{"key":"e_1_3_2_1_2_1","volume-title":"Reading wikipedia to answer open-domain questions. arXiv:1704.00051","author":"Chen Danqi","year":"2017","unstructured":"Danqi Chen , Adam Fisch , Jason Weston , and Antoine Bordes . 2017. Reading wikipedia to answer open-domain questions. arXiv:1704.00051 ( 2017 ). Danqi Chen, Adam Fisch, Jason Weston, and Antoine Bordes. 2017. Reading wikipedia to answer open-domain questions. arXiv:1704.00051 (2017)."},{"key":"e_1_3_2_1_3_1","unstructured":"Qian Chen Xiaodan Zhu Zhen-Hua Ling Si Wei and Hui Jiang. 2016. Enhancing and Combining Sequential and Tree LSTM for Natural Language Inference. CoRR abs\/1609.06038(2016). arxiv:1609.06038http:\/\/arxiv.org\/abs\/1609.06038  Qian Chen Xiaodan Zhu Zhen-Hua Ling Si Wei and Hui Jiang. 2016. Enhancing and Combining Sequential and Tree LSTM for Natural Language Inference. CoRR abs\/1609.06038(2016). arxiv:1609.06038http:\/\/arxiv.org\/abs\/1609.06038"},{"volume-title":"Recognizing Textual Entailment: Models and Applications","author":"Dagan Ido","key":"e_1_3_2_1_4_1","unstructured":"Ido Dagan , Dan Roth , Mark Sammons , and Fabio\u00a0Massimo Zanzotto . 2013. Recognizing Textual Entailment: Models and Applications . Morgan & Claypool Publishers . Ido Dagan, Dan Roth, Mark Sammons, and Fabio\u00a0Massimo Zanzotto. 2013. Recognizing Textual Entailment: Models and Applications. Morgan & Claypool Publishers."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/2488388.2488420"},{"key":"e_1_3_2_1_6_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. CoRR abs\/1810.04805(2018). arxiv:1810.04805http:\/\/arxiv.org\/abs\/1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2018 . BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. CoRR abs\/1810.04805(2018). arxiv:1810.04805http:\/\/arxiv.org\/abs\/1810.04805 Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. CoRR abs\/1810.04805(2018). arxiv:1810.04805http:\/\/arxiv.org\/abs\/1810.04805"},{"key":"e_1_3_2_1_7_1","first-page":"3","article-title":"Toward Veracity Assessment in RDF Knowledge Bases","volume":"9","author":"Esteves Diego","year":"2018","unstructured":"Diego Esteves , Anisa Rula , Aniketh\u00a0Janardhan Reddy , and Jens Lehmann . 2018 . Toward Veracity Assessment in RDF Knowledge Bases : An Exploratory Analysis. J. Data and Information Quality 9 , 3 (Feb. 2018). Diego Esteves, Anisa Rula, Aniketh\u00a0Janardhan Reddy, and Jens Lehmann. 2018. Toward Veracity Assessment in RDF Knowledge Bases: An Exploratory Analysis. J. Data and Information Quality 9, 3 (Feb. 2018).","journal-title":"An Exploratory Analysis. J. Data and Information Quality"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"crossref","unstructured":"Kawin Ethayarajh. 2019. How Contextual are Contextualized Word Representations? Comparing the Geometry of BERT ELMo and GPT-2 Embeddings. arxiv:1909.00512  Kawin Ethayarajh. 2019. How Contextual are Contextualized Word Representations? Comparing the Geometry of BERT ELMo and GPT-2 Embeddings. arxiv:1909.00512","DOI":"10.18653\/v1\/D19-1006"},{"key":"e_1_3_2_1_9_1","volume-title":"Axel-Cyrille Ngonga Ngomo, and Ren\u00e9 Speck","author":"Gerber Daniel","year":"2015","unstructured":"Daniel Gerber , Diego Esteves , Jens Lehmann , Lorenz B\u00fchmann , Ricardo Usbeck , Axel-Cyrille Ngonga Ngomo, and Ren\u00e9 Speck . 2015 . DeFacto - Temporal and Multilingual Deep Fact Validation. Web Semantics : Science, Services and Agents on the World Wide Web ( 2015). Daniel Gerber, Diego Esteves, Jens Lehmann, Lorenz B\u00fchmann, Ricardo Usbeck, Axel-Cyrille Ngonga Ngomo, and Ren\u00e9 Speck. 2015. DeFacto - Temporal and Multilingual Deep Fact Validation. Web Semantics: Science, Services and Agents on the World Wide Web (2015)."},{"key":"e_1_3_2_1_10_1","unstructured":"Andreas Hanselowski Hao Zhang Zile Li Daniil Sorokin Benjamin Schiller Claudia Schulz and Iryna Gurevych. 2018. UKP-Athene: Multi-Sentence Textual Entailment for Claim Verification. CoRR abs\/1809.01479(2018). arxiv:1809.01479  Andreas Hanselowski Hao Zhang Zile Li Daniil Sorokin Benjamin Schiller Claudia Schulz and Iryna Gurevych. 2018. UKP-Athene: Multi-Sentence Textual Entailment for Claim Verification. CoRR abs\/1809.01479(2018). arxiv:1809.01479"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_2_1_12_1","volume-title":"Adam: A method for stochastic optimization. arXiv:1412.6980","author":"Kingma P","year":"2014","unstructured":"Diederik\u00a0 P Kingma and Jimmy Ba . 2014 . Adam: A method for stochastic optimization. arXiv:1412.6980 (2014). Diederik\u00a0P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv:1412.6980 (2014)."},{"key":"e_1_3_2_1_13_1","volume-title":"Albert: A lite bert for self-supervised learning of language representations. arXiv:1909.11942","author":"Lan Zhenzhong","year":"2019","unstructured":"Zhenzhong Lan , Mingda Chen , Sebastian Goodman , Kevin Gimpel , Piyush Sharma , and Radu Soricut . 2019 . Albert: A lite bert for self-supervised learning of language representations. arXiv:1909.11942 (2019). Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. Albert: A lite bert for self-supervised learning of language representations. arXiv:1909.11942 (2019)."},{"key":"e_1_3_2_1_14_1","unstructured":"Vladimir\u00a0I Levenshtein. 1966. Binary codes capable of correcting deletions insertions and reversals. In Soviet physics doklady Vol.\u00a010. 707\u2013710.  Vladimir\u00a0I Levenshtein. 1966. Binary codes capable of correcting deletions insertions and reversals. In Soviet physics doklady Vol.\u00a010. 707\u2013710."},{"key":"e_1_3_2_1_15_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu , Myle Ott , Naman Goyal , Jingfei Du , Mandar Joshi , Danqi Chen , Omer Levy , Mike Lewis , Luke Zettlemoyer , and Veselin Stoyanov . 2019 . Roberta: A robustly optimized bert pretraining approach. arXiv:1907.11692 (2019). Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_1_16_1","volume-title":"Performance Measures For Information Extraction. In In Proceedings of DARPA Broadcast News Workshop. 249\u2013252","author":"Makhoul John","year":"1999","unstructured":"John Makhoul , Francis Kubala , Richard Schwartz , and Ralph Weischedel . 1999 . Performance Measures For Information Extraction. In In Proceedings of DARPA Broadcast News Workshop. 249\u2013252 . John Makhoul, Francis Kubala, Richard Schwartz, and Ralph Weischedel. 1999. Performance Measures For Information Extraction. In In Proceedings of DARPA Broadcast News Workshop. 249\u2013252."},{"key":"e_1_3_2_1_17_1","volume-title":"Team Papelo: Transformer Networks at FEVER. CoRR abs\/1901.02534(2019). arxiv:1901.02534","author":"Malon Christopher","year":"2019","unstructured":"Christopher Malon . 2019 . Team Papelo: Transformer Networks at FEVER. CoRR abs\/1901.02534(2019). arxiv:1901.02534 Christopher Malon. 2019. Team Papelo: Transformer Networks at FEVER. CoRR abs\/1901.02534(2019). arxiv:1901.02534"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33016859"},{"key":"e_1_3_2_1_19_1","volume-title":"A decomposable attention model for natural language inference. arXiv:1606.01933","author":"Parikh P","year":"2016","unstructured":"Ankur\u00a0 P Parikh , Oscar T\u00e4ckstr\u00f6m , Dipanjan Das , and Jakob Uszkoreit . 2016. A decomposable attention model for natural language inference. arXiv:1606.01933 ( 2016 ). Ankur\u00a0P Parikh, Oscar T\u00e4ckstr\u00f6m, Dipanjan Das, and Jakob Uszkoreit. 2016. A decomposable attention model for natural language inference. arXiv:1606.01933 (2016)."},{"key":"e_1_3_2_1_20_1","unstructured":"Aniketh\u00a0Janardhan Reddy Gil Rocha and Diego Esteves. 2018. DeFactoNLP: Fact Verification using Entity Recognition TFIDF Vector Comparison and Decomposable Attention. arxiv:1809.00509  Aniketh\u00a0Janardhan Reddy Gil Rocha and Diego Esteves. 2018. DeFactoNLP: Fact Verification using Entity Recognition TFIDF Vector Comparison and Decomposable Attention. arxiv:1809.00509"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1410"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.websem.2018.09.002"},{"volume-title":"Advances in Information Retrieval, Joemon\u00a0M","author":"Soleimani Amir","key":"e_1_3_2_1_23_1","unstructured":"Amir Soleimani , Christof Monz , and Marcel Worring . 2020. BERT for Evidence Retrieval and Claim Verification . In Advances in Information Retrieval, Joemon\u00a0M . Jose, Emine Yilmaz, Jo\u00e3o Magalh\u00e3es, Pablo Castells, Nicola Ferro, M\u00e1rio\u00a0J. Silva, and Fl\u00e1vio Martins (Eds.). Springer International Publishing , Cham , 359\u2013366. Amir Soleimani, Christof Monz, and Marcel Worring. 2020. BERT for Evidence Retrieval and Claim Verification. In Advances in Information Retrieval, Joemon\u00a0M. Jose, Emine Yilmaz, Jo\u00e3o Magalh\u00e3es, Pablo Castells, Nicola Ferro, M\u00e1rio\u00a0J. Silva, and Fl\u00e1vio Martins (Eds.). Springer International Publishing, Cham, 359\u2013366."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"James Thorne Andreas Vlachos Christos Christodoulopoulos and Arpit Mittal. 2018. FEVER: a large-scale dataset for Fact Extraction and VERification. arxiv:1803.05355  James Thorne Andreas Vlachos Christos Christodoulopoulos and Arpit Mittal. 2018. FEVER: a large-scale dataset for Fact Extraction and VERification. arxiv:1803.05355","DOI":"10.18653\/v1\/N18-1074"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"crossref","unstructured":"James Thorne Andreas Vlachos Oana Cocarascu Christos Christodoulopoulos and Arpit Mittal. 2018. The Fact Extraction and VERification (FEVER) Shared Task. arxiv:1811.10971  James Thorne Andreas Vlachos Oana Cocarascu Christos Christodoulopoulos and Arpit Mittal. 2018. The Fact Extraction and VERification (FEVER) Shared Task. arxiv:1811.10971","DOI":"10.18653\/v1\/W18-5501"},{"key":"e_1_3_2_1_26_1","volume-title":"HuggingFace\u2019s Transformers: State-of-the-art Natural Language Processing. arXiv:1910.03771","author":"Wolf Thomas","year":"2019","unstructured":"Thomas Wolf , Lysandre Debut , Victor Sanh , Julien Chaumond , Clement Delangue , Anthony Moi , Pierric Cistac , Tim Rault , R\u00e9mi Louf , Morgan Funtowicz , Joe Davison , Sam Shleifer , Patrick von Platen , Clara Ma , Yacine Jernite , Julien Plu , Canwen Xu , Teven\u00a0Le Scao , Sylvain Gugger , Mariama Drame , Quentin Lhoest , and Alexander\u00a0 M. Rush . 2019. HuggingFace\u2019s Transformers: State-of-the-art Natural Language Processing. arXiv:1910.03771 ( 2019 ). Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, R\u00e9mi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven\u00a0Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander\u00a0M. Rush. 2019. HuggingFace\u2019s Transformers: State-of-the-art Natural Language Processing. arXiv:1910.03771 (2019)."},{"key":"e_1_3_2_1_27_1","unstructured":"Zhilin Yang Zihang Dai Yiming Yang Jaime\u00a0G. Carbonell Ruslan Salakhutdinov and Quoc\u00a0V. Le. 2019. XLNet: Generalized Autoregressive Pretraining for Language Understanding. CoRR abs\/1906.08237(2019). arxiv:1906.08237  Zhilin Yang Zihang Dai Yiming Yang Jaime\u00a0G. Carbonell Ruslan Salakhutdinov and Quoc\u00a0V. Le. 2019. XLNet: Generalized Autoregressive Pretraining for Language Understanding. CoRR abs\/1906.08237(2019). arxiv:1906.08237"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Wanjun Zhong Jingjing Xu Duyu Tang Zenan Xu Nan Duan Ming Zhou Jiahai Wang and Jian Yin. 2019. Reasoning Over Semantic-Level Graph for Fact Checking. arxiv:1909.03745  Wanjun Zhong Jingjing Xu Duyu Tang Zenan Xu Nan Duan Ming Zhou Jiahai Wang and Jian Yin. 2019. Reasoning Over Semantic-Level Graph for Fact Checking. arxiv:1909.03745","DOI":"10.18653\/v1\/2020.acl-main.549"}],"event":{"name":"WI-IAT '21: IEEE\/WIC\/ACM International Conference on Web Intelligence","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence"],"location":"ESSENDON VIC Australia","acronym":"WI-IAT '21"},"container-title":["IEEE\/WIC\/ACM International Conference on Web Intelligence"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3486622.3493930","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3486622.3493930","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:12:06Z","timestamp":1750191126000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3486622.3493930"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12,14]]},"references-count":28,"alternative-id":["10.1145\/3486622.3493930","10.1145\/3486622"],"URL":"https:\/\/doi.org\/10.1145\/3486622.3493930","relation":{},"subject":[],"published":{"date-parts":[[2021,12,14]]}}}