{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T10:31:19Z","timestamp":1763202679687,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":25,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,9,7]],"date-time":"2022-09-07T00:00:00Z","timestamp":1662508800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,9,7]]},"DOI":"10.1145\/3549737.3549753","type":"proceedings-article","created":{"date-parts":[[2022,9,9]],"date-time":"2022-09-09T16:29:59Z","timestamp":1662740999000},"page":"1-8","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Transforming Drug-Drug Interaction Extraction from Biomedical Literature"],"prefix":"10.1145","author":[{"given":"Dimitrios","family":"Zaikis","sequence":"first","affiliation":[{"name":"School of Informatics, Aristotle University of Thessaloniki, Greece"}]},{"given":"Stylianos","family":"Kokkas","sequence":"additional","affiliation":[{"name":"School of Informatics, Aristotle University of Thessaloniki, Greece"}]},{"given":"Ioannis","family":"Vlahavas","sequence":"additional","affiliation":[{"name":"School of Informatics, Aristotle University of Thessaloniki, Greece"}]}],"member":"320","published-online":{"date-parts":[[2022,9,9]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Iz Beltagy Kyle Lo and Arman Cohan. 2019. SciBERT: Pretrained Language Model for Scientific Text. In EMNLP. arXiv:arXiv:1903.10676  Iz Beltagy Kyle Lo and Arman Cohan. 2019. SciBERT: Pretrained Language Model for Scientific Text. In EMNLP. arXiv:arXiv:1903.10676","DOI":"10.18653\/v1\/D19-1371"},{"key":"e_1_3_2_1_2_1","volume-title":"A Survey on Recent Named Entity Recognition and Relationship Extraction Techniques on Clinical Texts. Applied Sciences 11, 18","author":"Bose Priyankar","year":"2021","unstructured":"Priyankar Bose , Sriram Srinivasan , William\u00a0 C. Sleeman , Jatinder Palta , Rishabh Kapoor , and Preetam Ghosh . 2021. A Survey on Recent Named Entity Recognition and Relationship Extraction Techniques on Clinical Texts. Applied Sciences 11, 18 ( 2021 ). https:\/\/www.mdpi.com\/2076-3417\/11\/18\/8319 Priyankar Bose, Sriram Srinivasan, William\u00a0C. Sleeman, Jatinder Palta, Rishabh Kapoor, and Preetam Ghosh. 2021. A Survey on Recent Named Entity Recognition and Relationship Extraction Techniques on Clinical Texts. Applied Sciences 11, 18 (2021). https:\/\/www.mdpi.com\/2076-3417\/11\/18\/8319"},{"key":"e_1_3_2_1_3_1","volume-title":"ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In ICLR. https:\/\/openreview.net\/pdf?id=r1xMH1BtvB","author":"Clark Kevin","year":"2020","unstructured":"Kevin Clark , Minh-Thang Luong , Quoc\u00a0 V. Le , and Christopher\u00a0 D. Manning . 2020 . ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In ICLR. https:\/\/openreview.net\/pdf?id=r1xMH1BtvB Kevin Clark, Minh-Thang Luong, Quoc\u00a0V. Le, and Christopher\u00a0D. Manning. 2020. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In ICLR. https:\/\/openreview.net\/pdf?id=r1xMH1BtvB"},{"key":"e_1_3_2_1_4_1","volume-title":"NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference 1(2019)","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin , Ming\u00a0Wei Chang , Kenton Lee , and Kristina Toutanova . 2019 . BERT: Pre-training of deep bidirectional transformers for language understanding . NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference 1(2019) , 4171\u20134186. Issue Mlm. https:\/\/arxiv.org\/pdf\/ 1810.04805.pdf Jacob Devlin, Ming\u00a0Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference 1(2019), 4171\u20134186. Issue Mlm. https:\/\/arxiv.org\/pdf\/1810.04805.pdf"},{"key":"e_1_3_2_1_5_1","volume-title":"DeBERTa: Decoding-enhanced BERT with Disentangled Attention. (6","author":"He Pengcheng","year":"2020","unstructured":"Pengcheng He , Xiaodong Liu , Jianfeng Gao , and Weizhu Chen . 2020. DeBERTa: Decoding-enhanced BERT with Disentangled Attention. (6 2020 ). Pengcheng He, Xiaodong Liu, Jianfeng Gao, and Weizhu Chen. 2020. DeBERTa: Decoding-enhanced BERT with Disentangled Attention. (6 2020)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Weip\u00e9ng Hu\u00e1ng Xingyi Cheng Taifeng Wang and Wei Chu. 2019. BERT-Based Multi-Head Selection for Joint Entity-Relation Extraction. In NLPCC.  Weip\u00e9ng Hu\u00e1ng Xingyi Cheng Taifeng Wang and Wei Chu. 2019. BERT-Based Multi-Head Selection for Joint Entity-Relation Extraction. In NLPCC.","DOI":"10.1007\/978-3-030-32236-6_65"},{"key":"e_1_3_2_1_7_1","volume-title":"BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics (9","author":"Lee Jinhyuk","year":"2019","unstructured":"Jinhyuk Lee , Wonjin Yoon , Sungdong Kim , Donghyeon Kim , Sunkyu Kim , Chan\u00a0Ho So , and Jaewoo Kang . 2019. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics (9 2019 ). https:\/\/doi.org\/10.1093\/bioinformatics\/btz682 10.1093\/bioinformatics Jinhyuk Lee, Wonjin Yoon, Sungdong Kim, Donghyeon Kim, Sunkyu Kim, Chan\u00a0Ho So, and Jaewoo Kang. 2019. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics (9 2019). https:\/\/doi.org\/10.1093\/bioinformatics\/btz682"},{"key":"e_1_3_2_1_8_1","volume-title":"BioBERT: a pre-trained biomedical language representation model for biomedical text mining. (1","author":"Lee Jinhyuk","year":"2019","unstructured":"Jinhyuk Lee , Wonjin Yoon , Sungdong Kim , Donghyeon Kim , Sunkyu Kim , Chan\u00a0Ho So , and Jaewoo Kang . 2019. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. (1 2019 ). https:\/\/doi.org\/10.1093\/bioinformatics\/btz682 10.1093\/bioinformatics Jinhyuk Lee, Wonjin Yoon, Sungdong Kim, Donghyeon Kim, Sunkyu Kim, Chan\u00a0Ho So, and Jaewoo Kang. 2019. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. (1 2019). https:\/\/doi.org\/10.1093\/bioinformatics\/btz682"},{"key":"e_1_3_2_1_9_1","volume-title":"BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. (10","author":"Lewis Mike","year":"2019","unstructured":"Mike Lewis , Yinhan Liu , Naman Goyal , Marjan Ghazvininejad , Abdelrahman Mohamed , Omer Levy , Ves Stoyanov , and Luke Zettlemoyer . 2019 . BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. (10 2019). Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. 2019. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. (10 2019)."},{"key":"e_1_3_2_1_10_1","volume-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach. (7","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu , Myle Ott , Naman Goyal , Jingfei Du , Mandar Joshi , Danqi Chen , Omer Levy , Mike Lewis , Luke Zettlemoyer , and Veselin Stoyanov . 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. (7 2019 ). Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. (7 2019)."},{"key":"e_1_3_2_1_11_1","volume-title":"Decoupled Weight Decay Regularization. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Bkg6RiCqY7","author":"Loshchilov Ilya","year":"2019","unstructured":"Ilya Loshchilov and Frank Hutter . 2019 . Decoupled Weight Decay Regularization. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Bkg6RiCqY7 Ilya Loshchilov and Frank Hutter. 2019. Decoupled Weight Decay Regularization. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=Bkg6RiCqY7"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2020.103384"},{"key":"#cr-split#-e_1_3_2_1_13_1.1","doi-asserted-by":"crossref","unstructured":"Dinh\u00a0Phuong Nguyen and Tu Bao\u00a0Ho. 2020. Drug-Drug Interaction Extraction from Biomedical Texts via Relation BERT. (2020) 1-7. https:\/\/doi.org\/10.1109\/RIVF48685.2020.9140783 10.1109\/RIVF48685.2020.9140783","DOI":"10.1109\/RIVF48685.2020.9140783"},{"key":"#cr-split#-e_1_3_2_1_13_1.2","doi-asserted-by":"crossref","unstructured":"Dinh\u00a0Phuong Nguyen and Tu Bao\u00a0Ho. 2020. Drug-Drug Interaction Extraction from Biomedical Texts via Relation BERT. (2020) 1-7. https:\/\/doi.org\/10.1109\/RIVF48685.2020.9140783","DOI":"10.1109\/RIVF48685.2020.9140783"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.2979670"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1202"},{"key":"e_1_3_2_1_16_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog 1, 8","author":"Radford Alec","year":"2019","unstructured":"Alec Radford , Jeffrey Wu , Rewon Child , David Luan , Dario Amodei , Ilya Sutskever , 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 ( 2019 ), 9. Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, 2019. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9."},{"key":"e_1_3_2_1_17_1","volume-title":"a distilled version of BERT: smaller, faster, cheaper and lighter. (10","author":"Sanh Victor","year":"2019","unstructured":"Victor Sanh , Lysandre Debut , Julien Chaumond , and Thomas Wolf . 2019. DistilBERT , a distilled version of BERT: smaller, faster, cheaper and lighter. (10 2019 ). Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. (10 2019)."},{"key":"e_1_3_2_1_18_1","volume-title":"Extraction of Drug-Drug Interactions from Biomedical Texts (DDIExtraction","author":"Segura-Bedmar Isabel","year":"2013","unstructured":"Isabel Segura-Bedmar , Paloma Mart\u00ednez , and Mar\u00eda Herrero-Zazo . 2013. SemEval-2013 Task 9 : Extraction of Drug-Drug Interactions from Biomedical Texts (DDIExtraction 2013 ). (June 2013), 341\u2013350. https:\/\/aclanthology.org\/S13-2056 Isabel Segura-Bedmar, Paloma Mart\u00ednez, and Mar\u00eda Herrero-Zazo. 2013. SemEval-2013 Task 9 : Extraction of Drug-Drug Interactions from Biomedical Texts (DDIExtraction 2013). (June 2013), 341\u2013350. https:\/\/aclanthology.org\/S13-2056"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2019.103285"},{"key":"e_1_3_2_1_20_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N. Gomez Lukasz Kaiser and Illia Polosukhin. 2017. Attention Is All You Need. (2017). arxiv:1706.03762\u00a0[cs.CL]  Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N. Gomez Lukasz Kaiser and Illia Polosukhin. 2017. Attention Is All You Need. (2017). arxiv:1706.03762\u00a0[cs.CL]"},{"key":"e_1_3_2_1_21_1","unstructured":"Yonghui Wu Mike Schuster Zhifeng Chen Quoc\u00a0V. Le Mohammad Norouzi Wolfgang Macherey Maxim Krikun Yuan Cao Qin Gao Klaus Macherey Jeff Klingner Apurva Shah Melvin Johnson Xiaobing Liu \u0141ukasz Kaiser Stephan Gouws Yoshikiyo Kato Taku Kudo Hideto Kazawa Keith Stevens George Kurian Nishant Patil Wei Wang Cliff Young Jason Smith Jason Riesa Alex Rudnick Oriol Vinyals Greg Corrado Macduff Hughes and Jeffrey Dean. 2016. Google\u2019s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation. (2016) 1\u201323. arxiv:1609.08144http:\/\/arxiv.org\/abs\/1609.08144  Yonghui Wu Mike Schuster Zhifeng Chen Quoc\u00a0V. Le Mohammad Norouzi Wolfgang Macherey Maxim Krikun Yuan Cao Qin Gao Klaus Macherey Jeff Klingner Apurva Shah Melvin Johnson Xiaobing Liu \u0141ukasz Kaiser Stephan Gouws Yoshikiyo Kato Taku Kudo Hideto Kazawa Keith Stevens George Kurian Nishant Patil Wei Wang Cliff Young Jason Smith Jason Riesa Alex Rudnick Oriol Vinyals Greg Corrado Macduff Hughes and Jeffrey Dean. 2016. Google\u2019s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation. (2016) 1\u201323. arxiv:1609.08144http:\/\/arxiv.org\/abs\/1609.08144"},{"key":"e_1_3_2_1_22_1","volume-title":"Fine-tuning BERT for Joint Entity and Relation Extraction in Chinese Medical Text. 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)","author":"Xue Kui","year":"2019","unstructured":"Kui Xue , Yangming Zhou , Zhiyuan Ma , Tong Ruan , Huanhuan Zhang , and Ping He . 2019 . Fine-tuning BERT for Joint Entity and Relation Extraction in Chinese Medical Text. 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) (2019), 892\u2013897. Kui Xue, Yangming Zhou, Zhiyuan Ma, Tong Ruan, Huanhuan Zhang, and Ping He. 2019. Fine-tuning BERT for Joint Entity and Relation Extraction in Chinese Medical Text. 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) (2019), 892\u2013897."},{"key":"e_1_3_2_1_23_1","volume-title":"XLNet: Generalized Autoregressive Pretraining for Language Understanding. (6","author":"Yang Zhilin","year":"2019","unstructured":"Zhilin Yang , Zihang Dai , Yiming Yang , Jaime Carbonell , Ruslan Salakhutdinov , and Quoc\u00a0 V. Le. 2019. XLNet: Generalized Autoregressive Pretraining for Language Understanding. (6 2019 ). Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Ruslan Salakhutdinov, and Quoc\u00a0V. Le. 2019. XLNet: Generalized Autoregressive Pretraining for Language Understanding. (6 2019)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artmed.2021.102153"}],"event":{"name":"SETN 2022: 12th Hellenic Conference on Artificial Intelligence","acronym":"SETN 2022","location":"Corfu Greece"},"container-title":["Proceedings of the 12th Hellenic Conference on Artificial Intelligence"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3549737.3549753","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3549737.3549753","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:09:55Z","timestamp":1750183795000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3549737.3549753"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,9,7]]},"references-count":25,"alternative-id":["10.1145\/3549737.3549753","10.1145\/3549737"],"URL":"https:\/\/doi.org\/10.1145\/3549737.3549753","relation":{},"subject":[],"published":{"date-parts":[[2022,9,7]]},"assertion":[{"value":"2022-09-09","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}