{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,12,21]],"date-time":"2024-12-21T06:40:15Z","timestamp":1734763215056,"version":"3.32.0"},"reference-count":58,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,10,18]],"date-time":"2024-10-18T00:00:00Z","timestamp":1729209600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,10,18]],"date-time":"2024-10-18T00:00:00Z","timestamp":1729209600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,10,18]]},"DOI":"10.1109\/mlnlp63328.2024.10800230","type":"proceedings-article","created":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T18:56:07Z","timestamp":1734720967000},"page":"1-11","source":"Crossref","is-referenced-by-count":0,"title":["Ophthalmic Disease Zero-Shot Question Answering Through Knowledge Triple Augmented Language Model Prompting"],"prefix":"10.1109","author":[{"given":"Zihan","family":"Wang","sequence":"first","affiliation":[{"name":"Shanghai Jiao Tong University,Shanghai,China"}]},{"given":"Hong","family":"Huo","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,Shanghai,China"}]},{"given":"Renxin","family":"Xu","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,Shanghai,China"}]},{"given":"Shijie","family":"Yang","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,Shanghai,China"}]},{"given":"Tao","family":"Fang","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University,Shanghai,China"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Bert: Pretraining of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv preprint"},{"journal-title":"Improving language understanding by generative pretraining","year":"2018","author":"Radford","key":"ref2"},{"issue":"8","key":"ref3","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI blog"},{"key":"ref4","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref5","article-title":"Llama: Open and efficient foundation language models","author":"Touvron","year":"2023","journal-title":"arXiv preprint"},{"key":"ref6","article-title":"Publicly available clinical bert embeddings","author":"Alsentzer","year":"2019","journal-title":"arXiv preprint"},{"issue":"1.3","key":"ref7","first-page":"8","article-title":"Scibert: Pretrained contextu-alized embeddings for scientific text","volume-title":"arXiv preprint","volume":"1","author":"Beltagy","year":"2019"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btz682"},{"key":"ref9","article-title":"Clinical concept extraction with contextual word embedding","author":"Zhu","year":"2018","journal-title":"arXiv preprint"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1437"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380192"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401416"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.519"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401075"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.550"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401271"},{"issue":"10","key":"ref17","article-title":"Survey of open-domain knowledge graph question answering","volume":"15","author":"Zirui","year":"2021","journal-title":"Journal of Frontiers of Computer Science & Technology"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D13-1160"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P14-1133"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-1009"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P14-1090"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/N15-3014"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P14-1091"},{"key":"ref24","doi-asserted-by":"crossref","first-page":"1301","DOI":"10.1145\/2806416.2806542","article-title":"An-swering questions with complex semantic constraints on open knowledge bases","volume-title":"Proceedings of the 24th ACM international on conference on information and knowledge management","author":"Yin","year":"2015"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.14778\/3055540.3055549"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1145\/2588555.2610525"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/2187836.2187923"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/2806416.2806472"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1145\/2911451.2911536"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1067"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.12057"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.412"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1147"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1301"},{"key":"ref35","article-title":"Memory networks","author":"Weston","year":"2014","journal-title":"arXiv preprint"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P14-2105"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1015"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-1091"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00262"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N16-1181"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/p15-1128"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.91"},{"key":"ref43","first-page":"2503","article-title":"Constraint-based question answering with knowledge graph","volume-title":"Proceedings of COLING 2016, the 26th international conference on computational linguistics: technical papers","author":"Bao","year":"2016"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P15-1026"},{"key":"ref45","article-title":"Neural symbolic machines: Learning semantic parsers on freebase with weak supervision","author":"Liang","year":"2016","journal-title":"arXiv preprint"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1059"},{"key":"ref47","first-page":"727","article-title":"Pretrained language model for biomedical question answering","volume-title":"Joint European conference on machine learning and knowledge discovery in databases","author":"Yoon","year":"2019"},{"key":"ref48","article-title":"Doctorglm: Fine-tuning your chinese doctor is not a herculean task","author":"Xiong","year":"2023","journal-title":"arXiv preprint"},{"key":"ref49","article-title":"Chatdoctor: A medical chat model fine-tuned on llama model using medical domain knowledge","author":"Yunxiang","year":"2023","journal-title":"arXiv preprint"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.634"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1139"},{"key":"ref52","article-title":"Knowledge-aware language model pretraining","author":"Rosset","year":"2020","journal-title":"arXiv preprint"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00360"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.764"},{"key":"ref55","article-title":"Pretrained encyclopedia: Weakly supervised knowledge-pretrained language model","author":"Xiong","year":"2019","journal-title":"arXiv preprint"},{"key":"ref56","first-page":"9459","article-title":"Retrieval-augmented generation for knowledge-intensive nlp tasks","volume":"33","author":"Lewis","year":"2020","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref57","first-page":"16857","article-title":"Mp-net: Masked and permuted pretraining for language understanding","volume":"33","author":"Song","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref58","article-title":"Knowledge-consistent dialogue generation with knowledge graphs","volume-title":"ICML 2022 Workshop on Knowledge Retrieval and Language Models","author":"Kang","year":"2022"}],"event":{"name":"2024 7th International Conference on Machine Learning and Natural Language Processing (MLNLP)","start":{"date-parts":[[2024,10,18]]},"location":"Chengdu, China","end":{"date-parts":[[2024,10,20]]}},"container-title":["2024 7th International Conference on Machine Learning and Natural Language Processing (MLNLP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10799945\/10799990\/10800230.pdf?arnumber=10800230","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,21]],"date-time":"2024-12-21T05:58:04Z","timestamp":1734760684000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10800230\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,18]]},"references-count":58,"URL":"https:\/\/doi.org\/10.1109\/mlnlp63328.2024.10800230","relation":{},"subject":[],"published":{"date-parts":[[2024,10,18]]}}}