{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,29]],"date-time":"2025-11-29T08:03:22Z","timestamp":1764403402820},"reference-count":35,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,8,6]],"date-time":"2024-08-06T00:00:00Z","timestamp":1722902400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,8,6]],"date-time":"2024-08-06T00:00:00Z","timestamp":1722902400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,8,6]]},"DOI":"10.1109\/ccece59415.2024.10667143","type":"proceedings-article","created":{"date-parts":[[2024,9,12]],"date-time":"2024-09-12T17:42:46Z","timestamp":1726162966000},"page":"43-48","source":"Crossref","is-referenced-by-count":1,"title":["Mixture of Prompt Experts for Natural Language Inference"],"prefix":"10.1109","author":[{"given":"Zi\u2019ou","family":"Zheng","sequence":"first","affiliation":[{"name":"Queen&#x2019;s University,Ingenuity Labs Research Institute,Department of Electrical and Computer Engineering,Kingston,Canada"}]},{"given":"Xiaodan","family":"Zhu","sequence":"additional","affiliation":[{"name":"Queen&#x2019;s University,Ingenuity Labs Research Institute,Department of Electrical and Computer Engineering,Kingston,Canada"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-acl.91"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-emnlp.324"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00453"},{"article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","year":"2018","author":"Devlin","key":"ref4"},{"article-title":"Roberta: A robustly optimized bert pretraining approach","year":"2019","author":"Liu","key":"ref5"},{"article-title":"Language models are unsupervised multitask learners","year":"2019","author":"Radford","key":"ref6"},{"key":"ref7","article-title":"Language models are few-shot learners","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume-title":"JMLR","author":"Raffel","key":"ref8"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.aiopen.2023.08.012"},{"key":"ref12","article-title":"P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks","author":"Liu","year":"2021","journal-title":"CoRR"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.410"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/w19-4804"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6397"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1152"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D15-1075"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1101"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P18-2103"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/S19-1027"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.543"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/K19-1019"},{"article-title":"Stress test evaluation for natural language inference","year":"2018","author":"Naik","key":"ref23"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W18-5441"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.554"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.starsem-1.41"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.blackboxnlp-1.26"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.889"},{"key":"ref29","article-title":"Exploiting cloze questions for few-shot\u00a8 text classification and natural language inference","author":"Schick","year":"2020","journal-title":"Computing Research Repository"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.185"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.346"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.884"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.587"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1023\/A:1025164703468"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.33011\/lilt.v9i.1325"}],"event":{"name":"2024 IEEE Canadian Conference on Electrical and Computer Engineering (CCECE)","start":{"date-parts":[[2024,8,6]]},"location":"Kingston, ON, Canada","end":{"date-parts":[[2024,8,9]]}},"container-title":["2024 IEEE Canadian Conference on Electrical and Computer Engineering (CCECE)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10667055\/10667056\/10667143.pdf?arnumber=10667143","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,13]],"date-time":"2024-09-13T05:16:42Z","timestamp":1726204602000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10667143\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,6]]},"references-count":35,"URL":"https:\/\/doi.org\/10.1109\/ccece59415.2024.10667143","relation":{},"subject":[],"published":{"date-parts":[[2024,8,6]]}}}