{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T12:25:25Z","timestamp":1775910325182,"version":"3.50.1"},"publisher-location":"Stroudsburg, PA, USA","reference-count":0,"publisher":"Association for Computational Linguistics","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.18653\/v1\/d19-1441","type":"proceedings-article","created":{"date-parts":[[2019,11,3]],"date-time":"2019-11-03T02:21:11Z","timestamp":1572747671000},"page":"4322-4331","source":"Crossref","is-referenced-by-count":321,"title":["Patient Knowledge Distillation for BERT Model Compression"],"prefix":"10.18653","author":[{"given":"Siqi","family":"Sun","sequence":"first","affiliation":[]},{"given":"Yu","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Zhe","family":"Gan","sequence":"additional","affiliation":[]},{"given":"Jingjing","family":"Liu","sequence":"additional","affiliation":[]}],"member":"1643","event":{"name":"Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)","location":"Hong Kong, China","start":{"date-parts":[[2019,11]]},"end":{"date-parts":[[2019,11]]}},"container-title":["Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)"],"original-title":[],"deposited":{"date-parts":[[2019,12,19]],"date-time":"2019-12-19T22:13:10Z","timestamp":1576793590000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.aclweb.org\/anthology\/D19-1441"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"references-count":0,"URL":"https:\/\/doi.org\/10.18653\/v1\/d19-1441","relation":{},"subject":[],"published":{"date-parts":[[2019]]}}}