{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T16:26:59Z","timestamp":1769876819341,"version":"3.49.0"},"publisher-location":"Stroudsburg, PA, USA","reference-count":0,"publisher":"Association for Computational Linguistics","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.18653\/v1\/k19-1079","type":"proceedings-article","created":{"date-parts":[[2019,11,2]],"date-time":"2019-11-02T20:28:40Z","timestamp":1572726520000},"page":"843-861","source":"Crossref","is-referenced-by-count":38,"title":["Do Massively Pretrained Language Models Make Better Storytellers?"],"prefix":"10.18653","author":[{"given":"Abigail","family":"See","sequence":"first","affiliation":[]},{"given":"Aneesh","family":"Pappu","sequence":"additional","affiliation":[]},{"given":"Rohun","family":"Saxena","sequence":"additional","affiliation":[]},{"given":"Akhila","family":"Yerukola","sequence":"additional","affiliation":[]},{"given":"Christopher D.","family":"Manning","sequence":"additional","affiliation":[]}],"member":"1643","event":{"name":"Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL)","location":"Hong Kong, China","start":{"date-parts":[[2019,11]]},"end":{"date-parts":[[2019,11]]}},"container-title":["Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL)"],"original-title":[],"deposited":{"date-parts":[[2019,12,19]],"date-time":"2019-12-19T17:15:30Z","timestamp":1576775730000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.aclweb.org\/anthology\/K19-1079"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"references-count":0,"URL":"https:\/\/doi.org\/10.18653\/v1\/k19-1079","relation":{},"subject":[],"published":{"date-parts":[[2019]]}}}