{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,8]],"date-time":"2025-11-08T12:36:42Z","timestamp":1762605402392,"version":"build-2065373602"},"publisher-location":"Stroudsburg, PA, USA","reference-count":0,"publisher":"Association for Computational Linguistics","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.18653\/v1\/2025.emnlp-main.118","type":"proceedings-article","created":{"date-parts":[[2025,11,8]],"date-time":"2025-11-08T12:29:45Z","timestamp":1762604985000},"page":"2337-2354","source":"Crossref","is-referenced-by-count":0,"title":["Integral Transformer: Denoising Attention, Not Too Much Not Too Little"],"prefix":"10.18653","author":[{"given":"Ivan","family":"Kobyzev","sequence":"first","affiliation":[]},{"given":"Abbas","family":"Ghaddar","sequence":"additional","affiliation":[]},{"given":"Dingtao","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Boxing","family":"Chen","sequence":"additional","affiliation":[]}],"member":"1643","event":{"name":"Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing","location":"Suzhou, China","start":{"date-parts":[[2025,11]]},"end":{"date-parts":[[2025,11]]}},"container-title":["Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing"],"original-title":[],"deposited":{"date-parts":[[2025,11,8]],"date-time":"2025-11-08T12:31:39Z","timestamp":1762605099000},"score":1,"resource":{"primary":{"URL":"https:\/\/aclanthology.org\/2025.emnlp-main.118"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":0,"URL":"https:\/\/doi.org\/10.18653\/v1\/2025.emnlp-main.118","relation":{},"subject":[],"published":{"date-parts":[[2025]]}}}