{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T04:53:04Z","timestamp":1750913584172,"version":"3.37.3"},"reference-count":5,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,10,27]],"date-time":"2023-10-27T00:00:00Z","timestamp":1698364800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,10,27]],"date-time":"2023-10-27T00:00:00Z","timestamp":1698364800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100015956","name":"Key-Area Research and Development Program of Guangdong Province","doi-asserted-by":"publisher","award":["2019B010155003"],"award-info":[{"award-number":["2019B010155003"]}],"id":[{"id":"10.13039\/501100015956","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100021171","name":"Guangdong Basic and Applied Basic Research Foundation","doi-asserted-by":"publisher","award":["2020B1515120044"],"award-info":[{"award-number":["2020B1515120044"]}],"id":[{"id":"10.13039\/501100021171","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,10,27]]},"DOI":"10.1109\/icta60488.2023.10364277","type":"proceedings-article","created":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T19:21:00Z","timestamp":1703791260000},"page":"101-102","source":"Crossref","is-referenced-by-count":1,"title":["AttenTPU: Tensor Processor for Attention Mechanism with Fine-Grained Padding"],"prefix":"10.1109","author":[{"given":"Zhihao","family":"Du","sequence":"first","affiliation":[{"name":"Shenzhen Institute of Advanced Technology, ChineseAcademy of Sciences,Shenzhen,China"}]},{"given":"Yike","family":"Li","sequence":"additional","affiliation":[{"name":"Shenzhen Institute of Advanced Technology, ChineseAcademy of Sciences,Shenzhen,China"}]},{"given":"Chao","family":"Chen","sequence":"additional","affiliation":[{"name":"Shenzhen Institute of Advanced Technology, ChineseAcademy of Sciences,Shenzhen,China"}]},{"given":"Zheng","family":"Wang","sequence":"additional","affiliation":[{"name":"Shenzhen Institute of Advanced Technology, ChineseAcademy of Sciences,Shenzhen,China"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Attention is all you need","volume-title":"Advances in neural information processing systems","volume":"30","author":"Vaswani","year":"2017"},{"key":"ref2","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv preprint"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1145\/3604802"},{"doi-asserted-by":"publisher","key":"ref4","DOI":"10.1109\/CVPR.2018.00286"},{"doi-asserted-by":"publisher","key":"ref5","DOI":"10.18653\/v1\/w18-5446"}],"event":{"name":"2023 IEEE International Conference on Integrated Circuits, Technologies and Applications (ICTA)","start":{"date-parts":[[2023,10,27]]},"location":"Hefei, China","end":{"date-parts":[[2023,10,29]]}},"container-title":["2023 IEEE International Conference on Integrated Circuits, Technologies and Applications (ICTA)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10363774\/10363775\/10364277.pdf?arnumber=10364277","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,1]],"date-time":"2024-02-01T02:46:22Z","timestamp":1706755582000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10364277\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,27]]},"references-count":5,"URL":"https:\/\/doi.org\/10.1109\/icta60488.2023.10364277","relation":{},"subject":[],"published":{"date-parts":[[2023,10,27]]}}}