{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,7]],"date-time":"2026-01-07T18:44:19Z","timestamp":1767811459222,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":16,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,1,16]],"date-time":"2023-01-16T00:00:00Z","timestamp":1673827200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U19B2019, U21B2031, 61832007, 62104128"],"award-info":[{"award-number":["U19B2019, U21B2031, 61832007, 62104128"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Tsinghua EE Xilinx AI Research Fund"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,1,16]]},"DOI":"10.1145\/3566097.3567869","type":"proceedings-article","created":{"date-parts":[[2023,1,31]],"date-time":"2023-01-31T18:40:49Z","timestamp":1675190449000},"page":"645-650","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["NTGAT"],"prefix":"10.1145","author":[{"given":"Wentao","family":"Hou","sequence":"first","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"given":"Kai","family":"Zhong","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"given":"Shulin","family":"Zeng","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"given":"Guohao","family":"Dai","sequence":"additional","affiliation":[{"name":"Shanghai Jiao Tong University, Shanghai, China"}]},{"given":"Huazhong","family":"Yang","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"given":"Yu","family":"Wang","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2023,1,31]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Bingyi Zhang et al. 2021. Boostgcn:A framework for optimizing gcn inference on fpga. In FCCM. IEEE.","DOI":"10.1109\/FCCM51124.2021.00012"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Guyue Huang et al. 2020. GE-SpMM: General-purpose Sparse Matrix-Matrix Multiplication on GPUs for Graph Neural Networks. ArXiv (2020).","DOI":"10.1109\/SC41405.2020.00076"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Hengrui Zhang et al. 2021. Understanding GNN Computational Graph: A Coordinated Computation IO and Memory Perspective. ArXiv (2021).","DOI":"10.1155\/2021\/6882467"},{"key":"e_1_3_2_1_4_1","volume-title":"Resolve: Generation of high-performance sorting architectures from high-level synthesis. In FPGA.","author":"Janarbek Matai","year":"2016","unstructured":"Janarbek Matai et al. 2016. Resolve: Generation of high-performance sorting architectures from high-level synthesis. In FPGA."},{"key":"e_1_3_2_1_5_1","unstructured":"Minjie Wang et al. 2019. Deep Graph Library: Towards Efficient and Scalable Deep Learning on Graphs. ArXiv (2019)."},{"key":"e_1_3_2_1_6_1","volume-title":"Hygcn: A gcn accelerator with hybrid architecture","author":"Mingyu Yan","year":"2020","unstructured":"Mingyu Yan et al. 2020. Hygcn: A gcn accelerator with hybrid architecture. In HPCA. IEEE."},{"key":"e_1_3_2_1_7_1","unstructured":"Petar Veli\u010dkovi\u0107 et al. 2017. Graph attention networks. arXiv (2017)."},{"key":"e_1_3_2_1_8_1","volume-title":"Engn: A high-throughput and energy-efficient accelerator for large graph neural networks","author":"Shengwen Liang","year":"2020","unstructured":"Shengwen Liang et al. 2020. Engn: A high-throughput and energy-efficient accelerator for large graph neural networks. IEEE Trans. Comput. (2020)."},{"key":"e_1_3_2_1_9_1","unstructured":"Shuangli Li et al. 2021. Structure-aware Interactive Graph Neural Networks for the Prediction of Protein-Ligand Binding Affinity. SIGKDD (2021)."},{"key":"e_1_3_2_1_10_1","volume-title":"Social-bigat: Multimodal trajectory forecasting using bicycle-gan and graph attention networks. Advances in Neural Information Processing Systems","author":"Vineet Kosaraju","year":"2019","unstructured":"Vineet Kosaraju et al. 2019. Social-bigat: Multimodal trajectory forecasting using bicycle-gan and graph attention networks. Advances in Neural Information Processing Systems (2019)."},{"key":"e_1_3_2_1_11_1","unstructured":"Will Hamilton et al. 2017. Inductive representation learning on large graphs. (2017)."},{"key":"e_1_3_2_1_12_1","unstructured":"Weihua Hu et al. 2020. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems (2020)."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"Weian Yan et al. 2020. FPGAN: an FPGA accelerator for graph attention networks with software and hardware co-optimization. IEEE Access (2020).","DOI":"10.1109\/ACCESS.2020.3023946"},{"key":"e_1_3_2_1_14_1","unstructured":"Zhongming Yu et al. 2021. Exploiting Online Locality and Reduction Parallelism for Sampled Dense Matrix Multiplication on GPUs. ICCD (2021)."},{"key":"e_1_3_2_1_15_1","volume-title":"Oh","author":"Kim Dongkwan","year":"2021","unstructured":"Dongkwan Kim and Alice H. Oh. 2021. How to Find Your Friendly Neighborhood: Graph Attention Design with Self-Supervision. ArXiv (2021)."},{"key":"e_1_3_2_1_16_1","volume-title":"Semi-supervised classification with graph convolutional networks. arXiv","author":"Kipf Thomas N","year":"2016","unstructured":"Thomas N Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. arXiv (2016)."}],"event":{"name":"ASPDAC '23: 28th Asia and South Pacific Design Automation Conference","location":"Tokyo Japan","acronym":"ASPDAC '23","sponsor":["SIGDA ACM Special Interest Group on Design Automation","IEEE CEDA","IEICE","IEEE CAS","IPSJ"]},"container-title":["Proceedings of the 28th Asia and South Pacific Design Automation Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3566097.3567869","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3566097.3567869","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,7]],"date-time":"2026-01-07T17:33:25Z","timestamp":1767807205000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3566097.3567869"}},"subtitle":["A Graph Attention Network Accelerator with Runtime Node Tailoring"],"short-title":[],"issued":{"date-parts":[[2023,1,16]]},"references-count":16,"alternative-id":["10.1145\/3566097.3567869","10.1145\/3566097"],"URL":"https:\/\/doi.org\/10.1145\/3566097.3567869","relation":{},"subject":[],"published":{"date-parts":[[2023,1,16]]},"assertion":[{"value":"2023-01-31","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}