{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T18:48:27Z","timestamp":1755802107396,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":59,"publisher":"ACM","funder":[{"name":"The Shenzhen Start-Up Research Funds","award":["BL20230925"],"award-info":[{"award-number":["BL20230925"]}]},{"name":"Shenzhen Natural Science Fund","award":["GXWD20220811170436002"],"award-info":[{"award-number":["GXWD20220811170436002"]}]},{"name":"The start-up funding from the Harbin Institute of Technology, Shenzhen","award":["20210134"],"award-info":[{"award-number":["20210134"]}]},{"DOI":"10.13039\/501100006374","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["12205066"],"award-info":[{"award-number":["12205066"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]},{"name":"The Guangdong Basic and Applied Basic Research Foundation","award":["2024A1515011775"],"award-info":[{"award-number":["2024A1515011775"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,30]]},"DOI":"10.1145\/3731715.3733432","type":"proceedings-article","created":{"date-parts":[[2025,6,25]],"date-time":"2025-06-25T18:31:39Z","timestamp":1750876299000},"page":"760-769","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["SFi-Former: Sparse Flow Induced Attention for Graph Transformer"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-3170-2481","authenticated-orcid":false,"given":"Zhonghao","family":"Li","sequence":"first","affiliation":[{"name":"Harbin Institute of Technology, Shenzhen, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4889-195X","authenticated-orcid":false,"given":"Ji","family":"Shi","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology, Shenzhen, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3997-6425","authenticated-orcid":false,"given":"Xinming","family":"Zhang","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology, Shenzhen, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1262-4174","authenticated-orcid":false,"given":"Miao","family":"Zhang","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology, Shenzhen, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9743-9447","authenticated-orcid":false,"given":"Bo","family":"Li","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology, Shenzhen, Shenzhen, China"}]}],"member":"320","published-online":{"date-parts":[[2025,6,30]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Learning on Graphs Conference. PMLR, 5--1.","author":"Abboud Ralph","year":"2022","unstructured":"Ralph Abboud, Radoslav Dimitrov, and Ismail Ilkan Ceylan. 2022. Shortest path networks for graph property prediction. In Learning on Graphs Conference. PMLR, 5--1."},{"key":"e_1_3_2_1_2_1","volume-title":"On the bottleneck of graph neural networks and its practical implications. arXiv preprint arXiv:2006.05205","author":"Alon Uri","year":"2020","unstructured":"Uri Alon and Eran Yahav. 2020. On the bottleneck of graph neural networks and its practical implications. arXiv preprint arXiv:2006.05205 (2020)."},{"key":"e_1_3_2_1_3_1","volume-title":"Two-point step size gradient methods. IMA journal of numerical analysis","author":"Barzilai Jonathan","year":"1988","unstructured":"Jonathan Barzilai and Jonathan M Borwein. 1988. Two-point step size gradient methods. IMA journal of numerical analysis, Vol. 8, 1 (1988), 141--148."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637528.3672044"},{"key":"e_1_3_2_1_5_1","unstructured":"Cristian Bodnar Fabrizio Frasca Nina Otter Yuguang Wang Pietro Li\u00f2 Guido F Montufar and Michael Bronstein. 2021. Weisfeiler and Lehman go cellular: CW networks. In Advances in Neural Information Processing Systems (NeurIPS). 2625--2640."},{"key":"e_1_3_2_1_6_1","volume-title":"Improving graph neural network expressivity via subgraph isomorphism counting","author":"Bouritsas Giorgos","year":"2022","unstructured":"Giorgos Bouritsas, Fabrizio Frasca, Stefanos P Zafeiriou, and Michael Bronstein. 2022. Improving graph neural network expressivity via subgraph isomorphism counting. IEEE Transactions on Pattern Analysis and Machine Intelligence (2022)."},{"key":"e_1_3_2_1_7_1","volume-title":"Residual gated graph convnets. arXiv preprint arXiv:1711.07553","author":"Bresson Xavier","year":"2017","unstructured":"Xavier Bresson and Thomas Laurent. 2017. Residual gated graph convnets. arXiv preprint arXiv:1711.07553 (2017)."},{"key":"e_1_3_2_1_8_1","unstructured":"Michael M. Bronstein Joan Bruna Taco Cohen and Petar Veli\u010dkovi\u0107. 2021. Geometric Deep Learning: Grids Groups Graphs Geodesics and Gauges. arxiv: 2104.13478 [cs.LG] https:\/\/arxiv.org\/abs\/2104.13478"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3564269"},{"key":"e_1_3_2_1_10_1","volume-title":"Proceedings of the 38th International Conference on Machine Learning (Proceedings of Machine Learning Research","volume":"1418","author":"Chamberlain Ben","year":"2021","unstructured":"Ben Chamberlain, James Rowbottom, Maria I Gorinova, Michael Bronstein, Stefan Webb, and Emanuele Rossi. 2021. GRAND: Graph Neural Diffusion. In Proceedings of the 38th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 139), Marina Meila and Tong Zhang (Eds.). PMLR, 1407--1418."},{"key":"e_1_3_2_1_11_1","volume-title":"Proceedings of the 39th International Conference on Machine Learning (ICML).","author":"Chen Dexiong","year":"2022","unstructured":"Dexiong Chen, Leslie O'Bray, and Karsten Borgwardt. 2022b. Structure-aware transformer for graph representation learning. In Proceedings of the 39th International Conference on Machine Learning (ICML)."},{"key":"e_1_3_2_1_12_1","volume-title":"NAGphormer: A tokenized graph transformer for node classification in large graphs. arXiv preprint arXiv:2206.04910","author":"Chen Jinsong","year":"2022","unstructured":"Jinsong Chen, Kaiyuan Gao, Gaichao Li, and Kun He. 2022a. NAGphormer: A tokenized graph transformer for node classification in large graphs. arXiv preprint arXiv:2206.04910 (2022)."},{"key":"e_1_3_2_1_13_1","volume-title":"International Conference on Machine Learning. PMLR, 3648--3661","author":"Chen Qi","year":"2022","unstructured":"Qi Chen, Yifei Wang, Yisen Wang, Jiansheng Yang, and Zhouchen Lin. 2022c. Optimization-induced graph implicit nonlinear diffusion. In International Conference on Machine Learning. PMLR, 3648--3661."},{"key":"e_1_3_2_1_14_1","volume-title":"9th International Conference on Learning Representations (ICLR).","author":"Choromanski Krzysztof Marcin","year":"2021","unstructured":"Krzysztof Marcin Choromanski, Valerii Likhosherstov, David Dohan, Xingyou Song, Andreea Gane, Tam\u00e1s Sarl\u00f3s, Peter Hawkins, Jared Quincy Davis, Afroz Mohiuddin, Lukasz Kaiser, David Benjamin Belanger, Lucy J Colwell, and Adrian Weller. 2021. Rethinking attention with performers. In 9th International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19--1223"},{"key":"e_1_3_2_1_16_1","volume-title":"Garnett (Eds.)","volume":"29","author":"Defferrard Micha\u00ebl","year":"2016","unstructured":"Micha\u00ebl Defferrard, Xavier Bresson, and Pierre Vandergheynst. 2016. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering. In Advances in Neural Information Processing Systems, D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett (Eds.), Vol. 29. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2016\/file\/04df4d434d481c5bb723be1b6df1ee65-Paper.pdf"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"P.G. Doyle and J.L. Snell. 1984. Random walks and electric networks. Mathematical Association of America. 84061495","DOI":"10.5948\/UPO9781614440222"},{"key":"e_1_3_2_1_18_1","volume-title":"AAAI Workshop on Deep Learning on Graphs: Methods and Applications","author":"Dwivedi Vijay Prakash","year":"2021","unstructured":"Vijay Prakash Dwivedi and Xavier Bresson. 2021. A Generalization of Transformer Networks to Graphs. AAAI Workshop on Deep Learning on Graphs: Methods and Applications (2021)."},{"key":"e_1_3_2_1_19_1","first-page":"1","article-title":"Benchmarking graph neural networks","volume":"24","author":"Dwivedi Vijay Prakash","year":"2023","unstructured":"Vijay Prakash Dwivedi, Chaitanya K Joshi, Anh Tuan Luu, Thomas Laurent, Yoshua Bengio, and Xavier Bresson. 2023. Benchmarking graph neural networks. Journal of Machine Learning Research, Vol. 24, 43 (2023), 1--48.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_1_20_1","volume-title":"International Conference on Learning Representations (ICLR).","author":"Dwivedi Vijay Prakash","year":"2022","unstructured":"Vijay Prakash Dwivedi, Anh Tuan Luu, Thomas Laurent, Yoshua Bengio, and Xavier Bresson. 2022a. Graph neural networks with learnable structural and positional representations. In International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_21_1","volume-title":"Anh Tuan Luu, and Dominique Beaini","author":"Dwivedi Vijay Prakash","year":"2022","unstructured":"Vijay Prakash Dwivedi, Ladislav Ramp\u00e1\u0161ek, Mikhail Galkin, Ali Parviz, Guy Wolf, Anh Tuan Luu, and Dominique Beaini. 2022b. Long range graph benchmark. In Neural Information Processing Systems (NeurIPS 2022), Track on Datasets and Benchmarks."},{"key":"e_1_3_2_1_22_1","volume-title":"Convit: Improving Vision Transformers with Soft Convolutional Inductive Biases. In International Conference on Machine Learning. PMLR, 2286--2296","author":"Ascoli St\u00e9phane","year":"2021","unstructured":"St\u00e9phane d'Ascoli, Hugo Touvron, Matthew L Leavitt, Ari S Morcos, Giulio Biroli, and Levent Sagun. 2021. Convit: Improving Vision Transformers with Soft Convolutional Inductive Biases. In International Conference on Machine Learning. PMLR, 2286--2296."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3586074"},{"key":"e_1_3_2_1_24_1","volume-title":"Yanfei Dong, and Lee Wee Sun.","author":"Fu Guoji","year":"2023","unstructured":"Guoji Fu, Mohammed Haroon Dupty, Yanfei Dong, and Lee Wee Sun. 2023. Implicit graph neural diffusion based on constrained Dirichlet energy minimization. arXiv preprint arXiv:2308.03306 (2023)."},{"key":"e_1_3_2_1_25_1","volume-title":"CMT: Convolutional neural networks meet vision transformers. arXiv preprint arXiv:2107.06263","author":"Guo Jianyuan","year":"2021","unstructured":"Jianyuan Guo, Kai Han, Han Wu, Chang Xu, Yehui Tang, Chunjing Xu, and Yunhe Wang. 2021. CMT: Convolutional neural networks meet vision transformers. arXiv preprint arXiv:2107.06263 (2021)."},{"key":"e_1_3_2_1_26_1","volume-title":"Inductive representation learning on large graphs. Advances in neural information processing systems","author":"Hamilton Will","year":"2017","unstructured":"Will Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. Advances in neural information processing systems, Vol. 30 (2017)."},{"key":"e_1_3_2_1_27_1","volume-title":"From continuous dynamics to graph neural networks: Neural diffusion and beyond. arXiv preprint arXiv:2310.10121","author":"Han Andi","year":"2023","unstructured":"Andi Han, Dai Shi, Lequan Lin, and Junbin Gao. 2023. From continuous dynamics to graph neural networks: Neural diffusion and beyond. arXiv preprint arXiv:2310.10121 (2023)."},{"key":"e_1_3_2_1_28_1","unstructured":"Kai Han Yunhe Wang Hanting Chen Xinghao Chen Jianyuan Guo Zhenhua Liu Yehui Tang An Xiao Chunjing Xu Yixing Xu et al. 2022. A survey on vision transformer. IEEE Transactions on Pattern Analysis and Machine Intelligence (2022)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-0--387--84858--7"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_31_1","volume-title":"Ammus: A survey of transformer-based pretrained models in natural language processing. arXiv preprint arXiv:2108.05542","author":"Kalyan Katikapalli Subramanyam","year":"2021","unstructured":"Katikapalli Subramanyam Kalyan, Ajit Rajasekharan, and Sivanesan Sangeetha. 2021. Ammus: A survey of transformer-based pretrained models in natural language processing. arXiv preprint arXiv:2108.05542 (2021)."},{"key":"e_1_3_2_1_32_1","volume-title":"Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907","author":"Kipf Thomas N","year":"2016","unstructured":"Thomas N Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)."},{"key":"e_1_3_2_1_33_1","volume-title":"8th International Conference on Learning Representations (ICLR).","author":"Kitaev Nikita","year":"2020","unstructured":"Nikita Kitaev, Lukasz Kaiser, and Anselm Levskaya. 2020. Reformer: The efficient transformer. In 8th International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_34_1","volume-title":"Text generation from knowledge graphs with graph transformers. arXiv preprint arXiv:1904.02342","author":"Koncel-Kedziorski Rik","year":"2019","unstructured":"Rik Koncel-Kedziorski, Dhanush Bekal, Yi Luan, Mirella Lapata, and Hannaneh Hajishirzi. 2019. Text generation from knowledge graphs with graph transformers. arXiv preprint arXiv:1904.02342 (2019)."},{"key":"e_1_3_2_1_35_1","volume-title":"Wortman Vaughan (Eds.)","volume":"34","author":"Kreuzer Devin","year":"2021","unstructured":"Devin Kreuzer, Dominique Beaini, Will Hamilton, Vincent L\u00e9tourneau, and Prudencio Tossou. 2021a. Rethinking Graph Transformers with Spectral Attention. In Advances in Neural Information Processing Systems, M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (Eds.), Vol. 34. Curran Associates, Inc., 21618--21629. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2021\/file\/b4fd1d2cb085390fbbadae65e07876a7-Paper.pdf"},{"key":"e_1_3_2_1_36_1","unstructured":"Devin Kreuzer Dominique Beaini William L. Hamilton Vincent L\u00e9tourneau and Prudencio Tossou. 2021b. Rethinking graph transformers with spectral attention. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3591723"},{"key":"e_1_3_2_1_38_1","volume-title":"International Conference on Machine Learning. PMLR, 6837--6849","author":"Liu Xiaorui","year":"2021","unstructured":"Xiaorui Liu, Wei Jin, Yao Ma, Yaxin Li, Hua Liu, Yiqi Wang, Ming Yan, and Jiliang Tang. 2021. Elastic graph neural networks. In International Conference on Machine Learning. PMLR, 6837--6849."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.7717\/peerj-cs.357"},{"key":"e_1_3_2_1_40_1","volume-title":"A graph vae and graph transformer approach to generating molecular graphs. arXiv preprint arXiv:2104.04345","author":"Mitton Joshua","year":"2021","unstructured":"Joshua Mitton, Hans M Senn, Klaas Wynne, and Roderick Murray-Smith. 2021. A graph vae and graph transformer approach to generating molecular graphs. arXiv preprint arXiv:2104.04345 (2021)."},{"key":"e_1_3_2_1_41_1","volume-title":"Attending to Graph Transformers. Transactions on Machine Learning Research","author":"M\u00fcller Luis","year":"2024","unstructured":"Luis M\u00fcller, Mikhail Galkin, Christopher Morris, and Ladislav Ramp\u00e1\u0161ek. 2024. Attending to Graph Transformers. Transactions on Machine Learning Research (2024). https:\/\/openreview.net\/forum?id=HhbqHBBrfZ"},{"key":"e_1_3_2_1_42_1","volume-title":"Graph neural networks exponentially lose expressive power for node classification. arXiv preprint arXiv:1905.10947","author":"Oono Kenta","year":"2019","unstructured":"Kenta Oono and Taiji Suzuki. 2019. Graph neural networks exponentially lose expressive power for node classification. arXiv preprint arXiv:1905.10947 (2019)."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1561\/2400000003"},{"key":"e_1_3_2_1_44_1","volume-title":"Giacomo Wolf, and Dominique Beaini.","author":"Rampasek Lukas","year":"2022","unstructured":"Lukas Rampasek, Mikhail Galkin, Vijay P Dwivedi, Anh Tuan Luu, Giacomo Wolf, and Dominique Beaini. 2022. Recipe for a general, powerful, scalable graph transformer. CoRR, Vol. abs\/2205.12454 (2022)."},{"key":"e_1_3_2_1_45_1","volume-title":"Oh (Eds.)","volume":"35","author":"Ramp\u00e1\u0161ek Ladislav","year":"2022","unstructured":"Ladislav Ramp\u00e1\u0161ek, Michael Galkin, Vijay Prakash Dwivedi, Anh Tuan Luu, Guy Wolf, and Dominique Beaini. 2022. Recipe for a General, Powerful, Scalable Graph Transformer. In Advances in Neural Information Processing Systems, S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh (Eds.), Vol. 35. Curran Associates, Inc., 14501--14515. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2022\/file\/5d4834a159f1547b267a05a4e2b7cf5e-Paper-Conference.pdf"},{"key":"e_1_3_2_1_46_1","first-page":"1","article-title":"A new approach to Laplacian solvers and flow problems","volume":"20","author":"Rebeschini Patrick","year":"2019","unstructured":"Patrick Rebeschini and Sekhar Tatikonda. 2019. A new approach to Laplacian solvers and flow problems. Journal of Machine Learning Research, Vol. 20, 36 (2019), 1--37.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11647"},{"key":"e_1_3_2_1_48_1","volume-title":"Exphormer: Sparse Transformers for Graphs. arxiv: 2303.06147 [cs.LG] https:\/\/arxiv.org\/abs\/2303.06147","author":"Shirzad Hamed","year":"2023","unstructured":"Hamed Shirzad, Ameya Velingker, Balaji Venkatachalam, Danica J. Sutherland, and Ali Kemal Sinop. 2023. Exphormer: Sparse Transformers for Graphs. arxiv: 2303.06147 [cs.LG] https:\/\/arxiv.org\/abs\/2303.06147"},{"key":"e_1_3_2_1_49_1","volume-title":"Advances in Neural Information Processing Systems","volume":"30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is All You Need. In Advances in Neural Information Processing Systems, Vol. 30. Curran Associates, Inc."},{"key":"e_1_3_2_1_50_1","volume-title":"Graph attention networks. arXiv preprint arXiv:1710.10903","author":"Veli\u010dkovi\u0107 Petar","year":"2017","unstructured":"Petar Veli\u010dkovi\u0107, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. 2017. Graph attention networks. arXiv preprint arXiv:1710.10903 (2017)."},{"key":"e_1_3_2_1_51_1","volume-title":"Graph-mamba: Towards long-range graph sequence modeling with selective state spaces. arXiv preprint arXiv:2402.00789","author":"Wang Chloe","year":"2024","unstructured":"Chloe Wang, Oleksii Tsepa, Jun Ma, and Bo Wang. 2024. Graph-mamba: Towards long-range graph sequence modeling with selective state spaces. arXiv preprint arXiv:2402.00789 (2024)."},{"volume-title":"High-Dimensional Data Analysis with Low-Dimensional Models: Principles, Computation, and Applications","author":"Wright John","key":"e_1_3_2_1_52_1","unstructured":"John Wright and Yi Ma. 2022. High-Dimensional Data Analysis with Low-Dimensional Models: Principles, Computation, and Applications. Cambridge University Press."},{"key":"e_1_3_2_1_53_1","volume-title":"Difformer: Scalable (graph) transformers induced by energy constrained diffusion. arXiv preprint arXiv:2301.09474","author":"Wu Qitian","year":"2023","unstructured":"Qitian Wu, Chenxiao Yang, Wentao Zhao, Yixuan He, David Wipf, and Junchi Yan. 2023. Difformer: Scalable (graph) transformers induced by energy constrained diffusion. arXiv preprint arXiv:2301.09474 (2023)."},{"key":"e_1_3_2_1_54_1","volume-title":"Forty-first International Conference on Machine Learning. https:\/\/openreview.net\/forum?id=uKmcyyrZae","author":"Xing Yujie","year":"2024","unstructured":"Yujie Xing, Xiao Wang, Yibo Li, Hai Huang, and Chuan Shi. 2024. Less is More: on the Over-Globalizing Problem in Graph Transformers. In Forty-first International Conference on Machine Learning. https:\/\/openreview.net\/forum?id=uKmcyyrZae"},{"key":"e_1_3_2_1_55_1","volume-title":"How powerful are graph neural networks? arXiv preprint arXiv:1810.00826","author":"Xu Keyulu","year":"2018","unstructured":"Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2018. How powerful are graph neural networks? arXiv preprint arXiv:1810.00826 (2018)."},{"key":"e_1_3_2_1_56_1","first-page":"905","article-title":"Sparse graph attention networks","volume":"35","author":"Ye Yang","year":"2021","unstructured":"Yang Ye and Shihao Ji. 2021. Sparse graph attention networks. IEEE Transactions on Knowledge and Data Engineering, Vol. 35, 1 (2021), 905--916.","journal-title":"IEEE Transactions on Knowledge and Data Engineering"},{"key":"e_1_3_2_1_57_1","unstructured":"Chengxuan Ying Tianle Cai Shengjie Luo Shuxin Zheng Guolin Ke Di He Yanming Shen and Tie-Yan Liu. 2021. Do transformers really perform badly for graph representation?. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_58_1","volume-title":"Joshua Ainslie, Chris Alberti, Santiago Onta\u00f1\u00f3n","author":"Zaheer Manzil","year":"2020","unstructured":"Manzil Zaheer, Guru Guruganesh, Kumar Avinava Dubey, Joshua Ainslie, Chris Alberti, Santiago Onta\u00f1\u00f3n, Philip Pham, Anirudh Ravula, Qifan Wang, Li Yang, and Amr Ahmed. 2020. Big Bird: Transformers for longer sequences. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1186\/s40649-019-0069-y"}],"event":{"name":"ICMR '25: International Conference on Multimedia Retrieval","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Chicago IL USA","acronym":"ICMR '25"},"container-title":["Proceedings of the 2025 International Conference on Multimedia Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3731715.3733432","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T04:13:35Z","timestamp":1755749615000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3731715.3733432"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,30]]},"references-count":59,"alternative-id":["10.1145\/3731715.3733432","10.1145\/3731715"],"URL":"https:\/\/doi.org\/10.1145\/3731715.3733432","relation":{},"subject":[],"published":{"date-parts":[[2025,6,30]]},"assertion":[{"value":"2025-06-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}