{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,4]],"date-time":"2026-05-04T05:43:29Z","timestamp":1777873409369,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":105,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,8,3]]},"DOI":"10.1145\/3711896.3736558","type":"proceedings-article","created":{"date-parts":[[2025,8,3]],"date-time":"2025-08-03T20:52:41Z","timestamp":1754254361000},"page":"5999-6009","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Relational Deep Learning: Challenges, Foundations and Next-Generation Architectures"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1836-4323","authenticated-orcid":false,"given":"Vijay Prakash","family":"Dwivedi","sequence":"first","affiliation":[{"name":"Stanford University, Stanford, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0952-1561","authenticated-orcid":false,"given":"Charilaos","family":"Kanatsoulis","sequence":"additional","affiliation":[{"name":"Stanford University, Stanford, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4526-7097","authenticated-orcid":false,"given":"Shenyang","family":"Huang","sequence":"additional","affiliation":[{"name":"Mila, McGill University, Montreal, Quebec, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5411-923X","authenticated-orcid":false,"given":"Jure","family":"Leskovec","sequence":"additional","affiliation":[{"name":"Stanford University, Stanford, CA, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,8,3]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al., 2023. Gpt-4 technical report. arXiv preprint arXiv:2303.08774(2023)."},{"key":"e_1_3_2_1_2_1","volume-title":"The claude 3 model family: Opus, sonnet, haiku. Claude-3 Model Card","author":"Anthropic AI","year":"2024","unstructured":"AI Anthropic. 2024. The claude 3 model family: Opus, sonnet, haiku. Claude-3 Model Card, Vol. 1 (2024), 1."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i8.16826"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jcss.2020.04.003"},{"key":"e_1_3_2_1_5_1","unstructured":"Jinze Bai Shuai Bai Yunfei Chu Zeyu Cui Kai Dang Xiaodong Deng Yang Fan Wenbin Ge Yu Han Fei Huang et al. 2023. Qwen technical report. arXiv preprint arXiv:2309.16609(2023)."},{"key":"e_1_3_2_1_6_1","volume-title":"Learning on Graphs Conference. PMLR, 15-1.","author":"Besta Maciej","year":"2024","unstructured":"Maciej Besta, Afonso Claudino Catarino, Lukas Gianinazzi, Nils Blach, Piotr Nyczyk, Hubert Niewiadomski, and Torsten Hoefler. 2024. Hot: Higher-order dynamic graph representation learning with efficient transformers. In Learning on Graphs Conference. PMLR, 15-1."},{"key":"e_1_3_2_1_7_1","volume-title":"Holographic Node Representations: Pre-training Task-Agnostic Node Embeddings. In The Thirteenth International Conference on Learning Representations.","author":"Bevilacqua Beatrice","unstructured":"Beatrice Bevilacqua, Joshua Robinson, Jure Leskovec, and Bruno Ribeiro. [n.d.]. Holographic Node Representations: Pre-training Task-Agnostic Node Embeddings. In The Thirteenth International Conference on Learning Representations."},{"key":"e_1_3_2_1_8_1","unstructured":"Xiao Bi Deli Chen Guanting Chen Shanhuang Chen Damai Dai Chengqi Deng Honghui Ding Kai Dong Qiushi Du Zhe Fu et al. 2024. Deepseek llm: Scaling open-source language models with longtermism. arXiv preprint arXiv:2401.02954(2024)."},{"key":"e_1_3_2_1_9_1","unstructured":"Michael M Bronstein Joan Bruna Taco Cohen and Petar Veli\u010dkovi\u0107. 2021. Geometric deep learning: Grids groups graphs geodesics and gauges. arXiv preprint arXiv:2104.13478(2021)."},{"key":"e_1_3_2_1_10_1","unstructured":"Semih Cant\u00fcrk Renming Liu Olivier Lapointe-Gagn\u00e9 Vincent L\u00e9tourneau Guy Wolf Dominique Beaini and Ladislav Ramp\u00e1\u0161ek. 2023. Graph positional and structural encoder. arXiv preprint arXiv:2307.07107(2023)."},{"key":"e_1_3_2_1_11_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Chen Jinsong","year":"2022","unstructured":"Jinsong Chen, Kaiyuan Gao, Gaichao Li, and Kun He. 2022. NAGphormer: A tokenized graph transformer for node classification in large graphs. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_12_1","volume-title":"Trompt: Towards a Better Deep Neural Network for Tabular Data.","author":"Chen Kuan-Yu","year":"2023","unstructured":"Kuan-Yu Chen, Ping-Han Chiang, Hsin-Rung Chou, Ting-Wei Chen, and Tien-Hao Chang. 2023. Trompt: Towards a Better Deep Neural Network for Tabular Data."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939785"},{"key":"e_1_3_2_1_14_1","unstructured":"Tianlang Chen Charilaos Kanatsoulis and Jure Leskovec. 2025. RelGNN: Composite Message Passing for Relational Deep Learning. arXiv preprint arXiv:2502.06784(2025)."},{"key":"e_1_3_2_1_15_1","volume-title":"Can graph neural networks count substructures? Advances in neural information processing systems","author":"Chen Zhengdao","year":"2020","unstructured":"Zhengdao Chen, Lei Chen, Soledad Villar, and Joan Bruna. 2020. Can graph neural networks count substructures? Advances in neural information processing systems, Vol. 33 (2020), 10383-10395."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1179"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i6.25880"},{"key":"e_1_3_2_1_18_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Cong Weilin","unstructured":"Weilin Cong, Si Zhang, Jian Kang, Baichuan Yuan, Hao Wu, Xin Zhou, Hanghang Tong, and Mehrdad Mahdavi. [n.d.]. Do We Really Need Complicated Model Architectures For Temporal Networks?. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3481916"},{"key":"e_1_3_2_1_20_1","volume-title":"AAAI Workshop on Deep Learning on Graphs: Methods and Applications(2021)","author":"Dwivedi Vijay Prakash","year":"2021","unstructured":"Vijay Prakash Dwivedi and Xavier Bresson. 2021. A Generalization of Transformer Networks to Graphs. AAAI Workshop on Deep Learning on Graphs: Methods and Applications(2021)."},{"key":"e_1_3_2_1_21_1","unstructured":"Vijay Prakash Dwivedi Sri Jaladi Yangyi Shen Federico L\u00f3pez Charilaos I Kanatsoulis Rishi Puri Matthias Fey and Jure Leskovec. 2025. Relational Graph Transformer. arXiv preprint arXiv:2505.10960(2025)."},{"key":"e_1_3_2_1_22_1","unstructured":"Vijay Prakash Dwivedi Chaitanya K Joshi Thomas Laurent Yoshua Bengio and Xavier Bresson. 2020. Benchmarking graph neural networks. arXiv:2003.00982(2020)."},{"key":"e_1_3_2_1_23_1","volume-title":"Xavier Bresson, Neil Shah, and Tong Zhao.","author":"Dwivedi Vijay Prakash","year":"2023","unstructured":"Vijay Prakash Dwivedi, Yozen Liu, Anh Tuan Luu, Xavier Bresson, Neil Shah, and Tong Zhao. 2023. Graph transformers for large graphs. arXiv preprint arXiv:2312.11109(2023)."},{"key":"e_1_3_2_1_24_1","volume-title":"Graph Neural Networks with Learnable Structural and Positional Representations. In International Conference on Learning Representations.","author":"Dwivedi Vijay Prakash","year":"2022","unstructured":"Vijay Prakash Dwivedi, Anh Tuan Luu, Thomas Laurent, Yoshua Bengio, and Xavier Bresson. 2022. Graph Neural Networks with Learnable Structural and Positional Representations. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_25_1","unstructured":"Francesco Ferrini Antonio Longa Andrea Passerini and Manfred Jaeger. 2024. A Self-Explainable Heterogeneous GNN for Relational Deep Learning. arXiv preprint arXiv:2412.00521(2024)."},{"key":"e_1_3_2_1_26_1","volume-title":"Forty-first International Conference on Machine Learning.","author":"Fey Matthias","year":"2024","unstructured":"Matthias Fey, Weihua Hu, Kexin Huang, Jan Eric Lenssen, Rishabh Ranjan, Joshua Robinson, Rex Ying, Jiaxuan You, and Jure Leskovec. 2024. Position: Relational deep learning-graph representation learning on relational databases. In Forty-first International Conference on Machine Learning."},{"key":"e_1_3_2_1_27_1","unstructured":"Matthias Fey Vid Kocijan Federico Lopez Jan Eric Lenssen and Jure Leskovec. [n.d.]. KumoRFM: A Foundation Model for In-Context Learning on Relational Data. ( [n. d.])."},{"key":"e_1_3_2_1_28_1","unstructured":"Mikhail Galkin Xinyu Yuan Hesham Mostafa Jian Tang and Zhaocheng Zhu. 2023. Towards foundation models for knowledge graph reasoning. arXiv preprint arXiv:2310.04562(2023)."},{"key":"e_1_3_2_1_29_1","volume-title":"International Workshop on Social Network Mining and Analysis. Springer, 1-19","author":"Gallagher Brian","year":"2008","unstructured":"Brian Gallagher and Tina Eliassi-Rad. 2008. Leveraging label-independent features for classification in sparsely labeled networks: An empirical study. In International Workshop on Social Network Mining and Analysis. Springer, 1-19."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2020.3026980"},{"key":"e_1_3_2_1_31_1","volume-title":"International conference on machine learning. PMLR, 1263-1272","author":"Gilmer Justin","year":"2017","unstructured":"Justin Gilmer, Samuel S Schoenholz, Patrick F Riley, Oriol Vinyals, and George E Dahl. 2017. Neural message passing for quantum chemistry. In International conference on machine learning. PMLR, 1263-1272."},{"key":"e_1_3_2_1_32_1","first-page":"24991","article-title":"On embeddings for numerical features in tabular deep learning","volume":"35","author":"Gorishniy Yury","year":"2022","unstructured":"Yury Gorishniy, Ivan Rubachev, and Artem Babenko. 2022. On embeddings for numerical features in tabular deep learning. Advances in Neural Information Processing Systems, Vol. 35 (2022), 24991-25004.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_33_1","volume-title":"Revisiting deep learning models for tabular data","author":"Gorishniy Yury","year":"1893","unstructured":"Yury Gorishniy, Ivan Rubachev, Valentin Khrulkov, and Artem Babenko. 2021. Revisiting deep learning models for tabular data, Vol. 34. 18932-18943."},{"key":"e_1_3_2_1_34_1","unstructured":"Aaron Grattafiori Abhimanyu Dubey Abhinav Jauhri Abhinav Pandey Abhishek Kadian Ahmad Al-Dahle Aiesha Letman Akhil Mathur Alan Schelten Alex Vaughan et al. 2024. The llama 3 herd of models. arXiv preprint arXiv:2407.21783(2024)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939754"},{"key":"e_1_3_2_1_36_1","volume-title":"Inductive representation learning on large graphs. Advances in neural information processing systems","author":"Hamilton Will","year":"2017","unstructured":"Will Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. Advances in neural information processing systems, Vol. 30 (2017)."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/2339530.2339723"},{"key":"e_1_3_2_1_38_1","volume-title":"Long short-term memory. Neural computation","author":"Hochreiter Sepp","year":"1997","unstructured":"Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural computation, Vol. 9, 8 (1997), 1735-1780."},{"key":"e_1_3_2_1_39_1","volume-title":"The Eleventh International Conference on Learning Representations.","author":"Hollmann Noah","unstructured":"Noah Hollmann, Samuel M\u00fcller, Katharina Eggensperger, and Frank Hutter. [n.d.]. TabPFN: A Transformer That Solves Small Tabular Classification Problems in a Second. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_40_1","volume-title":"Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems","author":"Hu Weihua","year":"2020","unstructured":"Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. 2020b. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, Vol. 33 (2020), 22118-22133."},{"key":"e_1_3_2_1_41_1","unstructured":"Weihua Hu Yiwen Yuan Zecheng Zhang Akihiro Nitta Kaidi Cao Vid Kocijan Jinu Sunil Jure Leskovec and Matthias Fey. 2024. Pytorch frame: A modular framework for multi-modal tabular learning. arXiv preprint arXiv:2404.00776(2024)."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380027"},{"key":"e_1_3_2_1_43_1","first-page":"2056","article-title":"Temporal graph benchmark for machine learning on temporal graphs","volume":"36","author":"Huang Shenyang","year":"2023","unstructured":"Shenyang Huang, Farimah Poursafaei, Jacob Danovitch, Matthias Fey, Weihua Hu, Emanuele Rossi, Jure Leskovec, Michael Bronstein, Guillaume Rabusseau, and Reihaneh Rabbany. 2023. Temporal graph benchmark for machine learning on temporal graphs. Advances in Neural Information Processing Systems, Vol. 36 (2023), 2056-2073.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_44_1","volume-title":"The Third Learning on Graphs Conference.","author":"Huang Shenyang","unstructured":"Shenyang Huang, Farimah Poursafaei, Reihaneh Rabbany, Guillaume Rabusseau, and Emanuele Rossi. [n.d.]. UTG: Towards a Unified View of Snapshot and Event Based Models for Temporal Graphs. In The Third Learning on Graphs Conference."},{"key":"e_1_3_2_1_45_1","volume-title":"Tabtransformer: Tabular data modeling using contextual embeddings. arXiv preprint arXiv:2012.06678(2020).","author":"Huang Xin","year":"2020","unstructured":"Xin Huang, Ashish Khetan, Milan Cvitkovic, and Zohar Karnin. 2020. Tabtransformer: Tabular data modeling using contextual embeddings. arXiv preprint arXiv:2012.06678(2020)."},{"key":"e_1_3_2_1_46_1","volume-title":"On the Stability of Expressive Positional Encodings for Graphs. In The Twelfth International Conference on Learning Representations.","author":"Huang Yinan","year":"2024","unstructured":"Yinan Huang, William Lu, Joshua Robinson, Yu Yang, Muhan Zhang, Stefanie Jegelka, and Pan Li. 2024. On the Stability of Expressive Positional Encodings for Graphs. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_1_47_1","unstructured":"Ankit Jain and Piero Molino. 2019. Enhancing Recommendations on Uber Eats with Graph Convolutional Networks. (2019)."},{"key":"e_1_3_2_1_48_1","first-page":"17","article-title":"Transformers are graph neural networks","volume":"12","author":"Joshi Chaitanya","year":"2020","unstructured":"Chaitanya Joshi. 2020. Transformers are graph neural networks. The Gradient, Vol. 12 (2020), 17.","journal-title":"The Gradient"},{"key":"e_1_3_2_1_49_1","unstructured":"Charilaos Kanatsoulis and Alejandro Ribeiro. 2024a. Counting graph substructures with graph neural networks. In The twelfth international conference on learning representations."},{"key":"e_1_3_2_1_50_1","volume-title":"Learning Efficient Positional Encodings with Graph Neural Networks. In The Thirteenth International Conference on Learning Representations.","author":"Kanatsoulis Charilaos I","unstructured":"Charilaos I Kanatsoulis, Evelyn Choi, Stephanie Jegelka, Jure Leskovec, and Alejandro Ribeiro. [n.d.]. Learning Efficient Positional Encodings with Graph Neural Networks. In The Thirteenth International Conference on Learning Representations."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP48485.2024.10447704"},{"key":"e_1_3_2_1_52_1","unstructured":"Seyed Mehran Kazemi Rishab Goel Sepehr Eghbali Janahan Ramanan Jaspreet Sahota Sanjay Thakur Stella Wu Cathal Smyth Pascal Poupart and Marcus Brubaker. 2019. Time2vec: Learning a vector representation of time. arXiv preprint arXiv:1907.05321(2019)."},{"key":"e_1_3_2_1_53_1","volume-title":"Lightgbm: A highly efficient gradient boosting decision tree. Advances in neural information processing systems","author":"Ke Guolin","year":"2017","unstructured":"Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. 2017. Lightgbm: A highly efficient gradient boosting decision tree. Advances in neural information processing systems, Vol. 30 (2017)."},{"key":"e_1_3_2_1_54_1","unstructured":"Thomas N Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907(2016)."},{"key":"e_1_3_2_1_55_1","volume-title":"GOAT: A Global Transformer on Large-scale Graphs. In International Conference on Machine Learning.","author":"Kong Kezhi","year":"2023","unstructured":"Kezhi Kong, Jiuhai Chen, John Kirchenbauer, Renkun Ni, C Bayan Bruss, and Tom Goldstein. 2023. GOAT: A Global Transformer on Large-scale Graphs. In International Conference on Machine Learning."},{"key":"e_1_3_2_1_56_1","first-page":"21618","article-title":"Rethinking graph transformers with spectral attention","volume":"34","author":"Kreuzer Devin","year":"2021","unstructured":"Devin Kreuzer, Dominique Beaini, Will Hamilton, Vincent L\u00e9tourneau, and Prudencio Tossou. 2021. Rethinking graph transformers with spectral attention. Advances in Neural Information Processing Systems, Vol. 34 (2021), 21618-21629.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_57_1","unstructured":"Veronica Lachi Antonio Longa Beatrice Bevilacqua Bruno Lepri Andrea Passerini and Bruno Ribeiro. 2024. Over 100x Speedup in Relational Deep Learning via Static GNNs and Tabular Distillation. (2024)."},{"key":"e_1_3_2_1_58_1","volume-title":"Deep learning. nature","author":"LeCun Yann","year":"2015","unstructured":"Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. 2015. Deep learning. nature, Vol. 521, 7553 (2015), 436-444."},{"key":"e_1_3_2_1_59_1","first-page":"1","article-title":"Transferability of spectral graph convolutional neural networks","volume":"22","author":"Levie Ron","year":"2021","unstructured":"Ron Levie, Wei Huang, Lorenzo Bucci, Michael Bronstein, and Gitta Kutyniok. 2021. Transferability of spectral graph convolutional neural networks. Journal of Machine Learning Research, Vol. 22, 272 (2021), 1-59.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3615059"},{"key":"e_1_3_2_1_61_1","volume-title":"Sign and Basis Invariant Networks for Spectral Graph Representation Learning. In The Eleventh International Conference on Learning Representations.","author":"Lim Derek","year":"2022","unstructured":"Derek Lim, Joshua David Robinson, Lingxiao Zhao, Tess Smidt, Suvrit Sra, Haggai Maron, and Stefanie Jegelka. 2022. Sign and Basis Invariant Networks for Spectral Graph Representation Learning. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_62_1","volume-title":"Introduction to abstract harmonic analysis","author":"Loomis Lynn H","unstructured":"Lynn H Loomis. 2013. Introduction to abstract harmonic analysis. Courier Corporation."},{"key":"e_1_3_2_1_63_1","volume-title":"Learning on Graphs Conference. PMLR, 1-1.","author":"Luo Yuhong","year":"2022","unstructured":"Yuhong Luo and Pan Li. 2022. Neighborhood-aware scalable temporal network representation learning. In Learning on Graphs Conference. PMLR, 1-1."},{"key":"e_1_3_2_1_64_1","volume-title":"Forty-first International Conference on Machine Learning.","author":"Mao Haitao","year":"2024","unstructured":"Haitao Mao, Zhikai Chen, Wenzhuo Tang, Jianan Zhao, Yao Ma, Tong Zhao, Neil Shah, Mikhail Galkin, and Jiliang Tang. 2024. Position: Graph foundation models are already here. In Forty-first International Conference on Machine Learning."},{"key":"e_1_3_2_1_65_1","volume-title":"Invariant and Equivariant Graph Networks. In International Conference on Learning Representations.","author":"Maron Haggai","year":"2018","unstructured":"Haggai Maron, Heli Ben-Hamu, Nadav Shamir, and Yaron Lipman. 2018. Invariant and Equivariant Graph Networks. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_66_1","volume-title":"Graphit: Encoding graph structure in transformers. arXiv preprint arXiv:2106.05667(2021).","author":"Mialon Gr\u00e9goire","year":"2021","unstructured":"Gr\u00e9goire Mialon, Dexiong Chen, Margot Selosse, and Julien Mairal. 2021. Graphit: Encoding graph structure in transformers. arXiv preprint arXiv:2106.05667(2021)."},{"key":"e_1_3_2_1_67_1","volume-title":"Gaurav Rattan, and Martin Grohe.","author":"Morris Christopher","year":"2019","unstructured":"Christopher Morris, Martin Ritzert, Matthias Fey, William L Hamilton, Jan Eric Lenssen, Gaurav Rattan, and Martin Grohe. 2019. Weisfeiler and leman go neural: higher-order graph neural networks. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence and Thirty-First Innovative Applications of Artificial Intelligence Conference and Ninth AAAI Symposium on Educational Advances in Artificial Intelligence. 4602-4609."},{"key":"e_1_3_2_1_68_1","unstructured":"Jan Motl and Oliver Schulte. 2015. The CTU prague relational learning repository. arXiv preprint arXiv:1511.03086(2015)."},{"key":"e_1_3_2_1_69_1","volume-title":"Transformers Meet Relational Databases. arXiv preprint arXiv:2412","author":"Pele\u0161ka Jakub","year":"2024","unstructured":"Jakub Pele\u0161ka and Gustav \u0160\u00edr. 2024. Transformers Meet Relational Databases. arXiv preprint arXiv:2412.05218(2024)."},{"key":"e_1_3_2_1_70_1","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans Ilya Sutskever et al. 2018. Improving language understanding by generative pre-training.(2018)."},{"key":"e_1_3_2_1_71_1","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of machine learning research, Vol. 21, 140 (2020), 1-67.","journal-title":"Journal of machine learning research"},{"key":"e_1_3_2_1_72_1","first-page":"14501","article-title":"Recipe for a general, powerful, scalable graph transformer","volume":"35","author":"Ramp\u00e1\u0161ek Ladislav","year":"2022","unstructured":"Ladislav Ramp\u00e1\u0161ek, Michael Galkin, Vijay Prakash Dwivedi, Anh Tuan Luu, Guy Wolf, and Dominique Beaini. 2022. Recipe for a general, powerful, scalable graph transformer. Advances in Neural Information Processing Systems, Vol. 35 (2022), 14501-14515.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_73_1","first-page":"21330","article-title":"Relbench: A benchmark for deep learning on relational databases","volume":"37","author":"Robinson Joshua","year":"2024","unstructured":"Joshua Robinson, Rishabh Ranjan, Weihua Hu, Kexin Huang, Jiaqi Han, Alejandro Dobles, Matthias Fey, Jan Eric Lenssen, Yiwen Yuan, Zecheng Zhang, et al., 2024. Relbench: A benchmark for deep learning on relational databases. Advances in Neural Information Processing Systems, Vol. 37 (2024), 21330-21341.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_74_1","unstructured":"Emanuele Rossi Ben Chamberlain Fabrizio Frasca Davide Eynard Federico Monti and Michael Bronstein. 2020. Temporal graph networks for deep learning on dynamic graphs. arXiv preprint arXiv:2006.10637(2020)."},{"key":"e_1_3_2_1_75_1","first-page":"1702","article-title":"Graphon Neural Networks and the Transferability of Graph Neural Networks","volume":"33","author":"Ruiz Luana","year":"2020","unstructured":"Luana Ruiz, Luiz Chamon, and Alejandro Ribeiro. 2020. Graphon Neural Networks and the Transferability of Graph Neural Networks. In Advances in Neural Information Processing Systems, Vol. 33. 1702-1712.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"crossref","unstructured":"David E Rumelhart Geoffrey E Hinton Ronald J Williams et al. 1985. Learning internal representations by error propagation.","DOI":"10.21236\/ADA164453"},{"key":"e_1_3_2_1_77_1","first-page":"593","volume-title":"ESWC 2018, Heraklion, Crete, Greece, June 3-7, 2018, proceedings 15","author":"Schlichtkrull Michael","year":"2018","unstructured":"Michael Schlichtkrull, Thomas N Kipf, Peter Bloem, Rianne Van Den Berg, Ivan Titov, and Max Welling. 2018. Modeling relational data with graph convolutional networks. In The semantic web: 15th international conference, ESWC 2018, Heraklion, Crete, Greece, June 3-7, 2018, proceedings 15. Springer, 593-607."},{"key":"e_1_3_2_1_78_1","volume-title":"The Twelfth International Conference on Learning Representations.","author":"Shamsi Kiarash","year":"2024","unstructured":"Kiarash Shamsi, Farimah Poursafaei, Shenyang Huang, Bao Tran Gia Ngo, Baris Coskunuzer, and Cuneyt Gurcan Akcora. 2024. GraphPulse: Topological representations for temporal graph property prediction. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_1_79_1","volume-title":"Exphormer: Sparse transformers for graphs. arXiv preprint arXiv:2303.06147(2023).","author":"Shirzad Hamed","year":"2023","unstructured":"Hamed Shirzad, Ameya Velingker, Balaji Venkatachalam, Danica J Sutherland, and Ali Kemal Sinop. 2023. Exphormer: Sparse transformers for graphs. arXiv preprint arXiv:2303.06147(2023)."},{"key":"e_1_3_2_1_80_1","first-page":"19870","article-title":"On the scalability of gnns for molecular graphs","volume":"37","author":"Sypetkowski Maciej","year":"2025","unstructured":"Maciej Sypetkowski, Frederik Wenkel, Farimah Poursafaei, Nia Dickson, Karush Suri, Philip Fradkin, and Dominique Beaini. 2025. On the scalability of gnns for molecular graphs. Advances in Neural Information Processing Systems, Vol. 37 (2025), 19870-19906.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_81_1","volume-title":"Siamak Shakeri, Dara Bahri","author":"Tay Yi","year":"2022","unstructured":"Yi Tay, Mostafa Dehghani, Vinh Q Tran, Xavier Garcia, Jason Wei, Xuezhi Wang, Hyung Won Chung, Siamak Shakeri, Dara Bahri, Tal Schuster, et al., 2022. Ul2: Unifying language learning paradigms. arXiv preprint arXiv:2205.05131(2022)."},{"key":"e_1_3_2_1_82_1","unstructured":"Gemini Team Rohan Anil Sebastian Borgeaud Jean-Baptiste Alayrac Jiahui Yu Radu Soricut Johan Schalkwyk Andrew M Dai Anja Hauth Katie Millican et al. 2023. Gemini: a family of highly capable multimodal models. arXiv preprint arXiv:2312.11805(2023)."},{"key":"e_1_3_2_1_83_1","volume-title":"Juliette Love, et al.","author":"Team Gemma","year":"2024","unstructured":"Gemma Team, Thomas Mesnard, Cassidy Hardin, Robert Dadashi, Surya Bhupatiraju, Shreya Pathak, Laurent Sifre, Morgane Rivi\u00e8re, Mihir Sanjay Kale, Juliette Love, et al., 2024a. Gemma: Open models based on gemini research and technology. arXiv preprint arXiv:2403.08295(2024)."},{"key":"e_1_3_2_1_84_1","unstructured":"Reka Team Aitor Ormazabal Che Zheng Cyprien de Masson d'Autume Dani Yogatama Deyu Fu Donovan Ong Eric Chen Eugenie Lamprecht Hai Pham et al. 2024b. Reka core flash and edge: A series of powerful multimodal language models. arXiv preprint arXiv:2404.12387(2024)."},{"key":"e_1_3_2_1_85_1","volume-title":"Advances in Neural Information Processing Systems","volume":"30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in Neural Information Processing Systems, Vol. 30 (2017)."},{"key":"e_1_3_2_1_86_1","unstructured":"Petar Velickovic Guillem Cucurull Arantxa Casanova Adriana Romero Pietro Lio Yoshua Bengio et al. 2017. Graph attention networks. stat Vol. 1050 20 (2017) 10-48550."},{"key":"e_1_3_2_1_87_1","volume-title":"Tcl: Transformer-based dynamic graph modelling via contrastive learning. arXiv preprint arXiv:2105.07944(2021).","author":"Wang Lu","year":"2021","unstructured":"Lu Wang, Xiaofu Chang, Shuang Li, Yunfei Chu, Hui Li, Wei Zhang, Xiaofeng He, Le Song, Jingren Zhou, and Hongxia Yang. 2021. Tcl: Transformer-based dynamic graph modelling via contrastive learning. arXiv preprint arXiv:2105.07944(2021)."},{"key":"e_1_3_2_1_88_1","doi-asserted-by":"crossref","unstructured":"Xiao Wang Houye Ji Chuan Shi Bai Wang Yanfang Ye Peng Cui and Philip S Yu. 2019. Heterogeneous graph attention network. In The world wide web conference. 2022-2032.","DOI":"10.1145\/3308558.3313562"},{"key":"e_1_3_2_1_89_1","volume-title":"International Conference on Learning Representations.","author":"Wang Yanbang","year":"2020","unstructured":"Yanbang Wang, Yen-Yu Chang, Yunyu Liu, Jure Leskovec, and Pan Li. 2020. Inductive Representation Learning in Temporal Networks via Causal Anonymous Walks. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_90_1","unstructured":"Marek Wydmuch \u0141ukasz Borchmann and Filip Grali'nski. 2024. Tackling prediction tasks in relational databases with LLMs. arXiv preprint arXiv:2411.11829(2024)."},{"key":"e_1_3_2_1_91_1","unstructured":"Da Xu Chuanwei Ruan Evren Korpeoglu Sushant Kumar and Kannan Achan. 2020. Inductive representation learning on temporal graphs. arXiv preprint arXiv:2002.07962(2020)."},{"key":"e_1_3_2_1_92_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=ryGs6iA5Km","author":"Xu Keyulu","year":"2019","unstructured":"Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2019. How Powerful are Graph Neural Networks?. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=ryGs6iA5Km"},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3340404"},{"key":"e_1_3_2_1_94_1","first-page":"28877","article-title":"Do transformers really perform badly for graph representation","volume":"34","author":"Ying Chengxuan","year":"2021","unstructured":"Chengxuan Ying, Tianle Cai, Shengjie Luo, Shuxin Zheng, Guolin Ke, Di He, Yanming Shen, and Tie-Yan Liu. 2021. Do transformers really perform badly for graph representation? Advances in Neural Information Processing Systems, Vol. 34 (2021), 28877-28888.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219890"},{"key":"e_1_3_2_1_96_1","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539300"},{"key":"e_1_3_2_1_97_1","first-page":"67686","article-title":"Towards better dynamic graph learning: New architecture and unified library","volume":"36","author":"Yu Le","year":"2023","unstructured":"Le Yu, Leilei Sun, Bowen Du, and Weifeng Lv. 2023. Towards better dynamic graph learning: New architecture and unified library. Advances in Neural Information Processing Systems, Vol. 36 (2023), 67686-67700.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_98_1","unstructured":"Yiwen Yuan Zecheng Zhang Xinwei He Akihiro Nitta Weihua Hu Dong Wang Manan Shah Shenyang Huang Bla\u017e Stojanovi\u010d Alan Krumholz et al. 2024. ContextGNN: Beyond Two-Tower Recommendation Systems. arXiv preprint arXiv:2411.19513(2024)."},{"key":"e_1_3_2_1_99_1","volume-title":"NeurIPS 2023 Second Table Representation Learning Workshop.","author":"Zahradn\u00edk Luk\u00e1\u0161","year":"2023","unstructured":"Luk\u00e1\u0161 Zahradn\u00edk, Jan Neumann, and Gustav \u0160\u00edr. 2023. A deep learning blueprint for relational databases. In NeurIPS 2023 Second Table Representation Learning Workshop."},{"key":"e_1_3_2_1_100_1","unstructured":"Bohang Zhang Shengjie Luo Liwei Wang and Di He. 2023. Rethinking the expressive power of gnns via graph biconnectivity. arXiv preprint arXiv:2301.09505(2023)."},{"key":"e_1_3_2_1_101_1","first-page":"21171","article-title":"Hierarchical graph transformer with adaptive node sampling","volume":"35","author":"Zhang Zaixi","year":"2022","unstructured":"Zaixi Zhang, Qi Liu, Qingyong Hu, and Chee-Kong Lee. 2022. Hierarchical graph transformer with adaptive node sampling. Advances in Neural Information Processing Systems, Vol. 35 (2022), 21171-21183.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_102_1","volume-title":"Gophormer: Ego-graph transformer for node classification. arXiv preprint arXiv:2110.13094(2021).","author":"Zhao Jianan","year":"2021","unstructured":"Jianan Zhao, Chaozhuo Li, Qianlong Wen, Yiqi Wang, Yuming Liu, Hao Sun, Xing Xie, and Yanfang Ye. 2021. Gophormer: Ego-graph transformer for node classification. arXiv preprint arXiv:2110.13094(2021)."},{"key":"e_1_3_2_1_103_1","doi-asserted-by":"crossref","unstructured":"Tong Zhao Yozen Liu Matthew Kolodner Kyle Montemayor Elham Ghazizadeh Ankit Batra Zihao Fan Xiaobin Gao Xuan Guo Jiwen Ren et al. 2025. GiGL: Large-Scale Graph Neural Networks at Snapchat. arXiv preprint arXiv:2502.15054(2025).","DOI":"10.1145\/3711896.3737229"},{"key":"e_1_3_2_1_104_1","volume-title":"XTab: Cross-table Pretraining for Tabular Transformers. In ICML'23: Proceedings of the 40th International Conference on Machine Learning. JMLR. org.","author":"Zhu Bingzhao","year":"2023","unstructured":"Bingzhao Zhu, Xingjian Shi, Nick Erickson, Mu Li, George Karypis, and Mahsa Shoaran. 2023a. XTab: Cross-table Pretraining for Tabular Transformers. In ICML'23: Proceedings of the 40th International Conference on Machine Learning. JMLR. org."},{"key":"e_1_3_2_1_105_1","unstructured":"Wenhao Zhu Tianyu Wen Guojie Song Xiaojun Ma and Liang Wang. 2023b. Hierarchical Transformer for Scalable Graph Learning. arXiv preprint arXiv:2305.02866(2023)."}],"event":{"name":"KDD '25: The 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Toronto ON Canada","acronym":"KDD '25","sponsor":["SIGKDD ACM Special Interest Group on Knowledge Discovery in Data","SIGMOD ACM Special Interest Group on Management of Data"]},"container-title":["Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V.2"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3711896.3736558","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,30]],"date-time":"2026-04-30T17:55:54Z","timestamp":1777571754000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3711896.3736558"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,3]]},"references-count":105,"alternative-id":["10.1145\/3711896.3736558","10.1145\/3711896"],"URL":"https:\/\/doi.org\/10.1145\/3711896.3736558","relation":{},"subject":[],"published":{"date-parts":[[2025,8,3]]},"assertion":[{"value":"2025-08-03","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}