{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,18]],"date-time":"2026-01-18T13:04:48Z","timestamp":1768741488208,"version":"3.49.0"},"reference-count":81,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"12","license":[{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62222607"],"award-info":[{"award-number":["62222607"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shanghai Committee of Science and Technology","award":["22511105100"],"award-info":[{"award-number":["22511105100"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1109\/tpami.2024.3443110","type":"journal-article","created":{"date-parts":[[2024,8,14]],"date-time":"2024-08-14T17:35:31Z","timestamp":1723656931000},"page":"10845-10862","source":"Crossref","is-referenced-by-count":6,"title":["EasyDGL: Encode, Train and Interpret for Continuous-Time Dynamic Graph Learning"],"prefix":"10.1109","volume":"46","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3911-8711","authenticated-orcid":false,"given":"Chao","family":"Chen","sequence":"first","affiliation":[{"name":"Department of Computer Science and Engineering and MoE Key Lab of Artificial Intelligence, Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7808-3959","authenticated-orcid":false,"given":"Haoyu","family":"Geng","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering and MoE Key Lab of Artificial Intelligence, Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6099-9261","authenticated-orcid":false,"given":"Nianzu","family":"Yang","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering and MoE Key Lab of Artificial Intelligence, Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4029-3322","authenticated-orcid":false,"given":"Xiaokang","family":"Yang","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering and MoE Key Lab of Artificial Intelligence, Shanghai Jiao Tong University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9639-7679","authenticated-orcid":false,"given":"Junchi","family":"Yan","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering and MoE Key Lab of Artificial Intelligence, Shanghai Jiao Tong University, Shanghai, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"crossref","DOI":"10.1007\/978-0-387-68560-1","volume-title":"Survival and Event History Analysis: A Process Point of View","author":"Aalen","year":"2008"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i6.25831"},{"key":"ref3","article-title":"The Netflix prize","volume-title":"Proc. ACM SIGKDD Int. Conf. Knowl. Discov. Data Mining Cup","author":"Bennett"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3054830"},{"key":"ref5","article-title":"How attentive are graph attention networks?","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Brody"},{"key":"ref6","article-title":"Graph signal sampling for inductive one-bit matrix completion: A closed-form solution","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Chen"},{"key":"ref7","first-page":"1606","article-title":"Learning self-modulating attention in continuous time space with applications to sequential recommendation","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Chen"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2021.3050407"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3162711"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3449947"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4757-2001-3"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i4.25540"},{"key":"ref13","first-page":"3844","article-title":"Convolutional neural networks on graphs with fast localized spectral filtering","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Defferrard"},{"key":"ref14","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2007.1115"},{"key":"ref16","first-page":"703","article-title":"Analysis of learning from positive and unlabeled data","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Du Plessis"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM51629.2021.00123"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3482242"},{"key":"ref19","first-page":"6235","article-title":"Bayesian continuous-time tucker decomposition","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Fang"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2004.1262185"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1516"},{"key":"ref22","article-title":"Predict then propagate: Graph neural networks meet personalized pagerank","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Gasteiger"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5815"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1137\/090771806"},{"key":"ref25","first-page":"1025","article-title":"Inductive representation learning on large graphs","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Hamilton"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"ref27","article-title":"Session-based recommendations with recurrent neural networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Hidasi"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539321"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1145\/3511700"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3614868"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i7.25980"},{"key":"ref32","first-page":"19874","article-title":"Neural temporal walks: Motif-aware representation learning on continuous-time dynamic graphs","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Jin"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2018.00035"},{"issue":"70","key":"ref34","first-page":"1","article-title":"Representation learning for dynamic graphs: A survey","volume":"21","author":"Kazemi","year":"2020","journal-title":"J. Mach. Learn. Res."},{"key":"ref35","article-title":"Semi-supervised classification with graph convolutional networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Kipf"},{"key":"ref36","first-page":"304","article-title":"Sampling techniques for the nystrom method","volume-title":"Proc. Artif. Intell. Statist.","author":"Kumar"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330895"},{"key":"ref38","first-page":"11906","article-title":"DSTAGNN: Dynamic spatial-temporal aware graph neural network for traffic flow forecasting","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Lan"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1145\/3336191.3371786"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i7.26034"},{"key":"ref41","first-page":"4465","article-title":"Distance encoding: Design provably more powerful neural networks for graph representation learning","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Li"},{"key":"ref42","article-title":"Diffusion convolutional recurrent neural network: Data-driven traffic forecasting","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Li"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1145\/3178876.3186150"},{"key":"ref44","first-page":"6666","article-title":"Generative causal explanations for graph neural networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Lin"},{"key":"ref45","first-page":"9228","article-title":"Kalman filtering attention for user behavior modeling in CTR prediction","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Liu"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2022.3172903"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2018.2875886"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467379"},{"key":"ref49","first-page":"6757","article-title":"The neural Hawkes process: A neurally self-modulating multivariate point process","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Mei"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1080\/01621459.1949.10483310"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1145\/3184558.3191526"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2018.2820126"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.5984"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01103"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380073"},{"key":"ref56","article-title":"Temporal graph networks for deep learning on dynamic graphs","author":"Rossi","year":"2020"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1145\/3336191.3371845"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1145\/3511808.3557702"},{"key":"ref59","first-page":"5474","article-title":"Markov random fields for collaborative filtering","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Steck"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4757-2272-7"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357895"},{"key":"ref62","article-title":"Degree-quant: Quantization-aware training for graph neural networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Tailor"},{"key":"ref63","article-title":"Large-scale representation learning on graphs via bootstrapping","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Thakoor"},{"key":"ref64","article-title":"DyRep: Learning representations over dynamic graphs","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Trivedi"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref66","article-title":"Graph attention networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Veli\u010dkovi\u0107"},{"key":"ref67","first-page":"11329","article-title":"Towards open-world recommendation: An inductive model-based collaborative filtering approach","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Wu"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1145\/3616855.3635837"},{"key":"ref69","article-title":"Inductive representation learning on temporal graphs","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Xu"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1145\/3534678.3539300"},{"key":"ref71","first-page":"10871","article-title":"When does self-supervision help graph convolutional networks?","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"You"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2018\/505"},{"key":"ref73","doi-asserted-by":"publisher","DOI":"10.1145\/3366423.3380116"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403085"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3204236"},{"key":"ref76","first-page":"12241","article-title":"On explainability of graph neural networks via subgraph explorations","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Yuan"},{"key":"ref77","article-title":"GaaN: Gated attention networks for learning on large and spatiotemporal graphs","volume-title":"Proc. Conf. Uncertainty Artif. Intell.","author":"Zhang"},{"key":"ref78","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11257"},{"key":"ref79","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2016.2591009"},{"key":"ref80","first-page":"11692","article-title":"Transformer Hawkes process","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Zuo"},{"key":"ref81","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220054"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/34\/10746266\/10636764.pdf?arnumber=10636764","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,27]],"date-time":"2024-11-27T00:26:16Z","timestamp":1732667176000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10636764\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12]]},"references-count":81,"journal-issue":{"issue":"12"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2024.3443110","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12]]}}}