{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,21]],"date-time":"2026-02-21T01:30:04Z","timestamp":1771637404640,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":58,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T00:00:00Z","timestamp":1730073600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Fundamental Research Funds for the Central Universities of China","award":["226-2024-00140"],"award-info":[{"award-number":["226-2024-00140"]}]},{"name":"Zhejiang Natural Science Foundation","award":["LD24F020013"],"award-info":[{"award-number":["LD24F020013"]}]},{"name":"Ministry of Science and Technology of the People's Republic of China","award":["2022YFF0902600"],"award-info":[{"award-number":["2022YFF0902600"]}]},{"name":"Zhejiang University Education Foundation's Qizhen Talent program"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,28]]},"DOI":"10.1145\/3664647.3681018","type":"proceedings-article","created":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T06:59:33Z","timestamp":1729925973000},"page":"8556-8565","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["UniGM: Unifying Multiple Pre-trained Graph Models via Adaptive Knowledge Aggregation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9099-3792","authenticated-orcid":false,"given":"Jintao","family":"Chen","sequence":"first","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0953-6923","authenticated-orcid":false,"given":"Fan","family":"Wang","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-5510-2867","authenticated-orcid":false,"given":"Shengye","family":"Pang","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0634-8089","authenticated-orcid":false,"given":"Siwei","family":"Tan","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9663-7441","authenticated-orcid":false,"given":"Mingshuai","family":"Chen","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4166-6189","authenticated-orcid":false,"given":"Tiancheng","family":"Zhao","sequence":"additional","affiliation":[{"name":"Binjiang Institute of Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6335-8312","authenticated-orcid":false,"given":"Meng","family":"Xi","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4703-7348","authenticated-orcid":false,"given":"Jianwei","family":"Yin","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,28]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2015.7280769"},{"key":"e_1_3_2_1_2_1","volume-title":"Practical bilevel optimization: algorithms and applications","author":"Bard Jonathan F","unstructured":"Jonathan F Bard. 2013. Practical bilevel optimization: algorithms and applications. Vol. 30. Springer Science & Business Media."},{"key":"e_1_3_2_1_3_1","unstructured":"Luisa Bentivogli Peter Clark Ido Dagan and Danilo Giampiccolo. 2009. The Fifth PASCAL Recognizing Textual Entailment Challenge.. In TAC."},{"key":"e_1_3_2_1_4_1","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language models are few-shot learners. Advances in neural information processing systems 33 (2020) 1877--1901."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"e_1_3_2_1_6_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_7_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. north american","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. north american chapter of the association for computational linguistics (2019)."},{"key":"e_1_3_2_1_8_1","volume-title":"Unified language model pretraining for natural language understanding and generation. Advances in Neural Information Processing Systems 32","author":"Dong Li","year":"2019","unstructured":"Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, and Hsiao-Wuen Hon. 2019. Unified language model pretraining for natural language understanding and generation. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11704-019-8208-z"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i4.20313"},{"key":"e_1_3_2_1_11_1","volume-title":"International Conference on Machine Learning. PMLR, 1165--1173","author":"Franceschi Luca","year":"2017","unstructured":"Luca Franceschi, Michele Donini, Paolo Frasconi, and Massimiliano Pontil. 2017. Forward and reverse gradient-based hyperparameter optimization. In International Conference on Machine Learning. PMLR, 1165--1173."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467450"},{"key":"e_1_3_2_1_13_1","volume-title":"Proceedings of International Conference on Machine Learning. 3451--3461","author":"Hassani Kaveh","year":"2020","unstructured":"Kaveh Hassani and Amir Hosein Khasahmadi. 2020. Contrastive Multi-View Representation Learning on Graphs. In Proceedings of International Conference on Machine Learning. 3451--3461."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.322"},{"key":"e_1_3_2_1_16_1","volume-title":"GraphMAE: Self-Supervised Masked Graph Autoencoders. arXiv e-prints","author":"Hou Zhenyu","year":"2022","unstructured":"Zhenyu Hou, Xiao Liu, Yukuo Cen, Yuxiao Dong, Hongxia Yang, Chunjie Wang, and Jie Tang. 2022. GraphMAE: Self-Supervised Masked Graph Autoencoders. arXiv e-prints (2022), arXiv-2205."},{"key":"e_1_3_2_1_17_1","volume-title":"Strategies for Pre-training Graph Neural Networks. In International Conference on Learning Representations. https: \/\/openreview.net\/forum?id=HJlWWJSFDH","author":"Weihua","year":"2020","unstructured":"Weihua Hu*, Bowen Liu*, Joseph Gomes, Marinka Zitnik, Percy Liang, Vijay Pande, and Jure Leskovec. 2020. Strategies for Pre-training Graph Neural Networks. In International Conference on Learning Representations. https: \/\/openreview.net\/forum?id=HJlWWJSFDH"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403237"},{"key":"e_1_3_2_1_19_1","volume-title":"Noise stability regularization for improving BERT fine-tuning. arXiv preprint arXiv:2107.04835","author":"Hua Hang","year":"2021","unstructured":"Hang Hua, Xingjian Li, Dejing Dou, Cheng-Zhong Xu, and Jiebo Luo. 2021. Noise stability regularization for improving BERT fine-tuning. arXiv preprint arXiv:2107.04835 (2021)."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01249-6_38"},{"key":"e_1_3_2_1_21_1","unstructured":"Wei Jin Xiaorui Liu Xiangyu Zhao Yao Ma Neil Shah and Jiliang Tang. 2022. Automated Self-Supervised Learning for Graphs. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_22_1","unstructured":"Dongki Kim Jinheon Baek and Sung Ju Hwang. 2022. Graph Self-supervised Learning with Accurate Discrepancy Learning. In Advances in Neural Information Processing Systems Alice H. Oh Alekh Agarwal Danielle Belgrave and Kyunghyun Cho (Eds.). https:\/\/openreview.net\/forum?id=JY6fLgR8Yq"},{"key":"e_1_3_2_1_23_1","volume-title":"Sparse deep belief net model for visual area V2. Advances in neural information processing systems 20","author":"Lee Honglak","year":"2007","unstructured":"Honglak Lee, Chaitanya Ekanadham, and Andrew Ng. 2007. Sparse deep belief net model for visual area V2. Advances in neural information processing systems 20 (2007)."},{"key":"e_1_3_2_1_24_1","volume-title":"An effective self-supervised framework for learning expressive molecular global representations to drug discovery. Briefings in Bioinformatics 22, 6","author":"Li Pengyong","year":"2021","unstructured":"Pengyong Li, Jun Wang, Yixuan Qiao, Hao Chen, Yihuan Yu, Xiaojun Yao, Peng Gao, Guotong Xie, and Sen Song. 2021. An effective self-supervised framework for learning expressive molecular global representations to drug discovery. Briefings in Bioinformatics 22, 6 (2021), bbab109."},{"key":"e_1_3_2_1_25_1","volume-title":"ICML (Proceedings of Machine Learning Research","volume":"13065","author":"Li Sihang","year":"2022","unstructured":"Sihang Li, Xiang Wang, An Zhang, Yingxin Wu, Xiangnan He, and Tat-Seng Chua. 2022. Let Invariant Rationale Discovery Inspire Graph Contrastive Learning. In ICML (Proceedings of Machine Learning Research, Vol. 162). PMLR, 13052--13065."},{"key":"e_1_3_2_1_26_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/ forum?id=xQUe1pOKPam","author":"Liu Shengchao","year":"2022","unstructured":"Shengchao Liu, Hanchen Wang, Weiyang Liu, Joan Lasenby, Hongyu Guo, and Jian Tang. 2022. Pre-training Molecular Graph Representation with 3D Geometry. In International Conference on Learning Representations. https:\/\/openreview.net\/ forum?id=xQUe1pOKPam"},{"key":"e_1_3_2_1_27_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_1_28_1","volume-title":"Learning to Pre-train Graph Neural Networks","author":"Lu Yuanfu","unstructured":"Yuanfu Lu, Xunqiang Jiang, Yuan Fang, and Chuan Shi. 2021. Learning to Pre-train Graph Neural Networks. In AAAI. AAAI Press, 4276--4284."},{"key":"e_1_3_2_1_29_1","volume-title":"Recurrent neural networks: design and applications","author":"Medsker Larry","unstructured":"Larry Medsker and Lakhmi C Jain. 1999. Recurrent neural networks: design and applications. CRC press."},{"key":"e_1_3_2_1_30_1","volume-title":"Model Fusion of Heterogeneous Neural Networks via Cross-Layer Alignment. arXiv preprint arXiv:2110.15538","author":"Nguyen Dang","year":"2021","unstructured":"Dang Nguyen, Khai Nguyen, Dinh Phung, Hung Bui, and Nhat Ho. 2021. Model Fusion of Heterogeneous Neural Networks via Cross-Layer Alignment. arXiv preprint arXiv:2110.15538 (2021)."},{"key":"e_1_3_2_1_31_1","volume-title":"Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems 32","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. 2019. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems 32 (2019)."},{"key":"e_1_3_2_1_32_1","volume-title":"GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training. arXiv preprint arXiv:2006.09963","author":"Qiu Jiezhong","year":"2020","unstructured":"Jiezhong Qiu, Qibin Chen, Yuxiao Dong, Jing Zhang, Hongxia Yang, Ming Ding, Kuansan Wang, and Jie Tang. 2020. GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training. arXiv preprint arXiv:2006.09963 (2020)."},{"key":"e_1_3_2_1_33_1","first-page":"12559","article-title":"Self-supervised graph transformer on large-scale molecular data","volume":"33","author":"Rong Yu","year":"2020","unstructured":"Yu Rong, Yatao Bian, Tingyang Xu, Weiyang Xie, Ying Wei, Wenbing Huang, and Junzhou Huang. 2020. Self-supervised graph transformer on large-scale molecular data. Advances in Neural Information Processing Systems 33 (2020), 12559--12571.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"crossref","unstructured":"Olga Russakovsky Jia Deng Hao Su Jonathan Krause Sanjeev Satheesh Sean Ma Zhiheng Huang Andrej Karpathy Aditya Khosla Michael Bernstein et al. 2015. Imagenet large scale visual recognition challenge. International journal of computer vision 115 3 (2015) 211--252.","DOI":"10.1007\/s11263-015-0816-y"},{"key":"e_1_3_2_1_35_1","volume-title":"Ijcai","volume":"99","author":"Schapire Robert E","year":"1999","unstructured":"Robert E Schapire. 1999. A brief introduction to boosting. In Ijcai, Vol. 99. Citeseer, 1401--1406."},{"key":"e_1_3_2_1_36_1","first-page":"22045","article-title":"Model fusion via optimal transport","volume":"33","author":"Singh Sidak Pal","year":"2020","unstructured":"Sidak Pal Singh and Martin Jaggi. 2020. Model fusion via optimal transport. Advances in Neural Information Processing Systems 33 (2020), 22045--22055.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_37_1","volume-title":"Proceedings of the 2013 conference on empirical methods in natural language processing. 1631--1642","author":"Socher Richard","year":"2013","unstructured":"Richard Socher, Alex Perelygin, Jean Wu, Jason Chuang, Christopher D Manning, Andrew Y Ng, and Christopher Potts. 2013. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 conference on empirical methods in natural language processing. 1631--1642."},{"key":"e_1_3_2_1_38_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=r1lfF2NYvH","author":"Sun Fan-Yun","year":"2020","unstructured":"Fan-Yun Sun, Jordan Hoffman, Vikas Verma, and Jian Tang. 2020. InfoGraph: Unsupervised and Semi-supervised Graph-Level Representation Learning via Mutual Information Maximization. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=r1lfF2NYvH"},{"key":"e_1_3_2_1_39_1","volume-title":"MoCL: Contrastive Learning on Molecular Graphs with Multi-level Domain Knowledge. KDD","author":"Sun Mengying","year":"2021","unstructured":"Mengying Sun, Jing Xing, Huijun Wang, Bin Chen, and Jiayu Zhou. 2021. MoCL: Contrastive Learning on Molecular Graphs with Multi-level Domain Knowledge. KDD (2021)."},{"key":"e_1_3_2_1_40_1","volume-title":"Adversarial graph augmentation to improve graph contrastive learning. Advances in Neural Information Processing Systems 34","author":"Suresh Susheel","year":"2021","unstructured":"Susheel Suresh, Pan Li, Cong Hao, and Jennifer Neville. 2021. Adversarial graph augmentation to improve graph contrastive learning. Advances in Neural Information Processing Systems 34 (2021)."},{"key":"e_1_3_2_1_41_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=0UXT6PpRpW","author":"Thakoor Shantanu","year":"2022","unstructured":"Shantanu Thakoor, Corentin Tallec, Mohammad Gheshlaghi Azar, Mehdi Azabou, Eva L Dyer, Remi Munos, Petar Veli\u010dkovi\u0107, and Michal Valko. 2022. Large-Scale Representation Learning on Graphs via Bootstrapping. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=0UXT6PpRpW"},{"key":"e_1_3_2_1_42_1","volume-title":"Proc. AAAI-96 Workshop on Integrating Multiple Learned Models. AAAI Press. Citeseer, 133--138","author":"Utans Joachim","year":"1996","unstructured":"Joachim Utans. 1996. Weight averaging for neural networks and local resampling schemes. In Proc. AAAI-96 Workshop on Integrating Multiple Learned Models. AAAI Press. Citeseer, 133--138."},{"key":"e_1_3_2_1_43_1","volume-title":"Federated Learning with Matched Averaging. In International Conference on Learning Representations. https:\/\/openreview.net\/forum? id=BkluqlSFDS","author":"Wang Hongyi","year":"2020","unstructured":"Hongyi Wang, Mikhail Yurochkin, Yuekai Sun, Dimitris Papailiopoulos, and Yasaman Khazaeni. 2020. Federated Learning with Matched Averaging. In International Conference on Learning Representations. https:\/\/openreview.net\/forum? id=BkluqlSFDS"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-022-00447-x"},{"key":"e_1_3_2_1_45_1","volume-title":"Stacked generalization. Neural networks 5, 2","author":"Wolpert David H","year":"1992","unstructured":"David H Wolpert. 1992. Stacked generalization. Neural networks 5, 2 (1992), 241--259."},{"key":"e_1_3_2_1_46_1","volume-title":"MoleculeNet: a benchmark for molecular machine learning. Chemical science 9, 2","author":"Wu Zhenqin","year":"2018","unstructured":"Zhenqin Wu, Bharath Ramsundar, Evan N Feinberg, Joseph Gomes, Caleb Geniesse, Aneesh S Pappu, Karl Leswing, and Vijay Pande. 2018. MoleculeNet: a benchmark for molecular machine learning. Chemical science 9, 2 (2018), 513--530."},{"key":"e_1_3_2_1_47_1","volume-title":"Proceedings of TheWeb Conference","author":"Xia Jun","year":"2022","unstructured":"Jun Xia, Lirong Wu, Jintao Chen, Bozhen Hu, and Stan Z. Li. 2022. SimGRACE: A Simple Framework for Graph Contrastive Learning without Data Augmentation. In Proceedings of TheWeb Conference 2022. Association for Computing Machinery."},{"key":"e_1_3_2_1_48_1","volume-title":"Towards effective and generalizable fine-tuning for pre-trained molecular graph models. bioRxiv","author":"Xia Jun","year":"2022","unstructured":"Jun Xia, Jiangbin Zheng, Cheng Tan, Ge Wang, and Stan Z Li. 2022. Towards effective and generalizable fine-tuning for pre-trained molecular graph models. bioRxiv (2022)."},{"key":"e_1_3_2_1_49_1","volume-title":"Pre-training Graph Neural Networks for Molecular Representations: Retrospect and Prospect. In ICML 2022 2nd AI for Science Workshop. https:\/\/openreview.net\/forum?id=dhXLkrY2Nj3","author":"Xia Jun","unstructured":"Jun Xia, Yanqiao Zhu, Yuanqi Du, and Stan Z. Li. 2022. Pre-training Graph Neural Networks for Molecular Representations: Retrospect and Prospect. In ICML 2022 2nd AI for Science Workshop. https:\/\/openreview.net\/forum?id=dhXLkrY2Nj3"},{"key":"e_1_3_2_1_50_1","unstructured":"Keyulu Xu Weihua Hu Jure Leskovec and Stefanie Jegelka. 2019. How Powerful are Graph Neural Networks?. In ICLR."},{"key":"e_1_3_2_1_51_1","volume-title":"International Conference on Machine Learning. PMLR, 11548--11558","author":"Xu Minghao","year":"2021","unstructured":"Minghao Xu, Hang Wang, Bingbing Ni, Hongyu Guo, and Jian Tang. 2021. Selfsupervised graph-level representation learning with local and global structure. In International Conference on Machine Learning. PMLR, 11548--11558."},{"key":"e_1_3_2_1_52_1","volume-title":"Howtransferable are features in deep neural networks? Advances in neural information processing systems 27","author":"Yosinski Jason","year":"2014","unstructured":"Jason Yosinski, Jeff Clune, Yoshua Bengio, and Hod Lipson. 2014. Howtransferable are features in deep neural networks? Advances in neural information processing systems 27 (2014)."},{"key":"e_1_3_2_1_53_1","volume-title":"International Conference on Machine Learning. PMLR, 12121--12132","author":"You Yuning","year":"2021","unstructured":"Yuning You, Tianlong Chen, Yang Shen, and Zhangyang Wang. 2021. Graph contrastive learning automated. In International Conference on Machine Learning. PMLR, 12121--12132."},{"key":"e_1_3_2_1_54_1","first-page":"5812","article-title":"Graph contrastive learning with augmentations","volume":"33","author":"You Yuning","year":"2020","unstructured":"Yuning You, Tianlong Chen, Yongduo Sui, Ting Chen, Zhangyang Wang, and Yang Shen. 2020. Graph contrastive learning with augmentations. Advances in Neural Information Processing Systems 33 (2020), 5812--5823.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3292500.3330961"},{"key":"e_1_3_2_1_56_1","volume-title":"Revisiting few-sample BERT fine-tuning. arXiv preprint arXiv:2006.05987","author":"Zhang Tianyi","year":"2020","unstructured":"Tianyi Zhang, Felix Wu, Arzoo Katiyar, Kilian Q Weinberger, and Yoav Artzi. 2020. Revisiting few-sample BERT fine-tuning. arXiv preprint arXiv:2006.05987 (2020)."},{"key":"e_1_3_2_1_57_1","first-page":"15870","article-title":"Motifbased graph self-supervised learning for molecular property prediction","volume":"34","author":"Zhang Zaixi","year":"2021","unstructured":"Zaixi Zhang, Qi Liu, Hao Wang, Chengqiang Lu, and Chee-Kong Lee. 2021. Motifbased graph self-supervised learning for molecular property prediction. Advances in Neural Information Processing Systems 34 (2021), 15870--15882.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442381.3449802"}],"event":{"name":"MM '24: The 32nd ACM International Conference on Multimedia","location":"Melbourne VIC Australia","acronym":"MM '24","sponsor":["SIGMM ACM Special Interest Group on Multimedia"]},"container-title":["Proceedings of the 32nd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681018","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3664647.3681018","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:17:36Z","timestamp":1750295856000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681018"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,28]]},"references-count":58,"alternative-id":["10.1145\/3664647.3681018","10.1145\/3664647"],"URL":"https:\/\/doi.org\/10.1145\/3664647.3681018","relation":{},"subject":[],"published":{"date-parts":[[2024,10,28]]},"assertion":[{"value":"2024-10-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}