{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,16]],"date-time":"2026-04-16T08:59:17Z","timestamp":1776329957670,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":52,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,8,14]],"date-time":"2022-08-14T00:00:00Z","timestamp":1660435200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100004739","name":"Youth Innovation Promotion Association of the Chinese Academy of Sciences","doi-asserted-by":"publisher","award":["2018497"],"award-info":[{"award-number":["2018497"]}],"id":[{"id":"10.13039\/501100004739","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61836011"],"award-info":[{"award-number":["61836011"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,8,14]]},"DOI":"10.1145\/3534678.3539368","type":"proceedings-article","created":{"date-parts":[[2022,8,12]],"date-time":"2022-08-12T19:06:41Z","timestamp":1660331201000},"page":"2626-2636","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":43,"title":["Unified 2D and 3D Pre-Training of Molecular Representations"],"prefix":"10.1145","author":[{"given":"Jinhua","family":"Zhu","sequence":"first","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"given":"Yingce","family":"Xia","sequence":"additional","affiliation":[{"name":"Microsoft Research Asia, Beijing, China"}]},{"given":"Lijun","family":"Wu","sequence":"additional","affiliation":[{"name":"Microsoft Research Asia, Beijing, China"}]},{"given":"Shufang","family":"Xie","sequence":"additional","affiliation":[{"name":"Microsoft Research Asia, Beijing, China"}]},{"given":"Tao","family":"Qin","sequence":"additional","affiliation":[{"name":"Microsoft Research Asia, Beijing, China"}]},{"given":"Wengang","family":"Zhou","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"given":"Houqiang","family":"Li","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"given":"Tie-Yan","family":"Liu","sequence":"additional","affiliation":[{"name":"Microsoft Research Asia, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2022,8,14]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"Alvaro Sanchez-Gonzalez, Jacklynn Stott, Shantanu Thakoor, et al.","author":"Addanki Ravichandra","year":"2021","unstructured":"Ravichandra Addanki, Peter W Battaglia, David Budden, Andreea Deac, Jonathan Godwin, Thomas Keck, Wai Lok Sibon Li, Alvaro Sanchez-Gonzalez, Jacklynn Stott, Shantanu Thakoor, et al. 2021. Large-scale graph representation learning with very deep GNNs and self-supervision. arXiv:2107.09422 (2021)."},{"key":"e_1_3_2_2_2_1","volume-title":"NeurIPS","volume":"32","author":"Anderson Brandon","year":"2019","unstructured":"Brandon Anderson, Truong Son Hy, and Risi Kondor. 2019Cormorant: Covariant molecular neural networks. NeurIPS, Vol. 32 (2019)."},{"key":"e_1_3_2_2_3_1","volume-title":"Geom: Energy-annotated molecular conformations for property prediction and molecular generation. arXiv preprint arXiv:2006.05531","author":"Axelrod Simon","year":"2020","unstructured":"Simon Axelrod and Rafael Gomez-Bombarelli. 2020. Geom: Energy-annotated molecular conformations for property prediction and molecular generation. arXiv preprint arXiv:2006.05531 (2020)."},{"key":"e_1_3_2_2_4_1","first-page":"1","article-title":"Algebraic graph-assisted bidirectional transformers for molecular property prediction","volume":"12","author":"Chen Dong","year":"2021","unstructured":"Dong Chen, Kaifu Gao, Duc Duy Nguyen, Xin Chen, Yi Jiang, Guo-Wei Wei, and Feng Pan. 2021. Algebraic graph-assisted bidirectional transformers for molecular property prediction. Nature Communications, Vol. 12, 1 (2021), 1--9.","journal-title":"Nature Communications"},{"key":"e_1_3_2_2_5_1","volume-title":"Chemberta: Large-scale self-supervised pretraining for molecular property prediction. arXiv:2010.09885","author":"Chithrananda Seyone","year":"2020","unstructured":"Seyone Chithrananda, Gabriel Grand, and Bharath Ramsundar. 2020. Chemberta: Large-scale self-supervised pretraining for molecular property prediction. arXiv:2010.09885 (2020)."},{"key":"e_1_3_2_2_6_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171--4186.","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL. 4171--4186."},{"key":"e_1_3_2_2_7_1","unstructured":"Alexey Dosovitskiy Lucas Beyer Alexander Kolesnikov Dirk Weissenborn Xiaohua Zhai Thomas Unterthiner Mostafa Dehghani Matthias Minderer Georg Heigold Sylvain Gelly Jakob Uszkoreit and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In ICLR."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1021\/ci010132r"},{"key":"e_1_3_2_2_9_1","unstructured":"Fabian Fuchs Daniel Worrall Volker Fischer and Max Welling. [n. d.]. SE(3)-Transformers: 3D Roto-Translation Equivariant Attention Networks. In NeurIPS."},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1039\/D0CP00305K"},{"key":"e_1_3_2_2_11_1","volume-title":"Dahl","author":"Gilmer Justin","year":"2017","unstructured":"Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl. 2017. Neural Message Passing for Quantum Chemistry. In ICML. 1263--1272."},{"key":"e_1_3_2_2_12_1","volume-title":"Smiles transformer: Pre-trained molecular fingerprint for low data drug discovery. arXiv:1911.04738","author":"Honda Shion","year":"2019","unstructured":"Shion Honda, Shoi Shi, and Hiroki R Ueda. 2019. Smiles transformer: Pre-trained molecular fingerprint for low data drug discovery. arXiv:1911.04738 (2019)."},{"key":"e_1_3_2_2_13_1","volume-title":"OGB-LSC: A Large-Scale Challenge for Machine Learning on Graphs. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2).","author":"Hu Weihua","year":"2021","unstructured":"Weihua Hu, Matthias Fey, Hongyu Ren, Maho Nakata, Yuxiao Dong, and Jure Leskovec. 2021. OGB-LSC: A Large-Scale Challenge for Machine Learning on Graphs. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2)."},{"key":"e_1_3_2_2_14_1","volume-title":"2020 a. Open Graph Benchmark: Datasets for Machine Learning on Graphs. arxiv","author":"Hu Weihua","year":"2020","unstructured":"Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. 2020 a. Open Graph Benchmark: Datasets for Machine Learning on Graphs. arxiv, Vol. abs\/2005.00687 (2020)."},{"key":"e_1_3_2_2_15_1","unstructured":"Weihua Hu Bowen Liu Joseph Gomes Marinka Zitnik Percy Liang Vijay Pande and Jure Leskovec. 2020 b. Strategies for Pre-training Graph Neural Networks. In ICLR."},{"key":"e_1_3_2_2_16_1","volume-title":"ICML","volume":"37","author":"Ioffe Sergey","year":"2015","unstructured":"Sergey Ioffe and Christian Szegedy. 2015. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In ICML, Vol. 37. Lille, France, 448--456."},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2105070118"},{"key":"e_1_3_2_2_18_1","volume-title":"One Model To Learn Them All. CoRR","author":"Kaiser Lukasz","year":"2017","unstructured":"Lukasz Kaiser, Aidan N. Gomez, Noam Shazeer, Ashish Vaswani, Niki Parmar, Llion Jones, and Jakob Uszkoreit. 2017. One Model To Learn Them All. CoRR, Vol. abs\/1706.05137 (2017). showeprint[arXiv]1706.05137"},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jmgm.2006.04.002"},{"key":"e_1_3_2_2_20_1","unstructured":"Johannes Klicpera Janek Gro\u00df and Stephan G\u00fcnnemann. 2020. Directional Message Passing for Molecular Graphs. In ICLR ."},{"key":"e_1_3_2_2_21_1","volume-title":"FLAG: Adversarial Data Augmentation for Graph Neural Networks. CoRR","author":"Kong Kezhi","year":"2020","unstructured":"Kezhi Kong, Guohao Li, Mucong Ding, Zuxuan Wu, Chen Zhu, Bernard Ghanem, Gavin Taylor, and Tom Goldstein. 2020. FLAG: Adversarial Data Augmentation for Graph Neural Networks. CoRR, Vol. abs\/2010.09891 (2020)."},{"key":"e_1_3_2_2_22_1","unstructured":"Shengchao Liu Hanchen Wang Weiyang Liu Joan Lasenby Hongyu Guo and Jian Tang. 2022. Pre-training Molecular Graph Representation with 3D Geometry. In ICLR."},{"key":"e_1_3_2_2_23_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33011052"},{"key":"e_1_3_2_2_25_1","volume-title":"NeurIPS","volume":"34","author":"Luo Shitong","year":"2021","unstructured":"Shitong Luo, Chence Shi, Minkai Xu, and Jian Tang. 2021. Predicting Molecular Conformation via Dynamic Graph Score Matching. In NeurIPS, Vol. 34."},{"key":"e_1_3_2_2_26_1","volume-title":"Scientific Reports","volume":"9","author":"Mansimov Elman","year":"2019","unstructured":"Elman Mansimov, Omar Mahmood, Seokho Kang, and Kyunghyun Cho. 2019. Molecular Geometry Prediction using a Deep Generative Graph Neural Network. Scientific Reports, Vol. 9, 1 (31 Dec 2019), 20381."},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-020-00455-2"},{"key":"e_1_3_2_2_28_1","first-page":"8821","article-title":"Zero-Shot Text-to-Image Generation","volume":"139","author":"Ramesh Aditya","year":"2021","unstructured":"Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. 2021. Zero-Shot Text-to-Image Generation. In ICML, Vol. 139. 8821--8831.","journal-title":"ICML"},{"key":"e_1_3_2_2_29_1","first-page":"12559","article-title":"Self-Supervised Graph Transformer on Large-Scale Molecular Data","volume":"33","author":"Rong Yu","year":"2020","unstructured":"Yu Rong, Yatao Bian, Tingyang Xu, Weiyang Xie, Ying WEI, Wenbing Huang, and Junzhou Huang. 2020. Self-Supervised Graph Transformer on Large-Scale Molecular Data. In NeurIPS, Vol. 33. 12559--12571.","journal-title":"NeurIPS"},{"key":"e_1_3_2_2_30_1","unstructured":"Kristof Sch\u00fctt Oliver Unke and Michael Gastegger. 2021. Equivariant message passing for the prediction of tensorial properties and molecular spectra. In ICML."},{"key":"e_1_3_2_2_31_1","volume-title":"NeurIPS (Long Beach, California, USA)","author":"Sch\u00fctt K. T.","unstructured":"K. T. Sch\u00fctt, P.-J. Kindermans, H. E. Sauceda, S. Chmiela, A. Tkatchenko, and K.-R. M\u00fcller. 2017. SchNet: A Continuous-Filter Convolutional Neural Network for Modeling Quantum Interactions. In NeurIPS (Long Beach, California, USA). Curran Associates Inc., Red Hook, NY, USA, 992--1002."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-33-6191-1_21"},{"key":"e_1_3_2_2_33_1","unstructured":"Chence Shi Shitong Luo Minkai Xu and Jian Tang. 2021. Learning gradient fields for molecular conformation generation. In ICML. PMLR 9558--9568."},{"key":"e_1_3_2_2_34_1","volume-title":"Heterogeneous molecular graph neural networks for predicting molecule properties","author":"Shui Zeren","unstructured":"Zeren Shui and George Karypis. 2020. Heterogeneous molecular graph neural networks for predicting molecule properties. In ICDM. IEEE, 492--500."},{"key":"e_1_3_2_2_35_1","volume-title":"ICML","volume":"119","author":"Simm Gregor","year":"2020","unstructured":"Gregor Simm and Jose Miguel Hernandez-Lobato. 2020. A Generative Model for Molecular Distance Geometry. In ICML, Vol. 119. PMLR, 8949--8958."},{"key":"e_1_3_2_2_36_1","volume-title":"3D Infomax improves GNNs for Molecular Property Prediction. arXiv preprint arXiv:2110.04126","author":"St\u00e4rk Hannes","year":"2021","unstructured":"Hannes St\u00e4rk, Dominique Beaini, Gabriele Corso, Prudencio Tossou, Christian Dallago, Stephan G\u00fcnnemann, and Pietro Li\u00f2. 2021. 3D Infomax improves GNNs for Molecular Property Prediction. arXiv preprint arXiv:2110.04126 (2021)."},{"key":"e_1_3_2_2_37_1","volume-title":"3D Infomax improves GNNs for Molecular Property Prediction. CoRR","author":"St\u00e4rk Hannes","year":"2021","unstructured":"Hannes St\u00e4rk, Dominique Beaini, Gabriele Corso, Prudencio Tossou, Christian Dallago, Stephan G\u00fcnnemann, and Pietro Li\u00f2. 2021. 3D Infomax improves GNNs for Molecular Property Prediction. CoRR, Vol. abs\/2110.04126 (2021)."},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cell.2020.01.021"},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-020-00475-y"},{"key":"e_1_3_2_2_40_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is All you Need. In NeurIPS."},{"key":"e_1_3_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3307339.3342186"},{"key":"e_1_3_2_2_42_1","volume-title":"MolCLR: molecular contrastive learning of representations via graph neural networks. arXiv preprint arXiv:2102.10056","author":"Wang Yuyang","year":"2021","unstructured":"Yuyang Wang, Jianren Wang, Zhonglin Cao, and Amir Barati Farimani. 2021. MolCLR: molecular contrastive learning of representations via graph neural networks. arXiv preprint arXiv:2102.10056 (2021)."},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1021\/ci00057a005"},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.7b00558"},{"key":"e_1_3_2_2_45_1","volume-title":"MoleculeNet: a benchmark for molecular machine learning. Chemical science","author":"Wu Zhenqin","year":"2018","unstructured":"Zhenqin Wu, Bharath Ramsundar, Evan N Feinberg, Joseph Gomes, Caleb Geniesse, Aneesh S Pappu, Karl Leswing, and Vijay Pande. 2018. MoleculeNet: a benchmark for molecular machine learning. Chemical science, Vol. 9, 2 (2018), 513--530."},{"key":"e_1_3_2_2_46_1","volume-title":"ICLR","author":"Xu Keyulu","year":"2019","unstructured":"Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. 2019. How Powerful are Graph Neural Networks?. In ICLR 2019."},{"key":"e_1_3_2_2_47_1","unstructured":"Minkai Xu Lantao Yu Yang Song Chence Shi Stefano Ermon and Jian Tang. 2022. GeoDiff: A Geometric Diffusion Model for Molecular Conformation Generation. In ICLR."},{"key":"e_1_3_2_2_48_1","unstructured":"Chengxuan Ying Tianle Cai Shengjie Luo Shuxin Zheng Guolin Ke Di He Yanming Shen and Tie-Yan Liu. 2021. Do Transformers Really Perform Badly for Graph Representation?. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_2_49_1","unstructured":"Yuning You Tianlong Chen Yongduo Sui Ting Chen Zhangyang Wang and Yang Shen. 2020. Graph Contrastive Learning with Augmentations. In NeurIPS."},{"key":"e_1_3_2_2_50_1","volume-title":"Philip HS Torr, and Vladlen Koltun","author":"Zhao Hengshuang","year":"2021","unstructured":"Hengshuang Zhao, Li Jiang, Jiaya Jia, Philip HS Torr, and Vladlen Koltun. 2021. Point transformer. In ICCV. 16259--16268."},{"key":"e_1_3_2_2_51_1","volume-title":"Direct molecular conformation generation. arXiv preprint arXiv:2202.01356","author":"Zhu Jinhua","year":"2022","unstructured":"Jinhua Zhu, Yingce Xia, Chang Liu, Lijun Wu, Shufang Xie, Tong Wang, Yusong Wang, Wengang Zhou, Tao Qin, Houqiang Li, and Tie-Yan Liu. 2022. Direct molecular conformation generation. arXiv preprint arXiv:2202.01356 (2022)."},{"key":"e_1_3_2_2_52_1","volume-title":"Dual-view molecule pre-training. arXiv preprint arXiv:2106.10234","author":"Zhu Jinhua","year":"2021","unstructured":"Jinhua Zhu, Yingce Xia, Tao Qin, Wengang Zhou, Houqiang Li, and Tie-Yan Liu. 2021. Dual-view molecule pre-training. arXiv preprint arXiv:2106.10234 (2021)."}],"event":{"name":"KDD '22: The 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Washington DC USA","acronym":"KDD '22","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3534678.3539368","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3534678.3539368","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:47Z","timestamp":1750186967000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3534678.3539368"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,14]]},"references-count":52,"alternative-id":["10.1145\/3534678.3539368","10.1145\/3534678"],"URL":"https:\/\/doi.org\/10.1145\/3534678.3539368","relation":{},"subject":[],"published":{"date-parts":[[2022,8,14]]},"assertion":[{"value":"2022-08-14","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}