{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T13:15:40Z","timestamp":1775913340529,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":72,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,8,4]],"date-time":"2023-08-04T00:00:00Z","timestamp":1691107200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61836011,62021001"],"award-info":[{"award-number":["61836011,62021001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,8,6]]},"DOI":"10.1145\/3580305.3599317","type":"proceedings-article","created":{"date-parts":[[2023,8,4]],"date-time":"2023-08-04T18:13:58Z","timestamp":1691172838000},"page":"3615-3627","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":46,"title":["Dual-view Molecular Pre-training"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2157-9077","authenticated-orcid":false,"given":"Jinhua","family":"Zhu","sequence":"first","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9823-9033","authenticated-orcid":false,"given":"Yingce","family":"Xia","sequence":"additional","affiliation":[{"name":"Microsoft Research AI4Science, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3530-590X","authenticated-orcid":false,"given":"Lijun","family":"Wu","sequence":"additional","affiliation":[{"name":"Microsoft Research AI4Science, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7126-0139","authenticated-orcid":false,"given":"Shufang","family":"Xie","sequence":"additional","affiliation":[{"name":"Renmin University of China, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1690-9836","authenticated-orcid":false,"given":"Wengang","family":"Zhou","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9095-0776","authenticated-orcid":false,"given":"Tao","family":"Qin","sequence":"additional","affiliation":[{"name":"Microsoft Research AI4Science, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2188-3028","authenticated-orcid":false,"given":"Houqiang","family":"Li","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0476-8020","authenticated-orcid":false,"given":"Tie-Yan","family":"Liu","sequence":"additional","affiliation":[{"name":"Microsoft Research AI4Science, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2023,8,4]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/279943.279962"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1021\/jacsau.1c00246"},{"key":"e_1_3_2_2_3_1","volume-title":"International conference on machine learning. PMLR, 1597--1607","author":"Chen Ting","year":"2020","unstructured":"Ting Chen , Simon Kornblith , Mohammad Norouzi , and Geoffrey Hinton . 2020 . A simple framework for contrastive learning of visual representations . In International conference on machine learning. PMLR, 1597--1607 . Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. 2020. A simple framework for contrastive learning of visual representations. In International conference on machine learning. PMLR, 1597--1607."},{"key":"e_1_3_2_2_4_1","volume-title":"ChemBERTa: Large-Scale Self-Supervised Pretraining for Molecular Property Prediction. arXiv preprint arXiv:2010.09885","author":"Chithrananda Seyone","year":"2020","unstructured":"Seyone Chithrananda , Gabe Grand , and Bharath Ramsundar . 2020. ChemBERTa: Large-Scale Self-Supervised Pretraining for Molecular Property Prediction. arXiv preprint arXiv:2010.09885 ( 2020 ). Seyone Chithrananda, Gabe Grand, and Bharath Ramsundar. 2020. ChemBERTa: Large-Scale Self-Supervised Pretraining for Molecular Property Prediction. arXiv preprint arXiv:2010.09885 (2020)."},{"key":"e_1_3_2_2_5_1","volume-title":"Computer-assisted retrosynthesis based on molecular similarity. ACS central science","author":"Coley Connor W","year":"2017","unstructured":"Connor W Coley , Luke Rogers , William H Green , and Klavs F Jensen . 2017. Computer-assisted retrosynthesis based on molecular similarity. ACS central science , Vol. 3 , 12 ( 2017 ), 1237--1245. Connor W Coley, Luke Rogers, William H Green, and Klavs F Jensen. 2017. Computer-assisted retrosynthesis based on molecular similarity. ACS central science, Vol. 3, 12 (2017), 1237--1245."},{"key":"e_1_3_2_2_6_1","volume-title":"Support-vector networks. Machine learning","author":"Cortes Corinna","year":"1995","unstructured":"Corinna Cortes and Vladimir Vapnik . 1995. Support-vector networks. Machine learning , Vol. 20 , 3 ( 1995 ), 273--297. Corinna Cortes and Vladimir Vapnik. 1995. Support-vector networks. Machine learning, Vol. 20, 3 (1995), 273--297."},{"key":"e_1_3_2_2_7_1","volume-title":"Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextquotesingle Alch\u00e9-Buc","author":"Dai Hanjun","unstructured":"Hanjun Dai , Chengtao Li , Connor W Coley , Bo Dai , and Le Song . 2019. Retrosynthesis prediction with conditional graph logic network . In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextquotesingle Alch\u00e9-Buc , E. Fox, and R. Garnett (Eds.), Vol. 32 . Curran Associates, Inc. Hanjun Dai, Chengtao Li, Connor W Coley, Bo Dai, and Le Song. 2019. Retrosynthesis prediction with conditional graph logic network. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextquotesingle Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32. Curran Associates, Inc."},{"key":"e_1_3_2_2_8_1","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-020-00460-5"},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.1979.4766909"},{"key":"e_1_3_2_2_10_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2018 . Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018). Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_2_11_1","volume-title":"Speech-Transformer: A No-Recurrence Sequence-to-Sequence Model for Speech Recognition. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 5884--5888","author":"Dong Linhao","year":"2018","unstructured":"Linhao Dong , Shuang Xu , and Bo Xu . 2018 . Speech-Transformer: A No-Recurrence Sequence-to-Sequence Model for Speech Recognition. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 5884--5888 . https:\/\/doi.org\/10.1109\/ICASSP.2018.8462506 10.1109\/ICASSP.2018.8462506 Linhao Dong, Shuang Xu, and Bo Xu. 2018. Speech-Transformer: A No-Recurrence Sequence-to-Sequence Model for Speech Recognition. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 5884--5888. https:\/\/doi.org\/10.1109\/ICASSP.2018.8462506"},{"key":"e_1_3_2_2_12_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=YicbFdNTTy","author":"Dosovitskiy Alexey","year":"2021","unstructured":"Alexey Dosovitskiy , Lucas Beyer , Alexander Kolesnikov , Dirk Weissenborn , Xiaohua Zhai , Thomas Unterthiner , Mostafa Dehghani , Matthias Minderer , Georg Heigold , Sylvain Gelly , Jakob Uszkoreit , and Neil Houlsby . 2021 . An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale . In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=YicbFdNTTy Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=YicbFdNTTy"},{"key":"e_1_3_2_2_13_1","volume-title":"Molecular representation learning with language models and domain-relevant auxiliary tasks. arXiv preprint arXiv:2011.13230","author":"Fabian Benedek","year":"2020","unstructured":"Benedek Fabian , Thomas Edlich , H\u00e9l\u00e9na Gaspar , Marwin Segler , Joshua Meyers , Marco Fiscato , and Mohamed Ahmed . 2020a. Molecular representation learning with language models and domain-relevant auxiliary tasks. arXiv preprint arXiv:2011.13230 ( 2020 ). Benedek Fabian, Thomas Edlich, H\u00e9l\u00e9na Gaspar, Marwin Segler, Joshua Meyers, Marco Fiscato, and Mohamed Ahmed. 2020a. Molecular representation learning with language models and domain-relevant auxiliary tasks. arXiv preprint arXiv:2011.13230 (2020)."},{"key":"e_1_3_2_2_14_1","volume-title":"Machine Learning for Molecules Workshop at NeurIPS","author":"Fabian Benedek","year":"2020","unstructured":"Benedek Fabian , Thomas Edlich , H\u00e9l\u00e9na Gaspar , Marwin Segler , Joshua Meyers , Marco Fiscato , and Mohamed Ahmed . 2020 b. Molecular representation learning with language models and domain-relevant auxiliary tasks . In Machine Learning for Molecules Workshop at NeurIPS 2020. Benedek Fabian, Thomas Edlich, H\u00e9l\u00e9na Gaspar, Marwin Segler, Joshua Meyers, Marco Fiscato, and Mohamed Ahmed. 2020b. Molecular representation learning with language models and domain-relevant auxiliary tasks. In Machine Learning for Molecules Workshop at NeurIPS 2020."},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-021-00438-4"},{"key":"e_1_3_2_2_16_1","volume-title":"Lin (Eds.)","volume":"33","author":"Grill Jean-Bastien","year":"2020","unstructured":"Jean-Bastien Grill , Florian Strub , Florent Altch\u00e9 , Corentin Tallec , Pierre Richemond , Elena Buchatskaya , Carl Doersch , Bernardo Avila Pires , Zhaohan Guo , Mohammad Gheshlaghi Azar , Bilal Piot , koray kavukcuoglu, Remi Munos , and Michal Valko . 2020 a. Bootstrap Your Own Latent - A New Approach to Self-Supervised Learning. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H . Lin (Eds.) , Vol. 33 . Curran Associates, Inc., 21271--21284. https:\/\/proceedings.neurips.cc\/paper\/ 2020\/file\/f3ada80d5c4ee70142b17b8192b2958e-Paper.pdf Jean-Bastien Grill, Florian Strub, Florent Altch\u00e9, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, Bilal Piot, koray kavukcuoglu, Remi Munos, and Michal Valko. 2020a. Bootstrap Your Own Latent - A New Approach to Self-Supervised Learning. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 21271--21284. https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/f3ada80d5c4ee70142b17b8192b2958e-Paper.pdf"},{"key":"e_1_3_2_2_17_1","volume-title":"Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, et al.","author":"Grill Jean-Bastien","year":"2020","unstructured":"Jean-Bastien Grill , Florian Strub , Florent Altch\u00e9 , Corentin Tallec , Pierre H Richemond , Elena Buchatskaya , Carl Doersch , Bernardo Avila Pires , Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, et al. 2020 b. Bootstrap your own latent: A new approach to self-supervised learning. arXiv preprint arXiv:2006.07733 (2020). Jean-Bastien Grill, Florian Strub, Florent Altch\u00e9 , Corentin Tallec, Pierre H Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, et al. 2020b. Bootstrap your own latent: A new approach to self-supervised learning. arXiv preprint arXiv:2006.07733 (2020)."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2006.100"},{"key":"e_1_3_2_2_19_1","volume-title":"International Conference on Machine Learning. PMLR, 4116--4126","author":"Hassani Kaveh","year":"2020","unstructured":"Kaveh Hassani and Amir Hosein Khasahmadi . 2020 . Contrastive multi-view representation learning on graphs . In International Conference on Machine Learning. PMLR, 4116--4126 . Kaveh Hassani and Amir Hosein Khasahmadi. 2020. Contrastive multi-view representation learning on graphs. In International Conference on Machine Learning. PMLR, 4116--4126."},{"key":"e_1_3_2_2_20_1","volume-title":"Ola Engkvist, Christian Tyrchan, Werngard Czechtizky, et al.","author":"He Jiazhen","year":"2021","unstructured":"Jiazhen He , Felix Mattsson , Marcus Forsberg , Esben Jannik Bjerrum , Ola Engkvist, Christian Tyrchan, Werngard Czechtizky, et al. 2021 . Transformer Neural Network for Structure Constrained Molecular Optimization . (2021). Jiazhen He, Felix Mattsson, Marcus Forsberg, Esben Jannik Bjerrum, Ola Engkvist, Christian Tyrchan, Werngard Czechtizky, et al. 2021. Transformer Neural Network for Structure Constrained Molecular Optimization. (2021)."},{"key":"#cr-split#-e_1_3_2_2_21_1.1","doi-asserted-by":"crossref","unstructured":"Jiyan He Keyu Tian Shengjie Luo Yaosen Min Shuxin Zheng Yu Shi Di He Haiguang Liu Nenghai Yu Liwei Wang Ji Wu and Tie-Yan Liu. 2022. Masked Molecule Modeling: A New Paradigm of Molecular Representation Learning for Chemistry Understanding. https:\/\/doi.org\/10.21203\/rs.3.rs-1746019\/v1 10.21203\/rs.3.rs-1746019","DOI":"10.21203\/rs.3.rs-1746019\/v1"},{"key":"#cr-split#-e_1_3_2_2_21_1.2","doi-asserted-by":"crossref","unstructured":"Jiyan He Keyu Tian Shengjie Luo Yaosen Min Shuxin Zheng Yu Shi Di He Haiguang Liu Nenghai Yu Liwei Wang Ji Wu and Tie-Yan Liu. 2022. Masked Molecule Modeling: A New Paradigm of Molecular Representation Learning for Chemistry Understanding. https:\/\/doi.org\/10.21203\/rs.3.rs-1746019\/v1","DOI":"10.21203\/rs.3.rs-1746019\/v1"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_2_24_1","volume-title":"Proceedings of 3rd international conference on document analysis and recognition","volume":"1","author":"Ho Tin Kam","year":"1995","unstructured":"Tin Kam Ho . 1995 . Random decision forests . In Proceedings of 3rd international conference on document analysis and recognition , Vol. 1 . IEEE, 278--282. Tin Kam Ho. 1995. Random decision forests. In Proceedings of 3rd international conference on document analysis and recognition, Vol. 1. IEEE, 278--282."},{"key":"e_1_3_2_2_25_1","volume-title":"SMILES transformer: pre-trained molecular fingerprint for low data drug discovery. arXiv preprint arXiv:1911.04738","author":"Honda Shion","year":"2019","unstructured":"Shion Honda , Shoi Shi , and Hiroki R Ueda . 2019. SMILES transformer: pre-trained molecular fingerprint for low data drug discovery. arXiv preprint arXiv:1911.04738 ( 2019 ). Shion Honda, Shoi Shi, and Hiroki R Ueda. 2019. SMILES transformer: pre-trained molecular fingerprint for low data drug discovery. arXiv preprint arXiv:1911.04738 (2019)."},{"key":"e_1_3_2_2_26_1","volume-title":"Strategies for Pre-training Graph Neural Networks. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=HJlWWJSFDH","author":"Hu Weihua","year":"2020","unstructured":"Weihua Hu , Bowen Liu , Joseph Gomes , Marinka Zitnik , Percy Liang , Vijay Pande , and Jure Leskovec . 2020 . Strategies for Pre-training Graph Neural Networks. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=HJlWWJSFDH Weihua Hu, Bowen Liu, Joseph Gomes, Marinka Zitnik, Percy Liang, Vijay Pande, and Jure Leskovec. 2020. Strategies for Pre-training Graph Neural Networks. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=HJlWWJSFDH"},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jmedchem.5b01746"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1063\/1.4964627"},{"key":"e_1_3_2_2_29_1","volume-title":"International conference on machine learning. PMLR, 448--456","author":"Ioffe Sergey","year":"2015","unstructured":"Sergey Ioffe and Christian Szegedy . 2015 . Batch normalization: Accelerating deep network training by reducing internal covariate shift . In International conference on machine learning. PMLR, 448--456 . Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International conference on machine learning. PMLR, 448--456."},{"key":"e_1_3_2_2_30_1","volume-title":"Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980","author":"Kingma Diederik P","year":"2014","unstructured":"Diederik P Kingma and Jimmy Ba . 2014 . Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014). Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)."},{"key":"e_1_3_2_2_31_1","volume-title":"Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907","author":"Kipf Thomas N","year":"2016","unstructured":"Thomas N Kipf and Max Welling . 2016. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 ( 2016 ). Thomas N Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)."},{"key":"e_1_3_2_2_32_1","unstructured":"Greg Landrum. 2013. RDKit: A software suite for cheminformatics computational chemistry and predictive modeling.  Greg Landrum. 2013. RDKit: A software suite for cheminformatics computational chemistry and predictive modeling."},{"key":"e_1_3_2_2_33_1","volume-title":"Deepergcn: All you need to train deeper gcns. arXiv preprint arXiv:2006.07739","author":"Li Guohao","year":"2020","unstructured":"Guohao Li , Chenxin Xiong , Ali Thabet , and Bernard Ghanem . 2020 b. Deepergcn: All you need to train deeper gcns. arXiv preprint arXiv:2006.07739 (2020). Guohao Li, Chenxin Xiong, Ali Thabet, and Bernard Ghanem. 2020b. Deepergcn: All you need to train deeper gcns. arXiv preprint arXiv:2006.07739 (2020)."},{"key":"e_1_3_2_2_34_1","volume-title":"Learn molecular representations from large-scale unlabeled molecules for drug discovery. arXiv preprint arXiv:2012.11175","author":"Li Pengyong","year":"2020","unstructured":"Pengyong Li , Jun Wang , Yixuan Qiao , Hao Chen , Yihuan Yu , Xiaojun Yao , Peng Gao , Guotong Xie , and Sen Song . 2020a. Learn molecular representations from large-scale unlabeled molecules for drug discovery. arXiv preprint arXiv:2012.11175 ( 2020 ). Pengyong Li, Jun Wang, Yixuan Qiao, Hao Chen, Yihuan Yu, Xiaojun Yao, Peng Gao, Guotong Xie, and Sen Song. 2020a. Learn molecular representations from large-scale unlabeled molecules for drug discovery. arXiv preprint arXiv:2012.11175 (2020)."},{"key":"e_1_3_2_2_35_1","volume-title":"The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=KwmPfARgOTD","author":"Liao Yi-Lun","year":"2023","unstructured":"Yi-Lun Liao and Tess Smidt . 2023 . Equiformer: Equivariant Graph Attention Transformer for 3D Atomistic Graphs . In The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=KwmPfARgOTD Yi-Lun Liao and Tess Smidt. 2023. Equiformer: Equivariant Graph Attention Transformer for 3D Atomistic Graphs. In The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=KwmPfARgOTD"},{"key":"e_1_3_2_2_36_1","volume-title":"Stephen Ho, Jack Sloane, Paul Wender, and Vijay Pande.","author":"Liu Bowen","year":"2017","unstructured":"Bowen Liu , Bharath Ramsundar , Prasad Kawthekar , Jade Shi , Joseph Gomes , Quang Luu Nguyen , Stephen Ho, Jack Sloane, Paul Wender, and Vijay Pande. 2017 . Retrosynthetic reaction prediction using neural sequence-to-sequence models. ACS central science, Vol. 3 , 10 (2017), 1103--1113. Bowen Liu, Bharath Ramsundar, Prasad Kawthekar, Jade Shi, Joseph Gomes, Quang Luu Nguyen, Stephen Ho, Jack Sloane, Paul Wender, and Vijay Pande. 2017. Retrosynthetic reaction prediction using neural sequence-to-sequence models. ACS central science, Vol. 3, 10 (2017), 1103--1113."},{"key":"e_1_3_2_2_37_1","volume-title":"Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextquotesingle Alch\u00e9-Buc","author":"Liu Shengchao","year":"2019","unstructured":"Shengchao Liu , Mehmet F Demirel , and Yingyu Liang . 2019a. N-Gram Graph: Simple Unsupervised Representation for Graphs, with Applications to Molecules . In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextquotesingle Alch\u00e9-Buc , E. Fox, and R. Garnett (Eds.), Vol. 32 . Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/ 2019 \/file\/2f3926f0a9613f3c3cc21d52a3cdb4d9-Paper.pdf Shengchao Liu, Mehmet F Demirel, and Yingyu Liang. 2019a. N-Gram Graph: Simple Unsupervised Representation for Graphs, with Applications to Molecules. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextquotesingle Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/2f3926f0a9613f3c3cc21d52a3cdb4d9-Paper.pdf"},{"key":"e_1_3_2_2_38_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=xQUe1pOKPam","author":"Liu Shengchao","year":"2022","unstructured":"Shengchao Liu , Hanchen Wang , Weiyang Liu , Joan Lasenby , Hongyu Guo , and Jian Tang . 2022 . Pre-training Molecular Graph Representation with 3D Geometry . In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=xQUe1pOKPam Shengchao Liu, Hanchen Wang, Weiyang Liu, Joan Lasenby, Hongyu Guo, and Jian Tang. 2022. Pre-training Molecular Graph Representation with 3D Geometry. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=xQUe1pOKPam"},{"key":"e_1_3_2_2_39_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu , Myle Ott , Naman Goyal , Jingfei Du , Mandar Joshi , Danqi Chen , Omer Levy , Mike Lewis , Luke Zettlemoyer , and Veselin Stoyanov . 2019 b. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019). Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019b. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_2_40_1","unstructured":"Zhengying Liu Adrien Pavao Zhen Xu Sergio Escalera Fabio Ferreira Isabelle Guyon Sirui Hong Frank Hutter Rongrong Ji Julio C Junior etal 2021. Winning solutions and post-challenge analyses of the ChaLearn AutoDL challenge 2019. IEEE Transactions on Pattern Analysis and Machine Intelligence (2021).  Zhengying Liu Adrien Pavao Zhen Xu Sergio Escalera Fabio Ferreira Isabelle Guyon Sirui Hong Frank Hutter Rongrong Ji Julio C Junior et al. 2021. Winning solutions and post-challenge analyses of the ChaLearn AutoDL challenge 2019. IEEE Transactions on Pattern Analysis and Machine Intelligence (2021)."},{"key":"e_1_3_2_2_41_1","volume-title":"Multi-view graph neural networks for molecular property prediction. arXiv preprint arXiv:2005.13607","author":"Ma Hehuan","year":"2020","unstructured":"Hehuan Ma , Yatao Bian , Yu Rong , Wenbing Huang , Tingyang Xu , Weiyang Xie , Geyan Ye , and Junzhou Huang . 2020. Multi-view graph neural networks for molecular property prediction. arXiv preprint arXiv:2005.13607 ( 2020 ). Hehuan Ma, Yatao Bian, Yu Rong, Wenbing Huang, Tingyang Xu, Weiyang Xie, Geyan Ye, and Junzhou Huang. 2020. Multi-view graph neural networks for molecular property prediction. arXiv preprint arXiv:2005.13607 (2020)."},{"key":"e_1_3_2_2_42_1","volume-title":"Molecule attention transformer. arXiv preprint arXiv:2002.08264","author":"Maziarka \u0141ukasz","year":"2020","unstructured":"\u0141ukasz Maziarka , Tomasz Danel , S\u0142awomir Mucha , Krzysztof Rataj , Jacek Tabor , and Stanis\u0142aw Jastrzebski . 2020. Molecule attention transformer. arXiv preprint arXiv:2002.08264 ( 2020 ). \u0141ukasz Maziarka, Tomasz Danel, S\u0142awomir Mucha, Krzysztof Rataj, Jacek Tabor, and Stanis\u0142aw Jastrzebski. 2020. Molecule attention transformer. arXiv preprint arXiv:2002.08264 (2020)."},{"key":"e_1_3_2_2_43_1","volume-title":"Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781","author":"Mikolov Tomas","year":"2013","unstructured":"Tomas Mikolov , Kai Chen , Greg Corrado , and Jeffrey Dean . 2013. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 ( 2013 ). Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013)."},{"key":"e_1_3_2_2_44_1","volume-title":"Image Transformer. In Proceedings of the 35th International Conference on Machine Learning (Proceedings of Machine Learning Research","volume":"4064","author":"Parmar Niki","year":"2018","unstructured":"Niki Parmar , Ashish Vaswani , Jakob Uszkoreit , Lukasz Kaiser , Noam Shazeer , Alexander Ku , and Dustin Tran . 2018 . Image Transformer. In Proceedings of the 35th International Conference on Machine Learning (Proceedings of Machine Learning Research , Vol. 80), Jennifer Dy and Andreas Krause (Eds.). PMLR, 4055-- 4064 . http:\/\/proceedings.mlr.press\/v80\/parmar18a.html Niki Parmar, Ashish Vaswani, Jakob Uszkoreit, Lukasz Kaiser, Noam Shazeer, Alexander Ku, and Dustin Tran. 2018. Image Transformer. In Proceedings of the 35th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 80), Jennifer Dy and Andreas Krause (Eds.). PMLR, 4055--4064. http:\/\/proceedings.mlr.press\/v80\/parmar18a.html"},{"key":"e_1_3_2_2_45_1","volume-title":"Transfer learning enables the molecular transformer to predict regio-and stereoselective reactions on carbohydrates. Nature communications","author":"Pesciullesi Giorgio","year":"2020","unstructured":"Giorgio Pesciullesi , Philippe Schwaller , Teodoro Laino , and Jean-Louis Reymond . 2020. Transfer learning enables the molecular transformer to predict regio-and stereoselective reactions on carbohydrates. Nature communications , Vol. 11 , 1 ( 2020 ), 1--8. Giorgio Pesciullesi, Philippe Schwaller, Teodoro Laino, and Jean-Louis Reymond. 2020. Transfer learning enables the molecular transformer to predict regio-and stereoselective reactions on carbohydrates. Nature communications, Vol. 11, 1 (2020), 1--8."},{"key":"e_1_3_2_2_46_1","first-page":"14501","article-title":"Recipe for a general, powerful, scalable graph transformer","volume":"35","author":"Ladislav Ramp\u00e1vs","year":"2022","unstructured":"Ladislav Ramp\u00e1vs ek, Michael Galkin , Vijay Prakash Dwivedi , Anh Tuan Luu , Guy Wolf , and Dominique Beaini . 2022 . Recipe for a general, powerful, scalable graph transformer . Advances in Neural Information Processing Systems , Vol. 35 (2022), 14501 -- 14515 . Ladislav Ramp\u00e1vs ek, Michael Galkin, Vijay Prakash Dwivedi, Anh Tuan Luu, Guy Wolf, and Dominique Beaini. 2022. Recipe for a general, powerful, scalable graph transformer. Advances in Neural Information Processing Systems, Vol. 35 (2022), 14501--14515.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_47_1","volume-title":"Deep learning for the life sciences: applying deep learning to genomics, microscopy, drug discovery, and more","author":"Ramsundar Bharath","unstructured":"Bharath Ramsundar , Peter Eastman , Patrick Walters , and Vijay Pande . 2019. Deep learning for the life sciences: applying deep learning to genomics, microscopy, drug discovery, and more . O'Reilly Media . Bharath Ramsundar, Peter Eastman, Patrick Walters, and Vijay Pande. 2019. Deep learning for the life sciences: applying deep learning to genomics, microscopy, drug discovery, and more. O'Reilly Media."},{"key":"e_1_3_2_2_48_1","first-page":"9689","article-title":"Evaluating protein transfer learning with tape","volume":"32","author":"Rao Roshan","year":"2019","unstructured":"Roshan Rao , Nicholas Bhattacharya , Neil Thomas , Yan Duan , Xi Chen , John Canny , Pieter Abbeel , and Yun S Song . 2019 . Evaluating protein transfer learning with tape . Advances in Neural Information Processing Systems , Vol. 32 (2019), 9689 . Roshan Rao, Nicholas Bhattacharya, Neil Thomas, Yan Duan, Xi Chen, John Canny, Pieter Abbeel, and Yun S Song. 2019. Evaluating protein transfer learning with tape. Advances in Neural Information Processing Systems, Vol. 32 (2019), 9689.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_49_1","volume-title":"Garnett (Eds.)","volume":"32","author":"Ren Yi","year":"2019","unstructured":"Yi Ren , Yangjun Ruan , Xu Tan , Tao Qin , Sheng Zhao , Zhou Zhao , and Tie-Yan Liu . 2019 . FastSpeech: Fast, Robust and Controllable Text to Speech. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextquotesingle Alch\u00e9-Buc, E. Fox, and R . Garnett (Eds.) , Vol. 32 . https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/f63f65b503e22cb970527f23c9ad7db1-Paper.pdf Yi Ren, Yangjun Ruan, Xu Tan, Tao Qin, Sheng Zhao, Zhou Zhao, and Tie-Yan Liu. 2019. FastSpeech: Fast, Robust and Controllable Text to Speech. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextquotesingle Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32. https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/f63f65b503e22cb970527f23c9ad7db1-Paper.pdf"},{"key":"e_1_3_2_2_50_1","doi-asserted-by":"crossref","unstructured":"Alexander Rives Joshua Meier Tom Sercu Siddharth Goyal Zeming Lin Jason Liu Demi Guo Myle Ott C Lawrence Zitnick Jerry Ma etal 2020. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. bioRxiv (2020) 622803.  Alexander Rives Joshua Meier Tom Sercu Siddharth Goyal Zeming Lin Jason Liu Demi Guo Myle Ott C Lawrence Zitnick Jerry Ma et al. 2020. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. bioRxiv (2020) 622803.","DOI":"10.1101\/622803"},{"key":"e_1_3_2_2_51_1","first-page":"12559","article-title":"Self-supervised graph transformer on large-scale molecular data","volume":"33","author":"Rong Yu","year":"2020","unstructured":"Yu Rong , Yatao Bian , Tingyang Xu , Weiyang Xie , Ying Wei , Wenbing Huang , and Junzhou Huang . 2020 a. Self-supervised graph transformer on large-scale molecular data . Advances in Neural Information Processing Systems , Vol. 33 (2020), 12559 -- 12571 . Yu Rong, Yatao Bian, Tingyang Xu, Weiyang Xie, Ying Wei, Wenbing Huang, and Junzhou Huang. 2020a. Self-supervised graph transformer on large-scale molecular data. Advances in Neural Information Processing Systems, Vol. 33 (2020), 12559--12571.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_52_1","first-page":"12559","article-title":"Self-Supervised Graph Transformer on Large-Scale Molecular Data","volume":"33","author":"Rong Yu","year":"2020","unstructured":"Yu Rong , Yatao Bian , Tingyang Xu , Weiyang Xie , Ying WEI , Wenbing Huang , and Junzhou Huang . 2020 b. Self-Supervised Graph Transformer on Large-Scale Molecular Data . In NeurIPS , Vol. 33. 12559 -- 12571 . Yu Rong, Yatao Bian, Tingyang Xu, Weiyang Xie, Ying WEI, Wenbing Huang, and Junzhou Huang. 2020b. Self-Supervised Graph Transformer on Large-Scale Molecular Data. In NeurIPS, Vol. 33. 12559--12571.","journal-title":"NeurIPS"},{"key":"e_1_3_2_2_53_1","volume-title":"Molecular transformer: a model for uncertainty-calibrated chemical reaction prediction. ACS central science","author":"Schwaller Philippe","year":"2019","unstructured":"Philippe Schwaller , Teodoro Laino , Th\u00e9ophile Gaudin , Peter Bolgar , Christopher A Hunter , Costas Bekas , and Alpha A Lee . 2019. Molecular transformer: a model for uncertainty-calibrated chemical reaction prediction. ACS central science , Vol. 5 , 9 ( 2019 ), 1572--1583. Philippe Schwaller, Teodoro Laino, Th\u00e9ophile Gaudin, Peter Bolgar, Christopher A Hunter, Costas Bekas, and Alpha A Lee. 2019. Molecular transformer: a model for uncertainty-calibrated chemical reaction prediction. ACS central science, Vol. 5, 9 (2019), 1572--1583."},{"key":"e_1_3_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i1.16131"},{"key":"e_1_3_2_2_55_1","volume-title":"MoLGNN: Self-Supervised Motif Learning Graph Neural Network for Drug Discovery. Machine Learning for Molecules Workshop at NeurIPS 2020 (2020","author":"Shen Xiaoke","year":"2020","unstructured":"Xiaoke Shen , Yang Liu , You Wu , and Lei Xie . 2020 . MoLGNN: Self-Supervised Motif Learning Graph Neural Network for Drug Discovery. Machine Learning for Molecules Workshop at NeurIPS 2020 (2020 ). https:\/\/ml4molecules.github.io\/papers2020\/ML4Molecules_2020_paper_4.pdf Xiaoke Shen, Yang Liu, You Wu, and Lei Xie. 2020. MoLGNN: Self-Supervised Motif Learning Graph Neural Network for Drug Discovery. Machine Learning for Molecules Workshop at NeurIPS 2020 (2020). https:\/\/ml4molecules.github.io\/papers2020\/ML4Molecules_2020_paper_4.pdf"},{"key":"e_1_3_2_2_56_1","volume-title":"International Conference on Machine Learning. PMLR, 8818--8827","author":"Shi Chence","year":"2020","unstructured":"Chence Shi , Minkai Xu , Hongyu Guo , Ming Zhang , and Jian Tang . 2020 . A graph to graphs framework for retrosynthesis prediction . In International Conference on Machine Learning. PMLR, 8818--8827 . Chence Shi, Minkai Xu, Hongyu Guo, Ming Zhang, and Jian Tang. 2020. A graph to graphs framework for retrosynthesis prediction. In International Conference on Machine Learning. PMLR, 8818--8827."},{"key":"e_1_3_2_2_57_1","volume-title":"3D Infomax improves GNNs for Molecular Property Prediction. arXiv preprint arXiv:2110.04126","author":"St\u00e4rk Hannes","year":"2021","unstructured":"Hannes St\u00e4rk , Dominique Beaini , Gabriele Corso , Prudencio Tossou , Christian Dallago , Stephan G\u00fcnnemann , and Pietro Li\u00f2 . 2021. 3D Infomax improves GNNs for Molecular Property Prediction. arXiv preprint arXiv:2110.04126 ( 2021 ). Hannes St\u00e4rk, Dominique Beaini, Gabriele Corso, Prudencio Tossou, Christian Dallago, Stephan G\u00fcnnemann, and Pietro Li\u00f2. 2021. 3D Infomax improves GNNs for Molecular Property Prediction. arXiv preprint arXiv:2110.04126 (2021)."},{"key":"e_1_3_2_2_58_1","first-page":"10186","article-title":"Towards understanding retrosynthesis by energy-based models","volume":"34","author":"Sun Ruoxi","year":"2021","unstructured":"Ruoxi Sun , Hanjun Dai , Li Li , Steven Kearnes , and Bo Dai . 2021 . Towards understanding retrosynthesis by energy-based models . Advances in Neural Information Processing Systems , Vol. 34 (2021), 10186 -- 10194 . Ruoxi Sun, Hanjun Dai, Li Li, Steven Kearnes, and Bo Dai. 2021. Towards understanding retrosynthesis by energy-based models. Advances in Neural Information Processing Systems, Vol. 34 (2021), 10186--10194.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_59_1","volume-title":"Ruud Van Deursen, and Guillaume Godin","author":"Tetko Igor V.","year":"2020","unstructured":"Igor V. Tetko , Pavel Karpov , Ruud Van Deursen, and Guillaume Godin . 2020 . State-of-the-art augmented NLP transformer models for direct and single-step retrosynthesis. Nature Communications , Vol. 11 , 1 (04 Nov 2020), 5575. https:\/\/doi.org\/10.1038\/s41467-020-19266-y 10.1038\/s41467-020-19266-y Igor V. Tetko, Pavel Karpov, Ruud Van Deursen, and Guillaume Godin. 2020. State-of-the-art augmented NLP transformer models for direct and single-step retrosynthesis. Nature Communications, Vol. 11, 1 (04 Nov 2020), 5575. https:\/\/doi.org\/10.1038\/s41467-020-19266-y"},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.2c00321"},{"key":"e_1_3_2_2_61_1","first-page":"2579","article-title":"Visualizing Data using t-SNE","volume":"9","author":"van der Maaten Laurens","year":"2008","unstructured":"Laurens van der Maaten and Geoffrey Hinton . 2008 . Visualizing Data using t-SNE . Journal of Machine Learning Research , Vol. 9 (2008), 2579 -- 2605 . Laurens van der Maaten and Geoffrey Hinton. 2008. Visualizing Data using t-SNE. Journal of Machine Learning Research, Vol. 9 (2008), 2579--2605.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_2_62_1","volume-title":"Attention is all you need. arXiv preprint arXiv:1706.03762","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani , Noam Shazeer , Niki Parmar , Jakob Uszkoreit , Llion Jones , Aidan N Gomez , Lukasz Kaiser , and Illia Polosukhin . 2017. Attention is all you need. arXiv preprint arXiv:1706.03762 ( 2017 ). Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. arXiv preprint arXiv:1706.03762 (2017)."},{"key":"e_1_3_2_2_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3307339.3342186"},{"key":"e_1_3_2_2_64_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-022-00447-x"},{"key":"e_1_3_2_2_65_1","doi-asserted-by":"publisher","DOI":"10.1021\/ci00057a005"},{"key":"e_1_3_2_2_66_1","volume-title":"MoleculeNet: a benchmark for molecular machine learning. Chemical science","author":"Wu Zhenqin","year":"2018","unstructured":"Zhenqin Wu , Bharath Ramsundar , Evan N Feinberg , Joseph Gomes , Caleb Geniesse , Aneesh S Pappu , Karl Leswing , and Vijay Pande . 2018. MoleculeNet: a benchmark for molecular machine learning. Chemical science , Vol. 9 , 2 ( 2018 ), 513--530. Zhenqin Wu, Bharath Ramsundar, Evan N Feinberg, Joseph Gomes, Caleb Geniesse, Aneesh S Pappu, Karl Leswing, and Vijay Pande. 2018. MoleculeNet: a benchmark for molecular machine learning. Chemical science, Vol. 9, 2 (2018), 513--530."},{"key":"e_1_3_2_2_67_1","volume-title":"Lin (Eds.)","volume":"33","author":"Yan Chaochao","year":"2020","unstructured":"Chaochao Yan , Qianggang Ding , Peilin Zhao , Shuangjia Zheng , JINYU YANG , Yang Yu , and Junzhou Huang . 2020 . RetroXpert: Decompose Retrosynthesis Prediction Like A Chemist. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H . Lin (Eds.) , Vol. 33 . Curran Associates, Inc., 11248--11258. https:\/\/proceedings.neurips.cc\/paper\/ 2020\/file\/819f46e52c25763a55cc642422644317-Paper.pdf Chaochao Yan, Qianggang Ding, Peilin Zhao, Shuangjia Zheng, JINYU YANG, Yang Yu, and Junzhou Huang. 2020. RetroXpert: Decompose Retrosynthesis Prediction Like A Chemist. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 11248--11258. https:\/\/proceedings.neurips.cc\/paper\/2020\/file\/819f46e52c25763a55cc642422644317-Paper.pdf"},{"key":"e_1_3_2_2_68_1","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.9b00237"},{"key":"e_1_3_2_2_69_1","unstructured":"Yuning You Tianlong Chen Yongduo Sui Ting Chen Zhangyang Wang and Yang Shen. 2020. Graph Contrastive Learning with Augmentations. In NeurIPS.  Yuning You Tianlong Chen Yongduo Sui Ting Chen Zhangyang Wang and Yang Shen. 2020. Graph Contrastive Learning with Augmentations. In NeurIPS."},{"key":"e_1_3_2_2_70_1","doi-asserted-by":"publisher","DOI":"10.1039\/D2SC02763A"},{"key":"e_1_3_2_2_71_1","volume-title":"International Conference on Learning Representations. https:\/\/arxiv.org\/abs\/2002","author":"Zhu Jinhua","year":"2020","unstructured":"Jinhua Zhu , Yingce Xia , Lijun Wu , Di He , Tao Qin , Wengang Zhou , Houqiang Li , and Tie-Yan Liu . 2020 . Incorporating BERT into Neural Machine Translation . In International Conference on Learning Representations. https:\/\/arxiv.org\/abs\/2002 .06823 Jinhua Zhu, Yingce Xia, Lijun Wu, Di He, Tao Qin, Wengang Zhou, Houqiang Li, and Tie-Yan Liu. 2020. Incorporating BERT into Neural Machine Translation. In International Conference on Learning Representations. https:\/\/arxiv.org\/abs\/2002.06823"}],"event":{"name":"KDD '23: The 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Long Beach CA USA","acronym":"KDD '23","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3580305.3599317","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3580305.3599317","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:47Z","timestamp":1750178267000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3580305.3599317"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,4]]},"references-count":72,"alternative-id":["10.1145\/3580305.3599317","10.1145\/3580305"],"URL":"https:\/\/doi.org\/10.1145\/3580305.3599317","relation":{},"subject":[],"published":{"date-parts":[[2023,8,4]]},"assertion":[{"value":"2023-08-04","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}