{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,7]],"date-time":"2026-05-07T18:14:00Z","timestamp":1778177640626,"version":"3.51.4"},"reference-count":195,"publisher":"American Chemical Society (ACS)","issue":"11","license":[{"start":{"date-parts":[[2024,5,30]],"date-time":"2024-05-30T00:00:00Z","timestamp":1717027200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000078","name":"Division of Materials Research","doi-asserted-by":"publisher","award":["DMR-1933487"],"award-info":[{"award-number":["DMR-1933487"]}],"id":[{"id":"10.13039\/100000078","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["J. Chem. Inf. Model."],"published-print":{"date-parts":[[2024,6,10]]},"DOI":"10.1021\/acs.jcim.3c02070","type":"journal-article","created":{"date-parts":[[2024,5,30]],"date-time":"2024-05-30T20:27:39Z","timestamp":1717100859000},"page":"4392-4409","source":"Crossref","is-referenced-by-count":56,"title":["Application of Transformers in Cheminformatics"],"prefix":"10.1021","volume":"64","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-6919-4528","authenticated-orcid":true,"given":"Kha-Dinh","family":"Luong","sequence":"first","affiliation":[{"name":"Department of Computer Science, University of California Santa Barbara, Santa Barbara, CA 93106, United States"}]},{"given":"Ambuj","family":"Singh","sequence":"additional","affiliation":[{"name":"Department of Computer Science, University of California Santa Barbara, Santa Barbara, CA 93106, United States"}]}],"member":"316","published-online":{"date-parts":[[2024,5,30]]},"reference":[{"key":"ref1\/cit1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ddtec.2020.05.001"},{"key":"ref2\/cit2","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3084827"},{"key":"ref3\/cit3","doi-asserted-by":"publisher","DOI":"10.1186\/s40649-019-0069-y"},{"key":"ref4\/cit4","unstructured":"Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, L.; Polosukhin, I. Attention is All You Need. Proceedings of the 31st International Conference on Neural Information Processing Systems; Red Hook, NY, 2017; pp 6000\u20136010."},{"key":"ref5\/cit5","doi-asserted-by":"publisher","unstructured":"Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M. Transformers: State-of-the-art natural language processing. Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations; Association for Computational Linguistics, 2020; pp 38\u201345, 10.18653\/v1\/2020.emnlp-demos.6.","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"ref6\/cit6","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3152247"},{"key":"ref7\/cit7","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-2074"},{"key":"ref8\/cit8","unstructured":"Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.  Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Vol. 1 (Long and Short Papers); Minneapolis, MN, 2019; pp 4171\u20134186."},{"key":"ref9\/cit9","unstructured":"Lan, Z.; Chen, M.; Goodman, S.; Gimpel, K.; Sharma, P.; Soricut, R. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations. 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26\u201330, 2020."},{"key":"ref10\/cit10","volume-title":"arXiv:1907.11692","author":"Liu Y.","year":"2019"},{"key":"ref11\/cit11","volume-title":"arXiv:2302.13971","author":"Touvron H.","year":"2023"},{"key":"ref12\/cit12","unstructured":"Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J. D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A. Language models are few-shot learners. In  Advances in neural information processing systems; NeurIPS, 2020; Vol. 33, pp 1877\u20131901."},{"key":"ref13\/cit13","doi-asserted-by":"publisher","DOI":"10.1145\/3505244"},{"key":"ref14\/cit14","volume-title":"An Image is Worth 16 \u00d7 16 Words: Transformers for Image Recognition at Scale. International Conference on Learning Representations","author":"Dosovitskiy A.","year":"2021"},{"key":"ref15\/cit15","unstructured":"El-Nouby, A.; Touvron, H.; Caron, M.; Bojanowski, P.; Douze, M.; Joulin, A.; Laptev, I.; Neverova, N.; Synnaeve, G.; Verbeek, J.; Jegou, H. XCiT: Cross-Covariance Image Transformers. In  Advances in Neural Information Processing Systems; NeurIPS, 2021."},{"key":"ref16\/cit16","volume-title":"Rethinking Spatial Dimensions of Vision Transformers. International Conference on Computer Vision","author":"Heo B.","year":"2021"},{"key":"ref17\/cit17","unstructured":"Jiang, Z.H.; Hou, Q.; Yuan, L.; Zhou, D.; Shi, Y.; Jin, X.; Wang, A.; Feng, J. All Tokens Matter: Token Labeling for Training Better Vision Transformers. In  Advances in Neural Information Processing Systems; NeurIPS, 2021; pp 18590\u201318602."},{"key":"ref18\/cit18","unstructured":"Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jegou, H. Training data-efficient image transformers & distillation through attention. In  International Conference on Machine Learning; Proceedings of the 38th International Conference on Machine Learning, 2021; pp 10347\u201310357."},{"key":"ref187\/cit187","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btg1009"},{"key":"ref188\/cit188","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-022-29241-4"},{"key":"ref19\/cit19","doi-asserted-by":"crossref","unstructured":"Rao, R.; Bhattacharya, N.; Thomas, N.; Duan, Y.; Chen, X.; Canny, J.; Abbeel, P.; Song, Y. S. Evaluating Protein Transfer Learning with TAPE.  Proceedings of the 33rd International Conference on Neural Information Processing Systems, Red Hook, NY, USA, NeurIPS, 2019.","DOI":"10.1101\/676825"},{"key":"ref20\/cit20","doi-asserted-by":"crossref","unstructured":"Rives, A.; Meier, J.; Sercu, T.; Goyal, S.; Lin, Z.; Liu, J.; Guo, D.; Ott, M.; Zitnick, C. L.; Ma, J. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences.  Proceedings of the National Academy of Sciences; 2021, 118, e2016239118.","DOI":"10.1073\/pnas.2016239118"},{"key":"ref21\/cit21","doi-asserted-by":"publisher","DOI":"10.1093\/nar\/gkaa913"},{"key":"ref22\/cit22","doi-asserted-by":"crossref","unstructured":"Nambiar, A.; Heflin, M.; Liu, S.; Maslov, S.; Hopkins, M.; Ritz, A. Transforming the language of life: transformer neural networks for protein prediction tasks.  Proceedings of the 11th ACM international conference on bioinformatics, computational biology and health informatics; ACM, 2020; pp 1\u20138.","DOI":"10.1145\/3388440.3412467"},{"key":"ref23\/cit23","author":"Beltagy I.","year":"2020","journal-title":"arXiv:2004.05150"},{"key":"ref24\/cit24","author":"Filipavicius M.","year":"2020","journal-title":"arXiv:2012.03084"},{"key":"ref25\/cit25","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btac020"},{"key":"ref26\/cit26","doi-asserted-by":"publisher","DOI":"10.1101\/2020.07.12.199554"},{"key":"ref27\/cit27","first-page":"5485","volume":"21","author":"Raffel C.","year":"2020","journal-title":"Journal of Machine Learning Research"},{"key":"ref28\/cit28","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-022-00591-x"},{"key":"ref29\/cit29","unstructured":"Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R. R.; Le, Q. V. Xlnet: Generalized autoregressive pretraining for language understanding.  Advances in neural information processing systems; NeurIPS, 2019; Vol. 32."},{"key":"ref30\/cit30","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btac474"},{"key":"ref31\/cit31","doi-asserted-by":"crossref","unstructured":"Rao, R. M.; Liu, J.; Verkuil, R.; Meier, J.; Canny, J.; Abbeel, P.; Sercu, T.; Rives, A. MSA Transformer.  Proceedings of the 38th International Conference on Machine Learning; ICML, 2021; pp 8844\u20138856.","DOI":"10.1101\/2021.02.12.430858"},{"key":"ref32\/cit32","doi-asserted-by":"publisher","DOI":"10.1021\/acsbiomaterials.2c00737"},{"key":"ref33\/cit33","doi-asserted-by":"crossref","unstructured":"Meier, J.; Rao, R.; Verkuil, R.; Liu, J.; Sercu, T.; Rives, A. Language models enable zero-shot prediction of the effects of mutations on protein function. In  Advances in Neural Information Processing Systems, NeurIPS, 2021; Vol. 34, pp 29287\u201329303.","DOI":"10.1101\/2021.07.09.450648"},{"key":"ref34\/cit34","doi-asserted-by":"publisher","DOI":"10.1016\/j.xinn.2023.100487"},{"key":"ref35\/cit35","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btac053"},{"key":"ref36\/cit36","doi-asserted-by":"publisher","DOI":"10.1038\/s41587-022-01432-w"},{"key":"ref37\/cit37","doi-asserted-by":"publisher","DOI":"10.1038\/s43588-022-00373-3"},{"key":"ref38\/cit38","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-021-03819-2"},{"key":"ref39\/cit39","doi-asserted-by":"publisher","DOI":"10.1126\/science.abj8754"},{"key":"ref40\/cit40","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btab083"},{"key":"ref41\/cit41","doi-asserted-by":"publisher","DOI":"10.1038\/s41592-021-01252-x"},{"key":"ref42\/cit42","doi-asserted-by":"crossref","unstructured":"An, W.; Guo, Y.; Bian, Y.; Ma, H.; Yang, J.; Li, C.; Huang, J. MoDNA: motif-oriented pre-training for DNA language model.  Proceedings of the 13th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics; ACM, 2022; pp 1\u20135.","DOI":"10.1145\/3535508.3545512"},{"key":"ref43\/cit43","doi-asserted-by":"publisher","DOI":"10.1101\/2023.01.11.523679"},{"key":"ref44\/cit44","author":"Zhou Z.","year":"2023","journal-title":"arXiv:2306.15006"},{"key":"ref45\/cit45","doi-asserted-by":"publisher","DOI":"10.1016\/j.patter.2022.100588"},{"key":"ref46\/cit46","unstructured":"McNaught, A. D.; Wilkinson, A. Compendium of chemical terminology; Blackwell Science: Oxford, 1997; Vol. 1669."},{"key":"ref47\/cit47","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-015-0068-4"},{"key":"ref48\/cit48","doi-asserted-by":"publisher","DOI":"10.1021\/ci00057a005"},{"key":"ref49\/cit49","doi-asserted-by":"publisher","DOI":"10.26434\/chemrxiv.7097960.v1"},{"key":"ref50\/cit50","doi-asserted-by":"publisher","DOI":"10.1021\/acscentsci.9b00476"},{"key":"ref51\/cit51","doi-asserted-by":"publisher","DOI":"10.1088\/2632-2153\/aba947"},{"key":"ref52\/cit52","author":"Honda S.","year":"2019","journal-title":"arXiv:1911.04738"},{"key":"ref53\/cit53","doi-asserted-by":"crossref","unstructured":"Wang, S.; Guo, Y.; Wang, Y.; Sun, H.; Huang, J. Smiles-bert: large scale unsupervised pre-training for molecular property prediction.  Proceedings of the 10th ACM international conference on bioinformatics, computational biology and health informatics; ACM, 2019; pp 429\u2013436.","DOI":"10.1145\/3307339.3342186"},{"key":"ref54\/cit54","doi-asserted-by":"publisher","DOI":"10.1155\/2021\/4359369"},{"key":"ref55\/cit55","doi-asserted-by":"publisher","DOI":"10.1093\/bib\/bbac131"},{"key":"ref56\/cit56","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-023-39868-6"},{"key":"ref57\/cit57","doi-asserted-by":"publisher","DOI":"10.1038\/s41524-023-01016-5"},{"key":"ref58\/cit58","author":"Chithrananda S.","year":"2020","journal-title":"arXiv:2010.09885"},{"key":"ref59\/cit59","volume-title":"arXiv:2209.01712","author":"Ahmad W.","year":"2022"},{"key":"ref60\/cit60","doi-asserted-by":"publisher","DOI":"10.1088\/2632-2153\/ac3ffb"},{"key":"ref61\/cit61","doi-asserted-by":"publisher","DOI":"10.1016\/j.scib.2022.01.029"},{"key":"ref62\/cit62","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-022-00580-7"},{"key":"ref63\/cit63","doi-asserted-by":"publisher","DOI":"10.34133\/research.0004"},{"key":"ref64\/cit64","doi-asserted-by":"publisher","DOI":"10.1016\/j.jmgm.2022.108344"},{"key":"ref65\/cit65","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-022-00650-3"},{"key":"ref66\/cit66","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.7b00616"},{"key":"ref67\/cit67","author":"Mikolov T.","year":"2013","journal-title":"arXiv:1301.3781"},{"key":"ref68\/cit68","doi-asserted-by":"publisher","DOI":"10.1016\/j.ddtec.2020.11.009"},{"key":"ref69\/cit69","author":"M\u00fcller L.","year":"2023","journal-title":"arXiv:2302.04181"},{"key":"ref70\/cit70","volume-title":"arXiv:2001.05140","author":"Zhang J.","year":"2020"},{"key":"ref71\/cit71","unstructured":"Page, L. The pagerank citation ranking: Bringing order to the web. Stanford Digital Library Technologies Project, 1998."},{"key":"ref72\/cit72","doi-asserted-by":"crossref","unstructured":"Huang, N. T.; Villar, S. A short tutorial on the weisfeiler-lehman test and its variants.  ICASSP 2021\u20132021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); IEEE, 2021; pp 8533\u20138537.","DOI":"10.1109\/ICASSP39728.2021.9413523"},{"key":"ref73\/cit73","author":"Dwivedi V. P.","year":"2020","journal-title":"arXiv:2012.09699"},{"key":"ref74\/cit74","unstructured":"Kreuzer, D.; Beaini, D.; Hamilton, W.; L\u00e9tourneau, V.; Tossou, P. Rethinking graph transformers with spectral attention.  Advances in Neural Information Processing Systems; NeurIPS, 2021; Vol. 34, pp 21618\u201321629."},{"key":"ref75\/cit75","unstructured":"He, X.; Hooi, B.; Laurent, T.; Perold, A.; LeCun, Y.; Bresson, X. A generalization of vit\/mlp-mixer to graphs. International Conference on Machine Learning. ICML, 2023; pp 12724\u201312745."},{"key":"ref76\/cit76","unstructured":"Kim, J.; Nguyen, D.; Min, S.; Cho, S.; Lee, M.; Lee, H.; Hong, S. Pure transformers are powerful graph learners.  Advances in Neural Information Processing Systems; NeurIPS, 2022; Vol. 35, pp 14582\u201314595."},{"key":"ref77\/cit77","unstructured":"Chen, J.; Gao, K.; Li, G.; He, K. NAGphormer: A tokenized graph transformer for node classification in large graphs.  The Eleventh International Conference on Learning Representations, 2022."},{"key":"ref78\/cit78","unstructured":"Baek, J.; Kang, M.; Hwang, S. J. Accurate Learning of Graph Representations with Graph Multiset Pooling.  International Conference on Learning Representations, 2020."},{"key":"ref79\/cit79","unstructured":"Shirzad, H.; Velingker, A.; Venkatachalam, B.; Sutherland, D. J.; Sinop, A. K. Exphormer: Sparse transformers for graphs.  International Conference on Machine Learning, 2023."},{"key":"ref80\/cit80","unstructured":"Chen, D.; O\u2019Bray, L.; Borgwardt, K. Structure-aware transformer for graph representation learning.  International Conference on Machine Learning, 2022; pp 3469\u20133489."},{"key":"ref81\/cit81","unstructured":"Dwivedi, V. P.; Luu, A. T.; Laurent, T.; Bengio, Y.; Bresson, X. Graph Neural Networks with Learnable Structural and Positional Representations.  International Conference on Learning Representations, 2022."},{"key":"ref82\/cit82","unstructured":"Mialon, G.; Chen, D.; Selosse, M.; Mairal, J. Graphit: Encoding graph structure in transformers. arXiv:2106.05667, 2021."},{"key":"ref83\/cit83","doi-asserted-by":"crossref","unstructured":"Chen, B.; Barzilay, R.; Jaakkola, T. Path-augmented graph transformer network. arXiv:1905.12712, 2019.","DOI":"10.26434\/chemrxiv.8214422"},{"key":"ref84\/cit84","author":"Maziarka \u0141.","year":"2020","journal-title":"arXiv:2002.08264"},{"key":"ref85\/cit85","unstructured":"Gasteiger, J.; Gro\u00df, J.; G\u00fcnnemann, S. Directional Message Passing for Molecular Graphs; International Conference on Learning Representations (ICLR), 2020."},{"key":"ref86\/cit86","volume-title":"arXiv:2110.05841","author":"Maziarka \u0141.","year":"2021"},{"key":"ref87\/cit87","unstructured":"Hu, W.; Fey, M.; Zitnik, M.; Dong, Y.; Ren, H.; Liu, B.; Catasta, M.; Leskovec, J. Open graph benchmark: Datasets for machine learning on graphs.  Advances in neural information processing systems; NeurIPS, 2020; Vol. 33, pp 22118\u201322133."},{"key":"ref88\/cit88","unstructured":"Ying, C.; Cai, T.; Luo, S.; Zheng, S.; Ke, G.; He, D.; Shen, Y.; Liu, T.Y. Do transformers really perform badly for graph representation? Advances in Neural Information Processing Systems, 2021; Vol. 34, pp 28877\u201328888."},{"key":"ref89\/cit89","unstructured":"Park, W.; Chang, W.G.; Lee, D.; Kim, J.; Hwang, S. GRPE: Relative Positional Encoding for Graph Transformer.  ICLR2022 Machine Learning for Drug Discovery, 2022."},{"key":"ref90\/cit90","doi-asserted-by":"crossref","first-page":"655","DOI":"10.1145\/3534678.3539296","volume-title":"Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","author":"Hussain M. S.","year":"2022"},{"key":"ref91\/cit91","author":"Chen Z.","year":"2023","journal-title":"arXiv:2305.11424"},{"key":"ref92\/cit92","doi-asserted-by":"crossref","unstructured":"Wu, F.; Radev, D.; Li, S. Z. Molformer: Motif-based transformer on 3d heterogeneous molecular graphs.  Proceedings of the AAAI Conference on Artificial Intelligence; AAAI, 2023; pp 5312\u20135320.","DOI":"10.1609\/aaai.v37i4.25662"},{"key":"ref93\/cit93","unstructured":"Ramp\u00e1\u0161ek, L.; Galkin, M.; Dwivedi, V. P.; Luu, A. T.; Wolf, G.; Beaini, D. Recipe for a general, powerful, scalable graph transformer Advances in Neural Information Processing Systems; NeurIPS, 2022; Vol. 35, pp 14501\u201314515."},{"key":"ref94\/cit94","unstructured":"Alon, U.; Yahav, E. On the Bottleneck of Graph Neural Networks and its Practical Implications.  International Conference on Learning Representations, 2020."},{"key":"ref95\/cit95","doi-asserted-by":"publisher","unstructured":"Li, Q.; Han, Z.; Wu, X.M. Deeper insights into graph convolutional networks for semi-supervised learning.  Proceedings of the AAAI conference on artificial intelligence. AAAI, 2018; 1, 1, 10.1609\/aaai.v32i1.11604.","DOI":"10.1609\/aaai.v32i1.11604"},{"key":"ref96\/cit96","doi-asserted-by":"crossref","unstructured":"Srinivas, A.; Lin, T.Y.; Parmar, N.; Shlens, J.; Abbeel, P.; Vaswani, A. Bottleneck Transformers for Visual Recognition.  2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE, 2021; pp 16514\u201316524.","DOI":"10.1109\/CVPR46437.2021.01625"},{"key":"ref97\/cit97","unstructured":"Wu, Z.; Jain, P.; Wright, M.; Mirhoseini, A.; Gonzalez, J. E.; Stoica, I. Representing long-range context for graph neural networks with global attention.  Advances in Neural Information Processing Systems; NeurIPS, 2021; Vol. 34, pp 13266\u201313279."},{"key":"ref98\/cit98","unstructured":"Rong, Y.; Bian, Y.; Xu, T.; Xie, W.; Wei, Y.; Huang, W.; Huang, J. Self-supervised graph transformer on large-scale molecular data.  Advances in Neural Information Processing Systems; NeurIPS, 2020; Vol. 33, pp 12559\u201312571."},{"key":"ref99\/cit99","doi-asserted-by":"crossref","unstructured":"Nguyen, D. Q.; Nguyen, T. D.; Phung, D. Universal Graph Transformer Self-Attention Networks. Companion Proceedings of the Web Conference 2022: New York, NY, USA, 2022; pp 193\u2013196.","DOI":"10.1145\/3487553.3524258"},{"key":"ref100\/cit100","author":"Masters D.","year":"2022","journal-title":"arXiv:2212.02229"},{"key":"ref101\/cit101","doi-asserted-by":"publisher","DOI":"10.3390\/molecules25061292"},{"key":"ref102\/cit102","doi-asserted-by":"publisher","DOI":"10.1039\/C7SC02664A"},{"key":"ref103\/cit103","author":"Zhong Z.","year":"2024","journal-title":"arXiv:2403.05075"},{"key":"ref104\/cit104","doi-asserted-by":"crossref","unstructured":"Zhou, G.; Gao, Z.; Ding, Q.; Zheng, H.; Xu, H.; Wei, Z.; Zhang, L.; Ke, G. Uni-Mol: A Universal 3D Molecular Representation Learning Framework.  The Eleventh International Conference on Learning Representations, 2023.","DOI":"10.26434\/chemrxiv-2022-jjm0j-v4"},{"key":"ref105\/cit105","unstructured":"OGB-LSC @ KDD Cup 2021\n\u2500 ogb.stanford.edu. https:\/\/ogb.stanford.edu\/kddcup2021\/results, 2021 (accessed 01-04-2024)."},{"key":"ref106\/cit106","volume-title":"OGB-LSC Leaderboards \u2500 ogb.stanford.edu"},{"key":"ref107\/cit107","unstructured":"Dwivedi, V. P.; Ramp\u00e1\u0161ek, L.; Galkin, M.; Parviz, A.; Wolf, G.; Luu, A. T.; Beaini, D. Long range graph benchmark.  Advances in Neural Information Processing Systems, NeurIPS, 2022; Vol. 35, pp 22326\u201322340."},{"key":"ref108\/cit108","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s13321-020-00423-w","volume":"12","author":"Karpov P.","year":"2020","journal-title":"Journal of cheminformatics"},{"key":"ref109\/cit109","doi-asserted-by":"publisher","DOI":"10.1021\/acs.chemrestox.3c00032"},{"key":"ref110\/cit110","doi-asserted-by":"publisher","DOI":"10.1021\/acs.molpharmaceut.3c00129"},{"key":"ref111\/cit111","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2022.105772"},{"key":"ref112\/cit112","doi-asserted-by":"publisher","DOI":"10.3390\/pharmaceutics14081710"},{"key":"ref113\/cit113","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btad778"},{"key":"ref114\/cit114","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.122334"},{"key":"ref115\/cit115","doi-asserted-by":"publisher","DOI":"10.1093\/bioinformatics\/btaa880"},{"key":"ref116\/cit116","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-022-00584-w"},{"key":"ref117\/cit117","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiolchem.2023.107982"},{"key":"ref118\/cit118","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2023.11.018"},{"key":"ref119\/cit119","doi-asserted-by":"publisher","DOI":"10.1016\/j.cmpb.2023.108003"},{"key":"ref120\/cit120","doi-asserted-by":"publisher","DOI":"10.1021\/ci0255782"},{"key":"ref121\/cit121","doi-asserted-by":"publisher","DOI":"10.1002\/(SICI)1098-1128(199601)16:1<3::AID-MED1>3.0.CO;2-6"},{"key":"ref122\/cit122","doi-asserted-by":"publisher","DOI":"10.1093\/nar\/gkac956"},{"key":"ref123\/cit123","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.3c01496"},{"key":"ref124\/cit124","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-023-00759-z"},{"key":"ref125\/cit125","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-021-00403-1"},{"key":"ref126\/cit126","doi-asserted-by":"publisher","DOI":"10.1093\/bib\/bbad185"},{"key":"ref127\/cit127","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-023-00639-z"},{"key":"ref128\/cit128","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-020-79682-4"},{"key":"ref129\/cit129","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.3c00536"},{"key":"ref130\/cit130","doi-asserted-by":"crossref","unstructured":"Bao, J.; Duan, N.; Zhou, M.; Zhao, T. Knowledge-based question answering as machine translation.  Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Vol. 1: Long Papers); ACL, 2014; pp 967\u2013976.","DOI":"10.3115\/v1\/P14-1091"},{"key":"ref131\/cit131","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-021-94082-y"},{"key":"ref132\/cit132","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-021-00535-x"},{"key":"ref133\/cit133","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.9b01212"},{"key":"ref134\/cit134","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-022-00624-5"},{"key":"ref135\/cit135","doi-asserted-by":"publisher","DOI":"10.1002\/cmtd.202100069"},{"key":"ref136\/cit136","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-021-00538-8"},{"key":"ref137\/cit137","doi-asserted-by":"publisher","DOI":"10.3390\/biom11121793"},{"key":"ref138\/cit138","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-022-00599-3"},{"key":"ref139\/cit139","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-021-00565-5"},{"key":"ref140\/cit140","doi-asserted-by":"publisher","DOI":"10.1039\/D0SC02639E"},{"key":"ref141\/cit141","doi-asserted-by":"publisher","DOI":"10.1021\/acscentsci.9b00576"},{"key":"ref142\/cit142","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.2c01407"},{"key":"ref143\/cit143","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-020-18671-7"},{"key":"ref144\/cit144","doi-asserted-by":"publisher","DOI":"10.1039\/D2SC06798F"},{"key":"ref145\/cit145","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.2c00321"},{"key":"ref146\/cit146","doi-asserted-by":"crossref","unstructured":"Hu, H.; Jiang, Y.; Yang, Y.; Chen, J. X. Enhanced Template-Free Reaction Prediction with Molecular Graphs and Sequence-based Data Augmentation.  Proceedings of the 32nd ACM International Conference on Information and Knowledge Management; ACM, 2023; pp 813\u2013822.","DOI":"10.1145\/3583780.3614865"},{"key":"ref147\/cit147","doi-asserted-by":"crossref","unstructured":"Karpov, P.; Godin, G.; Tetko, I. V. A transformer model for retrosynthesis.  International Conference on Artificial Neural Networks; European Neural Network Society, 2019; pp 817\u2013830.","DOI":"10.1007\/978-3-030-30493-5_78"},{"key":"ref148\/cit148","doi-asserted-by":"publisher","DOI":"10.1039\/C9SC05704H"},{"key":"ref149\/cit149","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.9b00949"},{"key":"ref150\/cit150","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-020-19266-y"},{"key":"ref151\/cit151","doi-asserted-by":"publisher","DOI":"10.1016\/j.cej.2021.129845"},{"key":"ref152\/cit152","unstructured":"Wan, Y.; Hsieh, C.Y.; Liao, B.; Zhang, S. Retroformer: Pushing the limits of end-to-end retrosynthesis transformer.  International Conference on Machine Learning; ICML, 2022; pp 22475\u201322490."},{"key":"ref153\/cit153","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.1c00600"},{"key":"ref154\/cit154","doi-asserted-by":"publisher","DOI":"10.1186\/s13321-022-00638-z"},{"key":"ref155\/cit155","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.2c00344"},{"key":"ref156\/cit156","doi-asserted-by":"publisher","DOI":"10.1126\/sciadv.abe4166"},{"key":"ref157\/cit157","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-00284-w"},{"key":"ref158\/cit158","first-page":"28","volume":"3","author":"Reynaud E.","year":"2010","journal-title":"Nature Education"},{"key":"ref159\/cit159","doi-asserted-by":"publisher","DOI":"10.1016\/j.bbadis.2011.10.002"},{"key":"ref160\/cit160","doi-asserted-by":"publisher","DOI":"10.1038\/ncb1104-1054"},{"key":"ref161\/cit161","doi-asserted-by":"publisher","DOI":"10.1101\/2024.02.24.581708"},{"key":"ref162\/cit162","author":"Schwing G.","year":"2022","journal-title":"arXiv:2206.04683"},{"key":"ref163\/cit163","author":"Zeng W.","year":"2021","journal-title":"arXiv:2107.06573"},{"key":"ref164\/cit164","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jpcb.3c08195"},{"key":"ref165\/cit165","author":"Thomas N.","year":"2018","journal-title":"arXiv:1802.08219"},{"key":"ref166\/cit166","unstructured":"Fuchs, F.; Worrall, D.; Fischer, V.; Welling, M. Se (3)-transformers: 3d roto-translation equivariant attention networks.  Advances in neural information processing systems, NeurIPS, 2020; Vol. 33, pp 1970\u20131981."},{"key":"ref167\/cit167","unstructured":"Th\u00f6lke, P.; Fabritiis, G. D. Equivariant Transformers for Neural Network based Molecular Potentials.  International Conference on Learning Representations, 2022."},{"key":"ref168\/cit168","unstructured":"Liao, Y.L.; Smidt, T. Equiformer: Equivariant Graph Attention Transformer for 3D Atomistic Graphs.  The Eleventh International Conference on Learning Representations, 2022."},{"key":"ref169\/cit169","doi-asserted-by":"publisher","DOI":"10.1063\/1.4812323"},{"key":"ref170\/cit170","doi-asserted-by":"publisher","DOI":"10.1107\/S2052520616003954"},{"key":"ref171\/cit171","doi-asserted-by":"publisher","DOI":"10.1103\/PhysRevLett.120.145301"},{"key":"ref172\/cit172","unstructured":"Yan, K.; Liu, Y.; Lin, Y.; Ji, S. Periodic graph transformers for crystal material property prediction.  Advances in Neural Information Processing Systems, 2022; Vol. 35, pp 15066\u201315080."},{"key":"ref173\/cit173","doi-asserted-by":"publisher","DOI":"10.1021\/acs.jcim.2c00876"},{"key":"ref174\/cit174","author":"Bai J.","year":"2023","journal-title":"arXiv:2302.01486"},{"key":"ref175\/cit175","volume-title":"International Conference on Learning Representations","author":"Veli\u010dkovi\u0107 P.","year":"2018"},{"key":"ref176\/cit176","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-023-00628-2"},{"key":"ref177\/cit177","doi-asserted-by":"publisher","DOI":"10.1021\/jacs.2c11420"},{"key":"ref178\/cit178","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-022-28494-3"},{"key":"ref179\/cit179","doi-asserted-by":"crossref","unstructured":"Edwards, C.; Lai, T.; Ros, K.; Honke, G.; Cho, K.; Ji, H. Translation between Molecules and Natural Language.  Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing; Abu Dhabi, United Arab Emirates, 2022; pp 375\u2013413.","DOI":"10.18653\/v1\/2022.emnlp-main.26"},{"key":"ref180\/cit180","doi-asserted-by":"crossref","unstructured":"Liu, Z.; Zhang, W.; Xia, Y.; Wu, L.; Xie, S.; Qin, T.; Zhang, M.; Liu, T.Y. MolXPT: Wrapping Molecules with Text for Generative Pre-training. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Vol. 2: Short Papers). Toronto, Canada, 2023; pp 1606\u20131616.","DOI":"10.18653\/v1\/2023.acl-short.138"},{"key":"ref181\/cit181","author":"Li J.","year":"2023","journal-title":"Empowering Molecule Discovery for Molecule-Caption Translation with Large Language Models: A ChatGPT Perspective"},{"key":"ref182\/cit182","doi-asserted-by":"crossref","unstructured":"Edwards, C.; Zhai, C.; Ji, H. Text2Mol: Cross-Modal Molecule Retrieval with Natural Language Queries.  Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing; Online and Punta Cana, Dominican Republic, 2021; pp 595\u2013607.","DOI":"10.18653\/v1\/2021.emnlp-main.47"},{"key":"ref183\/cit183","doi-asserted-by":"crossref","unstructured":"Beltagy, I.; Lo, K.; Cohan, A. SciBERT: A Pretrained Language Model for Scientific Text.  Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing; EMNLP-IJCNLP: Hong Kong, China, 2019; pp 3615\u20133620.","DOI":"10.18653\/v1\/D19-1371"},{"key":"ref184\/cit184","author":"Su B.","year":"2022","journal-title":"arXiv:2209.05481"},{"key":"ref185\/cit185","unstructured":"Seidl, P.; Vall, A.; Hochreiter, S.; Klambauer, G. Enhancing Activity Prediction Models in Drug Discovery with the Ability to Understand Human Language.  Proceedings of the 40th International Conference on Machine Learning (ICML), 2023."},{"key":"ref186\/cit186","author":"Liu P.","year":"2023","journal-title":"Git-mol: A multi-modal large language model for molecular science with graph, image, and text"},{"key":"ref189\/cit189","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-024-02120-8"},{"key":"ref190\/cit190","first-page":"1877","volume":"33","author":"Brown T.","year":"2020","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref191\/cit191","first-page":"59662","volume":"36","author":"Guo T.","year":"2023","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref192\/cit192","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-023-00788-1"},{"key":"ref193\/cit193","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-023-00788-1"},{"key":"ref194\/cit194","doi-asserted-by":"publisher","DOI":"10.1039\/D2DD00087C"},{"key":"ref195\/cit195","doi-asserted-by":"publisher","DOI":"10.1039\/D3DD00113J"}],"container-title":["Journal of Chemical Information and Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/pubs.acs.org\/doi\/pdf\/10.1021\/acs.jcim.3c02070","content-type":"application\/pdf","content-version":"vor","intended-application":"unspecified"},{"URL":"https:\/\/pubs.acs.org\/doi\/pdf\/10.1021\/acs.jcim.3c02070","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,10]],"date-time":"2024-06-10T08:10:09Z","timestamp":1718007009000},"score":1,"resource":{"primary":{"URL":"https:\/\/pubs.acs.org\/doi\/10.1021\/acs.jcim.3c02070"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,30]]},"references-count":195,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2024,6,10]]}},"alternative-id":["10.1021\/acs.jcim.3c02070"],"URL":"https:\/\/doi.org\/10.1021\/acs.jcim.3c02070","relation":{},"ISSN":["1549-9596","1549-960X"],"issn-type":[{"value":"1549-9596","type":"print"},{"value":"1549-960X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,30]]}}}