{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,25]],"date-time":"2026-02-25T17:13:37Z","timestamp":1772039617866,"version":"3.50.1"},"reference-count":91,"publisher":"Association for Computing Machinery (ACM)","issue":"6","funder":[{"name":"Natural Science Foundation of Xiamen, China","award":["3502Z202471028"],"award-info":[{"award-number":["3502Z202471028"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62002303 and 42171456"],"award-info":[{"award-number":["62002303 and 42171456"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"CCF-Tencent Open Fund","award":["RAGR20210129"],"award-info":[{"award-number":["RAGR20210129"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Knowl. Discov. Data"],"published-print":{"date-parts":[[2025,7,31]]},"abstract":"<jats:p>\n            Software development is a repetitive task, as developers usually reuse or get inspiration from existing implementations. Code search, which refers to the retrieval of relevant code snippets from a codebase according to the developer\u2019s intent that has been expressed as a query, has become increasingly important in the software development process. Due to the success of deep learning in various applications, a great number of deep learning-based code search approaches have sprung up and achieved promising results. However, developers may not follow the same naming conventions and the same variable may have different variable names in different implementations, bringing a challenge to deep learning-based code search methods that rely on explicit variable correspondences to understand source code. To overcome this challenge, we propose a Naming-Agnostic Code Search (NACS) method based on contrastive multi-view code representation learning. NACS strips information bound to variable names from Abstract Syntax Tree (AST), the representation of the abstract syntactic structure of source code, and focuses on capturing intrinsic properties solely from AST structures. We use semantic-level and syntax-level augmentation techniques to prepare realistically rational data and adopt contrastive learning to design a graph-view modeling component in NACS to enhance the understanding of code snippets. We further model ASTs in a path view to strengthen the graph-view modeling component through multi-view learning. Extensive experiments show that NACS provides superior code search performance compared to baselines and NACS can be adapted to help existing code search methods overcome the impact of different naming conventions. Our implementation is available at\n            <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" ext-link-type=\"uri\" xlink:href=\"https:\/\/github.com\/KDEGroup\/NACS\">https:\/\/github.com\/KDEGroup\/NACS<\/jats:ext-link>\n            .\n          <\/jats:p>","DOI":"10.1145\/3737878","type":"journal-article","created":{"date-parts":[[2025,5,30]],"date-time":"2025-05-30T11:41:36Z","timestamp":1748605296000},"page":"1-26","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Deep Code Search with Naming-Agnostic Contrastive Multi-View Learning"],"prefix":"10.1145","volume":"19","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-7285-9033","authenticated-orcid":false,"given":"Jiadong","family":"Feng","sequence":"first","affiliation":[{"name":"Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, Xiamen University, Xiamen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-4928-5477","authenticated-orcid":false,"given":"Wei","family":"Li","sequence":"additional","affiliation":[{"name":"School of Electronic and Computer Engineering, Peking University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-2914-1458","authenticated-orcid":false,"given":"Suhuang","family":"Wu","sequence":"additional","affiliation":[{"name":"Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, Xiamen University, Xiamen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-4462-3153","authenticated-orcid":false,"given":"Zhao","family":"Wei","sequence":"additional","affiliation":[{"name":"Tencent, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-7488-3704","authenticated-orcid":false,"given":"Yong","family":"Xu","sequence":"additional","affiliation":[{"name":"Tencent, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-0864-0082","authenticated-orcid":false,"given":"Juhong","family":"Wang","sequence":"additional","affiliation":[{"name":"Tencent, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9139-3855","authenticated-orcid":false,"given":"Hui","family":"Li","sequence":"additional","affiliation":[{"name":"Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, Xiamen University, Xiamen, China"}]}],"member":"320","published-online":{"date-parts":[[2025,7,9]]},"reference":[{"issue":"4","key":"e_1_3_2_2_2","first-page":"81:1","article-title":"A survey of machine learning for big code and naturalness","volume":"51","author":"Allamanis Miltiadis","year":"2018","unstructured":"Miltiadis Allamanis, Earl T. Barr, Premkumar T. Devanbu, and Charles Sutton. 2018. A survey of machine learning for big code and naturalness. ACM Comput. Surv 51, 4 (2018), 81:1\u201381:37.","journal-title":"ACM Comput. Surv"},{"key":"e_1_3_2_3_2","volume-title":"ICLR","author":"Allamanis Miltiadis","year":"2018","unstructured":"Miltiadis Allamanis, Marc Brockschmidt, and Mahmoud Khademi. 2018. Learning to represent programs with graphs. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=BJOFETxR"},{"key":"e_1_3_2_4_2","volume-title":"ICLR","author":"Alon Uri","year":"2019","unstructured":"Uri Alon, Shaked Brody, Omer Levy, and Eran Yahav. 2019. code2seq: Generating sequences from structured representations of code. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=H1gKYo09tX"},{"issue":"40","key":"e_1_3_2_5_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3290353","article-title":"code2vec: Learning distributed representations of code","volume":"3","author":"Alon Uri","year":"2019","unstructured":"Uri Alon, Meital Zilberstein, Omer Levy, and Eran Yahav. 2019. code2vec: Learning distributed representations of code. Proc. ACM Program. Lang. 3, POPL, 40 (2019), 1\u201329.","journal-title":"Proc. ACM Program. Lang"},{"key":"e_1_3_2_6_2","first-page":"681","article-title":"Sourcerer: A search engine for open source code supporting structure-based search","author":"Bajracharya Sushil Krishna","year":"2006","unstructured":"Sushil Krishna Bajracharya, Trung Chi Ngo, Erik Linstead, Yimeng Dou, Paul Rigor, Pierre Baldi, and Cristina Videira Lopes. 2006. Sourcerer: A search engine for open source code supporting structure-based search. In OOPSLA Companion, 681\u2013682.","journal-title":"OOPSLA Companion"},{"key":"e_1_3_2_7_2","first-page":"157","volume-title":"SIGSOFT FSE","author":"Krishna Bajracharya Sushil","year":"2010","unstructured":"Sushil Krishna Bajracharya, Joel Ossher, and Cristina Videira Lopes. 2010. Leveraging usage similarity for effective retrieval of examples in code repositories. In SIGSOFT FSE, 157\u2013166."},{"key":"e_1_3_2_8_2","first-page":"3589","volume-title":"NeurIPS","author":"Ben-Nun Tal","year":"2018","unstructured":"Tal Ben-Nun, Alice Shoshana Jakobovits, and Torsten Hoefler. 2018. Neural code comprehension: A learnable representation of code semantics. In NeurIPS, 3589\u20133601."},{"key":"e_1_3_2_9_2","first-page":"1589","volume-title":"CHI","author":"Brandt Joel","year":"2009","unstructured":"Joel Brandt, Philip J. Guo, Joel Lewenstein, Mira Dontcheva, and Scott R. Klemmer. 2009. Two studies of opportunistic programming: Interleaving web foraging, learning, and writing code. In CHI, 1589\u20131598."},{"key":"e_1_3_2_10_2","first-page":"107","article-title":"The anatomy of a large-scale hypertextual web search engine","volume":"30","author":"Brin Sergey","year":"1998","unstructured":"Sergey Brin and Lawrence Page. 1998. The anatomy of a large-scale hypertextual web search engine. Comput. Netw. 30, 1\u20137 (1998), 107\u2013117.","journal-title":"Comput. Netw"},{"key":"e_1_3_2_11_2","volume-title":"ICLR","author":"Brockschmidt Marc","year":"2019","unstructured":"Marc Brockschmidt, Miltiadis Allamanis, Alexander L. Gaunt, and Oleksandr Polozov. 2019. Generative code modeling with graphs. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=Bke4KsA5FX"},{"key":"e_1_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462840"},{"key":"e_1_3_2_13_2","unstructured":"Luca Buratti Saurabh Pujar Mihaela A. Bornea J. Scott McCarley Yunhui Zheng Gaetano Rossiello Alessandro Morari Jim Laredo Veronika Thost Yufan Zhuang et al. 2020. Exploring software naturalness through neural language models. arXiv:2006.12641. Retrieved from https:\/\/arxiv.org\/abs\/2006.12641."},{"key":"e_1_3_2_14_2","doi-asserted-by":"crossref","unstructured":"Yitian Chai Hongyu Zhang Beijun Shen and Xiaodong Gu. 2022. Cross-domain deep code search with meta learning. In ICSE 487\u2013498.","DOI":"10.1145\/3510003.3510125"},{"key":"e_1_3_2_15_2","first-page":"533","volume-title":"ICPC","author":"Cheng Yi","year":"2022","unstructured":"Yi Cheng and Li Kuang. 2022. CSRS: Code search with relevance matching and semantic matching. In ICPC, 533\u2013542."},{"key":"e_1_3_2_16_2","first-page":"108","volume-title":"MSR","author":"Ciniselli Matteo","year":"2021","unstructured":"Matteo Ciniselli, Nathan Cooper, Luca Pascarella, Denys Poshyvanyk, Massimiliano Di Penta, and Gabriele Bavota. 2021. An empirical study on the usage of BERT models for code completion. In MSR, 108\u2013119."},{"key":"e_1_3_2_17_2","volume-title":"ICLR","author":"Clark Kevin","year":"2020","unstructured":"Kevin Clark, Minh-Thang Luong, Quoc V. Le, and Christopher D. Manning. 2020. ELECTRA: Pre-training text encoders as discriminators rather than generators. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=r1xMH1BtvB"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1023\/A:1006569829653"},{"key":"e_1_3_2_19_2","first-page":"1475","volume-title":"ICML","volume":"97","author":"Cvitkovic Milan","year":"2019","unstructured":"Milan Cvitkovic, Badal Singh, and Animashree Anandkumar. 2019. Open vocabulary learning on source code with a graph-structured cache. In ICML, Vol. 97, 1475\u20131485."},{"key":"e_1_3_2_20_2","first-page":"4171","volume-title":"NAACL-HLT (1)","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In NAACL-HLT (1), 4171\u20134186."},{"key":"e_1_3_2_21_2","unstructured":"Jeff Donahue Philipp Kr\u00e4henb\u00fchl and Trevor Darrell. 2017. Adversarial feature learning. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=BJtNZAFgg"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1109\/TBME.2016.2613502"},{"key":"e_1_3_2_23_2","first-page":"3070","volume-title":"WWW","author":"Fan Shaohua","year":"2020","unstructured":"Shaohua Fan, Xiao Wang, Chuan Shi, Emiao Lu, Ken Lin, and Bai Wang. 2020. One2Multi graph autoencoder for multi-view graph clustering. In WWW, 3070\u20133076."},{"key":"e_1_3_2_24_2","first-page":"7","volume-title":"ACM Multimedia","author":"Feng Fangxiang","year":"2014","unstructured":"Fangxiang Feng, Xiaojie Wang, and Ruifan Li. 2014. Cross-modal retrieval with correspondence autoencoder. In ACM Multimedia, 7\u201316."},{"key":"e_1_3_2_25_2","first-page":"264","volume-title":"CVPR","author":"Feng Yifan","year":"2018","unstructured":"Yifan Feng, Zizhao Zhang, Xibin Zhao, Rongrong Ji, and Yue Gao. 2018. GVCNN: Group-view convolutional neural networks for 3D shape recognition. In CVPR, 264\u2013272."},{"key":"e_1_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.139"},{"key":"e_1_3_2_27_2","first-page":"649","volume-title":"ICSE (Companion Volume)","author":"Fowkes Jaroslav M.","year":"2016","unstructured":"Jaroslav M. Fowkes, Pankajan Chanthirasegaran, Razvan Ranca, Miltiadis Allamanis, Mirella Lapata, and Charles Sutton. 2016. TASSAL: Autofolding for source code summarization. In ICSE (Companion Volume), 649\u2013652."},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1145\/3180155.3180167"},{"key":"e_1_3_2_29_2","first-page":"7212","volume-title":"ACL (1)","author":"Guo Daya","year":"2022","unstructured":"Daya Guo, Shuai Lu, Nan Duan, Yanlin Wang, Ming Zhou, and Jian Yin. 2022. UniXcoder: Unified cross-modal pre-training for code representation. In ACL (1), 7212\u20137225."},{"key":"e_1_3_2_30_2","volume-title":"ICLR","author":"Guo Daya","year":"2021","unstructured":"Daya Guo, Shuo Ren, Shuai Lu, Zhangyin Feng, Duyu Tang, Shujie Liu, Long Zhou, Nan Duan, Alexey Svyatkovskiy, Shengyu Fu, et al. 2021. GraphCodeBERT: Pre-training code representations with data flow. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=jLoC4ez43PZ"},{"key":"e_1_3_2_31_2","doi-asserted-by":"publisher","DOI":"10.5555\/3002812"},{"key":"e_1_3_2_32_2","first-page":"4116","article-title":"Contrastive multi-view representation learning on graphs","volume":"119","author":"Hassani Kaveh","year":"2020","unstructured":"Kaveh Hassani and Amir Hosein Khas Ahmadi. 2020. Contrastive multi-view representation learning on graphs. In ICML, Vol. 119, 4116\u20134126.","journal-title":"ICML"},{"key":"e_1_3_2_33_2","volume-title":"ICLR","author":"Hjelm R. Devon","year":"2019","unstructured":"R. Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Philip Bachman, Adam Trischler, and Yoshua Bengio. 2019. Learning deep representations by mutual information estimation and maximization. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=Bklr3j0cKX"},{"key":"e_1_3_2_34_2","volume-title":"ICLR","author":"Hu Weihua","year":"2020","unstructured":"Weihua Hu, Bowen Liu, Joseph Gomes, Marinka Zitnik, Percy Liang, Vijay S. Pande, and Jure Leskovec. 2020. Strategies for pre-training graph neural networks. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=HJlWWJSFDH"},{"key":"e_1_3_2_35_2","first-page":"5690","volume-title":"ACL\/IJCNLP (1)","author":"Huang Junjie","year":"2021","unstructured":"Junjie Huang, Duyu Tang, Linjun Shou, Ming Gong, Ke Xu, Daxin Jiang, Ming Zhou, and Nan Duan. 2021. CoSQA: 20,000+ web queries for code search and question answering. In ACL\/IJCNLP (1), 5690\u20135700."},{"key":"e_1_3_2_36_2","unstructured":"Hamel Husain Ho-Hsiang Wu Tiferet Gazit Miltiadis Allamanis and Marc Brockschmidt. 2019. CodeSearchNet challenge: Evaluating the state of semantic code search. arXiv:1909.09436. Retrieved from https:\/\/arxiv.org\/abs\/1909.09436"},{"key":"e_1_3_2_37_2","unstructured":"Vinoj Jayasundara Nghi Duy Quoc Bui Lingxiao Jiang and David Lo. 2019. TreeCaps: Tree-structured capsule networks for program source code processing. arXiv:1910.12306. Retrieved from https:\/\/arxiv.org\/abs\/1910.12306"},{"key":"e_1_3_2_38_2","first-page":"222","volume-title":"ICDM","author":"Jiao Yizhu","year":"2020","unstructured":"Yizhu Jiao, Yun Xiong, Jiawei Zhang, Yao Zhang, Tianqi Zhang, and Yangyong Zhu. 2020. Sub-graph contrast for scalable self-supervised graph representation learning. In ICDM, 222\u2013231."},{"key":"e_1_3_2_39_2","unstructured":"Wei Jin Tyler Derr Haochen Liu Yiqi Wang Suhang Wang Zitao Liu and Jiliang Tang. 2020. Self-supervised learning on graphs: Deep insights and new direction. arXiv:2006.10141. Retrieved from https:\/\/arxiv.org\/abs\/2006.10141"},{"key":"e_1_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1145\/3437963.3441735"},{"key":"e_1_3_2_41_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2598339"},{"key":"e_1_3_2_42_2","volume-title":"ICLR","author":"Kingma Diederik P.","year":"2015","unstructured":"Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In ICLR. Retrieved from https:\/\/arxiv.org\/abs\/1412.6980"},{"issue":"3","key":"e_1_3_2_43_2","first-page":"62:1","article-title":"Deep learning for source code modeling and generation: Models, applications, and challenges","volume":"53","author":"Huynh Triet","year":"2020","unstructured":"Triet Huynh, Minh Le, Hao Chen, and Muhammad Ali Babar. 2020. Deep learning for source code modeling and generation: Models, applications, and challenges. ACM Comput. Surv 53, 3 (2020), 62:1\u201362:38.","journal-title":"ACM Comput. Surv"},{"key":"e_1_3_2_44_2","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2018.2872063"},{"key":"e_1_3_2_45_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICPC52881.2021.00026"},{"key":"e_1_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3197238"},{"key":"e_1_3_2_47_2","first-page":"11174","volume-title":"CVPR","author":"Lin Yijie","year":"2021","unstructured":"Yijie Lin, Yuanbiao Gou, Zitao Liu, Boyun Li, Jiancheng Lv, and Xi Peng. 2021. COMPLETER: Incomplete multi-view clustering via contrastive prediction. In CVPR, 11174\u201311183."},{"key":"e_1_3_2_48_2","doi-asserted-by":"publisher","DOI":"10.1145\/3447571"},{"key":"e_1_3_2_49_2","doi-asserted-by":"publisher","DOI":"10.1145\/3480027"},{"key":"e_1_3_2_50_2","first-page":"740","volume-title":"ISSTA","author":"Liu Hao","year":"2023","unstructured":"Hao Liu, Yanlin Wang, Zhao Wei, Yong Xu, Juhong Wang, Hui Li, and Rongrong Ji. 2023. RefBERT: A two-stage pre-trained framework for automatic rename refactoring. In ISSTA, 740\u2013752."},{"key":"e_1_3_2_51_2","first-page":"1","article-title":"Self-supervised learning: Generative or contrastive","author":"Liu Xiao","year":"2021","unstructured":"Xiao Liu, Fanjin Zhang, Zhenyu Hou, Zhaoyu Wang, Li Mian, Jing Zhang, and Jie Tang. 2021. Self-supervised learning: Generative or contrastive. IEEE Trans. Knowl. Data Eng., 1\u20131.","journal-title":"IEEE Trans. Knowl. Data Eng"},{"key":"e_1_3_2_52_2","unstructured":"Yinhan Liu Myle Ott Naman Goyal Jingfei Du Mandar Joshi Danqi Chen Omer Levy Mike Lewis Luke Zettlemoyer and Veselin Stoyanov. 2019. RoBERTa: A robustly optimized BERT pretraining approach. arXiv:1907.11692. Retrieved from https:\/\/arxiv.org\/abs\/1907.11692"},{"issue":"6","key":"e_1_3_2_53_2","first-page":"5879","article-title":"Graph self-supervised learning: A survey","volume":"35","author":"Liu Yixin","year":"2023","unstructured":"Yixin Liu, Shirui Pan, Ming Jin, Chuan Zhou, Feng Xia, and Philip S. Yu. 2023. Graph self-supervised learning: A survey. IEEE Trans. Neural Networks Learn. Syst. 35, 6 (2023), 5879\u20135900.","journal-title":"IEEE Trans. Neural Networks Learn. Syst"},{"key":"e_1_3_2_54_2","first-page":"260","volume-title":"ASE","author":"Lv Fei","year":"2015","unstructured":"Fei Lv, Hongyu Zhang, Jian-Guang Lou, Shaowei Wang, Dongmei Zhang, and Jianjun Zhao. 2015. CodeHow: Effective code search based on API understanding and extended Boolean model (E). In ASE, 260\u2013270."},{"key":"e_1_3_2_55_2","unstructured":"Hehuan Ma Yatao Bian Yu Rong Wenbing Huang Tingyang Xu Weiyang Xie Geyan Ye and Junzhou Huang. 2020. Dual message passing neural network for molecular property prediction. arXiv:2005.13607. Retrieved from https:\/\/arxiv.org\/abs\/2005.13607"},{"key":"e_1_3_2_56_2","volume-title":"ICLR","author":"Mao Junhua","year":"2015","unstructured":"Junhua Mao, Wei Xu, Yi Yang, Jiang Wang, and Alan L. Yuille. 2015. Deep captioning with multimodal recurrent neural networks (m-RNN). In ICLR. Retrieved from https:\/\/arxiv.org\/abs\/1412.6632"},{"key":"e_1_3_2_57_2","first-page":"541","volume-title":"PAKDD (1)","volume":"12712","author":"Mavromatis Costas","year":"2021","unstructured":"Costas Mavromatis and George Karypis. 2021. Graph InfoClust: Maximizing coarse-grain mutual information in graphs. In PAKDD (1), Vol. 12712, 541\u2013553."},{"key":"e_1_3_2_58_2","doi-asserted-by":"publisher","DOI":"10.1145\/1985793.1985809"},{"key":"e_1_3_2_59_2","doi-asserted-by":"publisher","DOI":"10.1145\/2522920.2522930"},{"key":"e_1_3_2_60_2","first-page":"689","volume-title":"ICML","author":"Ngiam Jiquan","year":"2011","unstructured":"Jiquan Ngiam, Aditya Khosla, Mingyu Kim, Juhan Nam, Honglak Lee, and Andrew Y. Ng. 2011. Multimodal deep learning. In ICML, 689\u2013696."},{"key":"e_1_3_2_61_2","first-page":"8476","volume-title":"ICML. PMLR","volume":"139","author":"Peng Dinglan","year":"2021","unstructured":"Dinglan Peng, Shuxin Zheng, Yatao Li, Guolin Ke, Di He, and Tie-Yan Liu. 2021. How could neural networks understand programs? In ICML. PMLR, Vol. 139, 8476\u20138486."},{"key":"e_1_3_2_62_2","unstructured":"Zhen Peng Yixiang Dong Minnan Luo Xiao-Ming Wu and Qinghua Zheng. 2020. Self-supervised graph representation learning via global context prediction. arXiv:2003.01604. Retrieved from https:\/\/arxiv.org\/abs\/2003.01604"},{"key":"e_1_3_2_63_2","first-page":"1150","volume-title":"KDD","author":"Qiu Jiezhong","year":"2020","unstructured":"Jiezhong Qiu, Qibin Chen, Yuxiao Dong, Jing Zhang, Hongxia Yang, Ming Ding, Kuansan Wang, and Jie Tang. 2020. GCC: Graph contrastive coding for graph neural network pre-training. In KDD, 1150\u20131160."},{"key":"e_1_3_2_64_2","unstructured":"Yuxiang Ren Bo Liu Chao Huang Peng Dai Liefeng Bo and Jiawei Zhang. 2019. Heterogeneous deep graph infomax. arXiv:1911.08538. Retrieved from https:\/\/arxiv.org\/abs\/1911.08538"},{"key":"e_1_3_2_65_2","unstructured":"Joshua David Robinson Ching-Yao Chuang Suvrit Sra and Stefanie Jegelka. 2021. Contrastive learning with hard negative samples. In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=CR1XOQ0UTh"},{"key":"e_1_3_2_66_2","doi-asserted-by":"publisher","DOI":"10.1145\/361219.361220"},{"key":"e_1_3_2_67_2","first-page":"593","volume-title":"ESWC","volume":"10843","author":"Sejr Schlichtkrull Michael","year":"2018","unstructured":"Michael Sejr Schlichtkrull, Thomas N. Kipf, Peter Bloem, Rianne van den Berg, Ivan Titov, and Max Welling. 2018. Modeling relational data with graph convolutional networks. In ESWC, Vol. 10843, 593\u2013607."},{"key":"e_1_3_2_68_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.114"},{"key":"e_1_3_2_69_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i04.6048"},{"key":"e_1_3_2_70_2","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510140"},{"key":"e_1_3_2_71_2","unstructured":"Zhensu Sun Yan Liu Chen Yang and Yu Qian. 2020. PSCS: A path-based neural model for semantic code search. arXiv:2008.03042. Retrieved from https:\/\/arxiv.org\/abs\/2008.03042"},{"key":"e_1_3_2_72_2","first-page":"2727","volume-title":"KDD","author":"Svyatkovskiy Alexey","year":"2019","unstructured":"Alexey Svyatkovskiy, Ying Zhao, Shengyu Fu, and Neel Sundaresan. 2019. Pythia: AI-assisted code completion system. In KDD, 2727\u20132735."},{"key":"e_1_3_2_73_2","first-page":"942","volume-title":"IJCAI","author":"Tian Yu","year":"2018","unstructured":"Yu Tian, Xi Peng, Long Zhao, Shaoting Zhang, and Dimitris N. Metaxas. 2018. CR-GAN: Learning complete representations for multi-view generation. In IJCAI, 942\u2013948."},{"key":"e_1_3_2_74_2","unstructured":"A\u00e4ron van den Oord Yazhe Li and Oriol Vinyals. 2018. Representation learning with contrastive predictive coding. arXiv:1807.03748. Retrieved from https:\/\/arxiv.org\/abs\/1807.03748"},{"key":"e_1_3_2_75_2","first-page":"354","volume-title":"ICPC","author":"Wang Xiao","year":"2022","unstructured":"Xiao Wang, Qiong Wu, Hongyu Zhang, Chen Lyu, Xue Jiang, Zhuoran Zheng, Lei Lyu, and Songlin Hu. 2022. HELoC: Hierarchical contrastive learning of source code representation. In ICPC, 354\u2013365."},{"key":"e_1_3_2_76_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i16.17650"},{"key":"e_1_3_2_77_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-16-6054-2"},{"key":"e_1_3_2_78_2","first-page":"4216","article-title":"Self-supervised on graphs: Contrastive, generative, or predictive","author":"Wu Lirong","year":"2023","unstructured":"Lirong Wu, Haitao Lin, Zhangyang Gao, Cheng Tan, and Stan Z. Li. 2023. Self-supervised on graphs: Contrastive, generative, or predictive. IEEE Trans. Knowl. Data Eng. 35, 4 (2023), 4216\u20134235.","journal-title":"IEEE Trans. Knowl. Data Eng"},{"issue":"2","key":"e_1_3_2_79_2","first-page":"54:1","article-title":"Survey of code search based on deep learning","volume":"33","author":"Xie Yutao","year":"2024","unstructured":"Yutao Xie, Jiayi Lin, Hande Dong, Lei Zhang, and Zhonghai Wu. 2024. Survey of code search based on deep learning. ACM Trans. Softw. Eng. Methodol. 33, 2 (2024), 54:1\u201354:42.","journal-title":"ACM Trans. Softw. Eng. Methodol"},{"key":"e_1_3_2_80_2","first-page":"2412","article-title":"Self-supervised learning of graph neural networks: A unified review","author":"Xie Yaochen","year":"2023","unstructured":"Yaochen Xie, Zhao Xu, Jingtun Zhang, Zhengyang Wang, and Shuiwang Ji. 2023. Self-supervised learning of graph neural networks: A unified review. IEEE Trans. Pattern Anal. Mach. Intell. 45, 2 (2023), 2412\u20132429.","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell"},{"key":"e_1_3_2_81_2","unstructured":"Keyulu Xu Weihua Hu Jure Leskovec and Stefanie Jegelka. 2019. How powerful are graph neural networks? In ICLR. Retrieved from https:\/\/openreview.net\/pdf?id=ryGs6iA5Km"},{"key":"e_1_3_2_82_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2021.03.090"},{"key":"e_1_3_2_83_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3155499"},{"issue":"10","key":"e_1_3_2_84_2","first-page":"206:1","article-title":"A survey on deep learning for software engineering","volume":"54","author":"Yang Yanming","year":"2022","unstructured":"Yanming Yang, Xin Xia, David Lo, and John C. Grundy. 2022. A survey on deep learning for software engineering. ACM Comput. Surv. 54, 10s (2022), 206:1\u2013206:73.","journal-title":"ACM Comput. Surv"},{"key":"e_1_3_2_85_2","article-title":"Graph contrastive learning with augmentations","author":"You Yuning","year":"2020","unstructured":"Yuning You, Tianlong Chen, Yongduo Sui, Ting Chen, Zhangyang Wang, and Yang Shen. 2020. Graph contrastive learning with augmentations. In NeurIPS.","journal-title":"NeurIPS"},{"key":"e_1_3_2_86_2","first-page":"10871","article-title":"When does self-supervision help graph convolutional networks","author":"You Yuning","year":"2020","unstructured":"Yuning You, Tianlong Chen, Zhangyang Wang, and Yang Shen. 2020. When does self-supervision help graph convolutional networks? In ICML, 10871\u201310880.","journal-title":"ICML"},{"issue":"2","key":"e_1_3_2_87_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3546066","article-title":"deGraphCS: Embedding variable-based flow graph for neural code search","volume":"32","author":"Zeng Chen","year":"2022","unstructured":"Chen Zeng, Yue Yu, Shanshan Li, Xin Xia, Zhiming Wang, Mingyang Geng, Bailin Xiao, Wei Dong, and Xiangke Liao. 2022. deGraphCS: Embedding variable-based flow graph for neural code search. ACM Trans. Softw. Eng. Methodol. 32, 2 (2022), 1\u201327.","journal-title":"ACM Trans. Softw. Eng. Methodol"},{"key":"e_1_3_2_88_2","first-page":"217","volume-title":"ASE","author":"Zhang Chenyuan","year":"2023","unstructured":"Chenyuan Zhang, Yanlin Wang, Zhao Wei, Yong Xu, Juhong Wang, Hui Li, and Rongrong Ji. 2023. EALink: An efficient and accurate pre-trained framework for issue-commit link recovery. In ASE, 217\u2013229."},{"key":"e_1_3_2_89_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2019.00086"},{"key":"e_1_3_2_90_2","unstructured":"Sheng Zhang Hui Li Yanlin Wang Zhao Wei Yong Xiu Juhong Wang and Rongong Ji. 2023. Code search debiasing: Improve search results beyond overall ranking performance. arXiv:2311.14901. Retrieved from https:\/\/arxiv.org\/abs\/2311.14901"},{"key":"e_1_3_2_91_2","doi-asserted-by":"publisher","DOI":"10.1145\/3324884.3416530"},{"key":"e_1_3_2_92_2","unstructured":"Yanqiao Zhu Yichen Xu Feng Yu Qiang Liu Shu Wu and Liang Wang. 2020. Deep graph contrastive representation learning. arXiv:2006.04131. Retrieved from https:\/\/arxiv.org\/abs\/2006.04131"}],"container-title":["ACM Transactions on Knowledge Discovery from Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3737878","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,9]],"date-time":"2025-07-09T15:02:48Z","timestamp":1752073368000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3737878"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,9]]},"references-count":91,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,7,31]]}},"alternative-id":["10.1145\/3737878"],"URL":"https:\/\/doi.org\/10.1145\/3737878","relation":{},"ISSN":["1556-4681","1556-472X"],"issn-type":[{"value":"1556-4681","type":"print"},{"value":"1556-472X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,9]]},"assertion":[{"value":"2022-10-18","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-05-08","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-07-09","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}