{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,4]],"date-time":"2026-05-04T10:06:04Z","timestamp":1777889164831,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":55,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,4,14]],"date-time":"2024-04-14T00:00:00Z","timestamp":1713052800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,4,14]]},"DOI":"10.1145\/3639477.3639719","type":"proceedings-article","created":{"date-parts":[[2024,5,31]],"date-time":"2024-05-31T09:27:26Z","timestamp":1717147646000},"page":"418-429","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":16,"title":["CodeFuse-13B: A Pretrained Multi-lingual Code Large Language Model"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5799-5876","authenticated-orcid":false,"given":"Peng","family":"Di","sequence":"first","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8645-0680","authenticated-orcid":false,"given":"Jianguo","family":"Li","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5639-0912","authenticated-orcid":false,"given":"Hang","family":"Yu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-6605-9793","authenticated-orcid":false,"given":"Wei","family":"Jiang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-1229-3388","authenticated-orcid":false,"given":"Wenting","family":"Cai","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-4723-293X","authenticated-orcid":false,"given":"Yang","family":"Cao","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6133-4324","authenticated-orcid":false,"given":"Chaoyu","family":"Chen","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-9532-7636","authenticated-orcid":false,"given":"Dajun","family":"Chen","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-8624-9056","authenticated-orcid":false,"given":"Hongwei","family":"Chen","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-9655-1072","authenticated-orcid":false,"given":"Liang","family":"Chen","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8633-6036","authenticated-orcid":false,"given":"Gang","family":"Fan","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-2226-6821","authenticated-orcid":false,"given":"Jie","family":"Gong","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-7142-8433","authenticated-orcid":false,"given":"Zi","family":"Gong","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4874-2260","authenticated-orcid":false,"given":"Wen","family":"Hu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-1732-8631","authenticated-orcid":false,"given":"Tingting","family":"Guo","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-1599-4675","authenticated-orcid":false,"given":"Zhichao","family":"Lei","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2521-3240","authenticated-orcid":false,"given":"Ting","family":"Li","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-1264-7284","authenticated-orcid":false,"given":"Zheng","family":"Li","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3622-1510","authenticated-orcid":false,"given":"Ming","family":"Liang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-6393-9035","authenticated-orcid":false,"given":"Cong","family":"Liao","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-9380-6168","authenticated-orcid":false,"given":"Bingchang","family":"Liu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-0053-1733","authenticated-orcid":false,"given":"Jiachen","family":"Liu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4829-0751","authenticated-orcid":false,"given":"Zhiwei","family":"Liu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-1775-7943","authenticated-orcid":false,"given":"Shaojun","family":"Lu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-8418-1877","authenticated-orcid":false,"given":"Min","family":"Shen","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-8652-6294","authenticated-orcid":false,"given":"Guangpei","family":"Wang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-3333-7195","authenticated-orcid":false,"given":"Huan","family":"Wang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1095-9632","authenticated-orcid":false,"given":"Zhi","family":"Wang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-3975-2481","authenticated-orcid":false,"given":"Zhaogui","family":"Xu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-7739-8247","authenticated-orcid":false,"given":"Jiawei","family":"Yang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-8969-1291","authenticated-orcid":false,"given":"Qing","family":"Ye","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-8036-5749","authenticated-orcid":false,"given":"Gehao","family":"Zhang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-8097-4976","authenticated-orcid":false,"given":"Yu","family":"Zhang","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-2187-0229","authenticated-orcid":false,"given":"Zelin","family":"Zhao","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-8146-4861","authenticated-orcid":false,"given":"Xunjin","family":"Zheng","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-0476-4449","authenticated-orcid":false,"given":"Hailian","family":"Zhou","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-1738-4525","authenticated-orcid":false,"given":"Lifu","family":"Zhu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-8219-0495","authenticated-orcid":false,"given":"Xianying","family":"Zhu","sequence":"additional","affiliation":[{"name":"Ant Group, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2024,5,31]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Loubna Ben Allal Raymond Li Denis Kocetkov et al. 2023. SantaCoder: don't reach for the stars! arXiv:2301.03988 [cs.SE]"},{"key":"e_1_3_2_1_2_1","unstructured":"Rohan Anil Andrew M. Dai Orhan Firat et al. 2023. PaLM 2 Technical Report. arXiv:2305.10403 [cs.CL]"},{"key":"e_1_3_2_1_3_1","unstructured":"Jacob Austin Augustus Odena Maxwell Nye et al. 2021. Program Synthesis with Large Language Models. arXiv:2108.07732 [cs.PL]"},{"key":"e_1_3_2_1_4_1","volume-title":"Jamie Ryan Kiros, and Geoffrey E. Hinton","author":"Ba Jimmy Lei","year":"2016","unstructured":"Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E. Hinton. 2016. Layer Normalization. arXiv:1607.06450 [stat.ML]"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"crossref","unstructured":"Sid Black Stella Biderman Eric Hallahan et al. 2022. GPT-NeoX-20B: An Open-Source Autoregressive Language Model. arXiv:2204.06745 [cs.CL]","DOI":"10.18653\/v1\/2022.bigscience-1.9"},{"key":"e_1_3_2_1_6_1","unstructured":"Tom B. Brown Benjamin Mann Nick Ryder et al. 2020. Language Models are Few-Shot Learners. arXiv:2005.14165 [cs.CL]"},{"key":"e_1_3_2_1_7_1","unstructured":"Mark Chen Jerry Tworek Heewoo Jun et al. 2021. Evaluating Large Language Models Trained on Code. arXiv:2107.03374 [cs.LG]"},{"key":"e_1_3_2_1_8_1","unstructured":"Mark Chen Jerry Tworek Heewoo Jun et al. 2021. Evaluating Large Language Models Trained on Code. arXiv:2107.03374 [cs.LG]"},{"key":"e_1_3_2_1_9_1","unstructured":"Aakanksha Chowdhery Sharan Narang Jacob Devlin et al. 2022. PaLM: Scaling Language Modeling with Pathways. arXiv:2204.02311 [cs.CL]"},{"key":"e_1_3_2_1_10_1","unstructured":"Tri Dao Daniel Y. Fu Stefano Ermon et al. 2022. FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness. arXiv:2205.14135 [cs.LG]"},{"key":"e_1_3_2_1_11_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805 [cs.CL]","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805 [cs.CL]"},{"key":"e_1_3_2_1_12_1","volume-title":"GPTQ: Accurate post-training quantization for generative pre-trained transformers. arXiv preprint arXiv:2210.17323","author":"Frantar Elias","year":"2022","unstructured":"Elias Frantar, Saleh Ashkboos, Torsten Hoefler, and Dan Alistarh. 2022. GPTQ: Accurate post-training quantization for generative pre-trained transformers. arXiv preprint arXiv:2210.17323 (2022)."},{"key":"e_1_3_2_1_13_1","unstructured":"Daniel Fried Armen Aghajanyan Jessy Lin et al. 2023. InCoder: A Generative Model for Code Infilling and Synthesis. arXiv:2204.05999 [cs.SE]"},{"key":"e_1_3_2_1_14_1","volume-title":"The Pile: An 800GB Dataset of Diverse Text for Language Modeling. arXiv:2101.00027 [cs.CL]","author":"Gao Leo","year":"2020","unstructured":"Leo Gao, Stella Biderman, Sid Black, et al. 2020. The Pile: An 800GB Dataset of Diverse Text for Language Modeling. arXiv:2101.00027 [cs.CL]"},{"key":"e_1_3_2_1_15_1","unstructured":"Ant Group. 2023. Sparrow. http:\/\/sparrow.alipay.com."},{"key":"e_1_3_2_1_16_1","unstructured":"Dan Hendrycks and Kevin Gimpel. 2023. Gaussian Error Linear Units (GELUs). arXiv:1606.08415 [cs.LG]"},{"key":"e_1_3_2_1_17_1","unstructured":"Jordan Hoffmann Sebastian Borgeaud Arthur Mensch et al. 2022. Training Compute-Optimal Large Language Models. arXiv:2203.15556 [cs.CL]"},{"key":"e_1_3_2_1_18_1","volume-title":"Kingma and Jimmy Ba","author":"Diederik","year":"2017","unstructured":"Diederik P. Kingma and Jimmy Ba. 2017. Adam: A Method for Stochastic Optimization. arXiv:1412.6980 [cs.LG]"},{"key":"e_1_3_2_1_19_1","volume-title":"Loubna Ben Allal, et al","author":"Kocetkov Denis","year":"2022","unstructured":"Denis Kocetkov, Raymond Li, Loubna Ben Allal, et al. 2022. The Stack: 3 TB of permissively licensed source code. arXiv:2211.15533 [cs.CL]"},{"key":"e_1_3_2_1_20_1","volume-title":"Yangtian Zi, et al.","author":"Li Raymond","year":"2023","unstructured":"Raymond Li, Loubna Ben Allal, Yangtian Zi, et al. 2023. StarCoder: may the source be with you! arXiv:2305.06161 [cs.CL]"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1126\/science.abq1158"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3597926.3598042"},{"key":"e_1_3_2_1_23_1","unstructured":"Ziyang Luo Can Xu Pu Zhao et al. 2023. WizardCoder: Empowering Code Large Language Models with Evol-Instruct. arXiv:2306.08568 [cs.CL]"},{"key":"e_1_3_2_1_24_1","unstructured":"MicroSoft. 2023. ChatML. https:\/\/github.com\/openai\/openai-python\/blob\/main\/chatml.md."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/SCAM.2007.31"},{"key":"e_1_3_2_1_26_1","unstructured":"Erik Nijkamp Bo Pang Hiroaki Hayashi et al. 2023. CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis. arXiv:2203.13474 [cs.LG]"},{"key":"e_1_3_2_1_27_1","unstructured":"OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774 [cs.CL]"},{"key":"e_1_3_2_1_28_1","unstructured":"Jack W. Rae Sebastian Borgeaud Trevor Cai et al. 2022. Scaling Language Models: Methods Analysis & Insights from Training Gopher. arXiv:2112.11446 [cs.CL]"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Samyam Rajbhandari Jeff Rasley Olatunji Ruwase and Yuxiong He. 2020. ZeRO: Memory Optimizations Toward Training Trillion Parameter Models. arXiv:1910.02054 [cs.LG]","DOI":"10.1109\/SC41405.2020.00024"},{"key":"e_1_3_2_1_30_1","volume-title":"Code Llama: Open Foundation Models for Code. arXiv:2308.12950 [cs.CL]","author":"Rozi\u00e8re Baptiste","year":"2023","unstructured":"Baptiste Rozi\u00e8re, Jonas Gehring, Fabian Gloeckle, et al. 2023. Code Llama: Open Foundation Models for Code. arXiv:2308.12950 [cs.CL]"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","unstructured":"Max Sch\u00e4fer Sarah Nadi Aryaz Eghbali and Frank Tip. 2023. An Empirical Evaluation of Using Large Language Models for Automated Unit Test Generation. arXiv:2302.06527 [cs.SE]","DOI":"10.1109\/TSE.2023.3334955"},{"key":"e_1_3_2_1_32_1","unstructured":"Bo Shen Jiaxin Zhang Taihong Chen et al. 2023. PanGu-Coder2: Boosting Large Language Models for Code with Ranking Feedback. arXiv:2307.14936 [cs.CL]"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3192366.3192418"},{"key":"e_1_3_2_1_34_1","unstructured":"Yusuke Shibata Takuya Kida Shuichi Fukamachi et al. 1999. Byte Pair Encoding: A Text Compression Scheme That Accelerates Pattern Matching. (09 1999)."},{"key":"e_1_3_2_1_35_1","unstructured":"Mohammad Shoeybi Mostofa Patwary Raul Puri et al. 2020. Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism. arXiv:1909.08053 [cs.CL]"},{"key":"e_1_3_2_1_36_1","unstructured":"Jianlin Su Yu Lu Shengfeng Pan et al. 2022. RoFormer: Enhanced Transformer with Rotary Position Embedding. arXiv:2104.09864 [cs.CL]"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"crossref","unstructured":"Yu Sun Shuohuan Wang Yukun Li et al. 2019. ERNIE 2.0: A Continual Pre-training Framework for Language Understanding. arXiv:1907.12412 [cs.CL]","DOI":"10.1609\/aaai.v34i05.6428"},{"key":"e_1_3_2_1_38_1","volume-title":"ERNIE: Enhanced Representation through Knowledge Integration. arXiv:1904.09223 [cs.CL]","author":"Sun Yu","year":"2019","unstructured":"Yu Sun, Shuohuan Wang, Yukun Li, et al. 2019. ERNIE: Enhanced Representation through Knowledge Integration. arXiv:1904.09223 [cs.CL]"},{"key":"e_1_3_2_1_39_1","volume-title":"Stanford Alpaca: An Instruction-following LLaMA model. https:\/\/github.com\/tatsu-lab\/stanford_alpaca.","author":"Taori Rohan","year":"2023","unstructured":"Rohan Taori, Ishaan Gulrajani, Tianyi Zhang, et al. 2023. Stanford Alpaca: An Instruction-following LLaMA model. https:\/\/github.com\/tatsu-lab\/stanford_alpaca."},{"key":"e_1_3_2_1_40_1","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard et al. 2023. LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971 [cs.CL]"},{"key":"e_1_3_2_1_41_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone et al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_1_42_1","unstructured":"Ben Wang and Aran Komatsuzaki. 2021. GPT-J-6B: A 6 Billion Parameter Autoregressive Language Model. https:\/\/github.com\/kingoflolz\/mesh-transformer-jax."},{"key":"e_1_3_2_1_43_1","unstructured":"Junjie Wang Yuchao Huang Chunyang Chen et al. 2023. Software Testing with Large Language Model: Survey Landscape and Vision. arXiv:2307.07221 [cs.SE]"},{"key":"e_1_3_2_1_44_1","volume-title":"Titan: Exploring Larger-scale Knowledge Enhanced Pre-training for Language Understanding and Generation. arXiv:2112.12731 [cs.CL]","author":"Wang Shuohuan","year":"2021","unstructured":"Shuohuan Wang, Yu Sun, Yang Xiang, et al. 2021. ERNIE 3.0 Titan: Exploring Larger-scale Knowledge Enhanced Pre-training for Language Understanding and Generation. arXiv:2112.12731 [cs.CL]"},{"key":"e_1_3_2_1_45_1","volume-title":"Akhilesh Deepak Gotmare, et al","author":"Wang Yue","year":"2023","unstructured":"Yue Wang, Hung Le, Akhilesh Deepak Gotmare, et al. 2023. CodeT5+: Open Code Large Language Models for Code Understanding and Generation. arXiv:2305.07922 [cs.CL]"},{"key":"e_1_3_2_1_46_1","volume-title":"Hoi","author":"Wang Yue","year":"2021","unstructured":"Yue Wang, Weishi Wang, Shafiq Joty, and Steven C. H. Hoi. 2021. CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation. arXiv:2109.00859 [cs.CL]"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"e_1_3_2_1_48_1","unstructured":"Ruibin Xiong Yunchang Yang Di He et al. 2020. On Layer Normalization in the Transformer Architecture. arXiv:2002.04745 [cs.LG]"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1145\/3453483.3454054"},{"key":"e_1_3_2_1_50_1","unstructured":"Aohan Zeng Xiao Liu Zhengxiao Du et al. 2022. GLM-130B: An Open Bilingual Pre-trained Model. arXiv:2210.02414 [cs.CL]"},{"key":"e_1_3_2_1_51_1","unstructured":"Wayne Xin Zhao Kun Zhou Junyi Li et al. 2023. A Survey of Large Language Models. arXiv:2303.18223 [cs.CL]"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"crossref","unstructured":"Qinkai Zheng Xiao Xia Xu Zou et al. 2023. CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Evaluations on HumanEval-X. arXiv:2303.17568 [cs.LG]","DOI":"10.1145\/3580305.3599790"},{"key":"e_1_3_2_1_53_1","volume-title":"Field-Based Static Taint Analysis for Industrial Microservices. In 44th IEEE\/ACM International Conference on Software Engineering: Software Engineering in Practice, ICSE (SEIP) 2022","author":"Zhong Zexin","year":"2022","unstructured":"Zexin Zhong, Jiangchao Liu, Diyu Wu, Peng Di, et al. 2022. Field-Based Static Taint Analysis for Industrial Microservices. In 44th IEEE\/ACM International Conference on Software Engineering: Software Engineering in Practice, ICSE (SEIP) 2022, Pittsburgh, PA, USA, May 22-24, 2022. IEEE, 149--150."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE-SEIP58684.2023.00015"},{"key":"e_1_3_2_1_55_1","volume-title":"Tails: How Long-Tailed Code Distributions Impact Large Language Models. arXiv:2309.03567 [cs.SE]","author":"Zhou Xin","year":"2023","unstructured":"Xin Zhou, Kisub Kim, Bowen Xu, et al. 2023. The Devil is in the Tails: How Long-Tailed Code Distributions Impact Large Language Models. arXiv:2309.03567 [cs.SE]"}],"event":{"name":"ICSE-SEIP '24: 46th International Conference on Software Engineering: Software Engineering in Practice","location":"Lisbon Portugal","acronym":"ICSE-SEIP '24","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering","IEEE CS","Faculty of Engineering of University of Porto"]},"container-title":["Proceedings of the 46th International Conference on Software Engineering: Software Engineering in Practice"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3639477.3639719","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3639477.3639719","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,28]],"date-time":"2025-08-28T17:36:36Z","timestamp":1756402596000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3639477.3639719"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,14]]},"references-count":55,"alternative-id":["10.1145\/3639477.3639719","10.1145\/3639477"],"URL":"https:\/\/doi.org\/10.1145\/3639477.3639719","relation":{},"subject":[],"published":{"date-parts":[[2024,4,14]]},"assertion":[{"value":"2024-05-31","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}