{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:14:51Z","timestamp":1750220091748,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":48,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T00:00:00Z","timestamp":1652659200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Academic Research Fund Tier 2","award":["MOE2019-T2-1-193"],"award-info":[{"award-number":["MOE2019-T2-1-193"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,5,16]]},"DOI":"10.1145\/3524610.3527916","type":"proceedings-article","created":{"date-parts":[[2022,10,20]],"date-time":"2022-10-20T15:19:30Z","timestamp":1666279170000},"page":"343-353","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["Benchmarking library recognition in tweets"],"prefix":"10.1145","author":[{"given":"Ting","family":"Zhang","sequence":"first","affiliation":[{"name":"Singapore Management University"}]},{"given":"Divya Prabha","family":"Chandrasekaran","sequence":"additional","affiliation":[{"name":"Singapore Management University"}]},{"given":"Ferdian","family":"Thung","sequence":"additional","affiliation":[{"name":"Singapore Management University"}]},{"given":"David","family":"Lo","sequence":"additional","affiliation":[{"name":"Singapore Management University"}]}],"member":"320","published-online":{"date-parts":[[2022,10,20]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"[n.d.]. Common Crawl. http:\/\/commoncrawl.org\/. (Accessed on 10\/22\/2021).  [n.d.]. Common Crawl. http:\/\/commoncrawl.org\/. (Accessed on 10\/22\/2021)."},{"key":"e_1_3_2_1_2_1","unstructured":"[n.d.]. Download | OpenWebTextCorpus. https:\/\/skylion007.github.io\/OpenWebTextCorpus\/. (Accessed on 10\/22\/2021).  [n.d.]. Download | OpenWebTextCorpus. https:\/\/skylion007.github.io\/OpenWebTextCorpus\/. (Accessed on 10\/22\/2021)."},{"key":"e_1_3_2_1_3_1","unstructured":"[n.d.]. News Dataset Available - Common Crawl. https:\/\/commoncrawl.org\/2016\/10\/news-dataset-available\/. (Accessed on 10\/22\/2021).  [n.d.]. News Dataset Available - Common Crawl. https:\/\/commoncrawl.org\/2016\/10\/news-dataset-available\/. (Accessed on 10\/22\/2021)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/2351676.2351740"},{"key":"e_1_3_2_1_5_1","volume-title":"Peter V Desouza, Jennifer C Lai, and Robert L Mercer.","author":"Brown Peter F","year":"1992","unstructured":"Peter F Brown , Vincent J Della Pietra , Peter V Desouza, Jennifer C Lai, and Robert L Mercer. 1992 . Class-based n-gram models of natural language. Computational linguistics 18, 4 (1992), 467--480. Peter F Brown, Vincent J Della Pietra, Peter V Desouza, Jennifer C Lai, and Robert L Mercer. 1992. Class-based n-gram models of natural language. Computational linguistics 18, 4 (1992), 467--480."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10664-017-9546-9"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/2786805.2786819"},{"key":"e_1_3_2_1_8_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2018 . Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018). Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE43902.2021.00133"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3379597.3387466"},{"key":"e_1_3_2_1_11_1","volume-title":"Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155","author":"Feng Zhangyin","year":"2020","unstructured":"Zhangyin Feng , Daya Guo , Duyu Tang , Nan Duan , Xiaocheng Feng , Ming Gong , Linjun Shou , Bing Qin , Ting Liu , Daxin Jiang , 2020 . Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155 (2020). Zhangyin Feng, Daya Guo, Duyu Tang, Nan Duan, Xiaocheng Feng, Ming Gong, Linjun Shou, Bing Qin, Ting Liu, Daxin Jiang, et al. 2020. Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155 (2020)."},{"key":"e_1_3_2_1_12_1","volume-title":"Graphcodebert: Pre-training code representations with data flow. arXiv preprint arXiv:2009.08366","author":"Guo Daya","year":"2020","unstructured":"Daya Guo , Shuo Ren , Shuai Lu , Zhangyin Feng , Duyu Tang , Shujie Liu , Long Zhou , Nan Duan , Alexey Svyatkovskiy , Shengyu Fu , 2020 . Graphcodebert: Pre-training code representations with data flow. arXiv preprint arXiv:2009.08366 (2020). Daya Guo, Shuo Ren, Shuai Lu, Zhangyin Feng, Duyu Tang, Shujie Liu, Long Zhou, Nan Duan, Alexey Svyatkovskiy, Shengyu Fu, et al. 2020. Graphcodebert: Pre-training code representations with data flow. arXiv preprint arXiv:2009.08366 (2020)."},{"volume-title":"A needle in a haystack: What do twitter users say about software?. In 2016 IEEE 24th international requirements engineering conference (RE)","author":"Guzman Emitza","key":"e_1_3_2_1_13_1","unstructured":"Emitza Guzman , Rana Alkadhi , and Norbert Seyff . 2016. A needle in a haystack: What do twitter users say about software?. In 2016 IEEE 24th international requirements engineering conference (RE) . IEEE , 96--105. Emitza Guzman, Rana Alkadhi, and Norbert Seyff. 2016. A needle in a haystack: What do twitter users say about software?. In 2016 IEEE 24th international requirements engineering conference (RE). IEEE, 96--105."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00766-017-0274-x"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/RE.2017.88"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2016.2597849"},{"key":"e_1_3_2_1_17_1","volume-title":"International Conference on Machine Learning. PMLR, 5110--5121","author":"Kanade Aditya","year":"2020","unstructured":"Aditya Kanade , Petros Maniatis , Gogul Balakrishnan , and Kensen Shi . 2020 . Learning and evaluating contextual embedding of source code . In International Conference on Machine Learning. PMLR, 5110--5121 . Aditya Kanade, Petros Maniatis, Gogul Balakrishnan, and Kensen Shi. 2020. Learning and evaluating contextual embedding of source code. In International Conference on Machine Learning. PMLR, 5110--5121."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE43902.2021.00040"},{"key":"e_1_3_2_1_19_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu , Myle Ott , Naman Goyal , Jingfei Du , Mandar Joshi , Danqi Chen , Omer Levy , Mike Lewis , Luke Zettlemoyer , and Veselin Stoyanov . 2019 . Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019). Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."},{"key":"e_1_3_2_1_20_1","volume-title":"Decoupled weight decayregularization. arXiv preprint arXiv:1711.05101","author":"Loshchilov Ilya","year":"2017","unstructured":"Ilya Loshchilov and Frank Hutter . 2017. Decoupled weight decayregularization. arXiv preprint arXiv:1711.05101 ( 2017 ). Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decayregularization. arXiv preprint arXiv:1711.05101 (2017)."},{"key":"e_1_3_2_1_21_1","volume-title":"Interrater reliability: the kappa statistic. Biochemia medica 22, 3","author":"McHugh Mary L","year":"2012","unstructured":"Mary L McHugh . 2012. Interrater reliability: the kappa statistic. Biochemia medica 22, 3 ( 2012 ), 276--282. Mary L McHugh. 2012. Interrater reliability: the kappa statistic. Biochemia medica 22, 3 (2012), 276--282."},{"key":"e_1_3_2_1_22_1","unstructured":"Tomas Mikolov Ilya Sutskever Kai Chen Greg S Corrado and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems. 3111--3119.  Tomas Mikolov Ilya Sutskever Kai Chen Greg S Corrado and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems. 3111--3119."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/219717.219748"},{"key":"e_1_3_2_1_24_1","volume-title":"Word sense disambiguation: A survey. ACM computing surveys (CSUR) 41, 2","author":"Navigli Roberto","year":"2009","unstructured":"Roberto Navigli . 2009. Word sense disambiguation: A survey. ACM computing surveys (CSUR) 41, 2 ( 2009 ), 1--69. Roberto Navigli. 2009. Word sense disambiguation: A survey. ACM computing surveys (CSUR) 41, 2 (2009), 1--69."},{"key":"e_1_3_2_1_25_1","volume-title":"BERTweet: A pre-trained language model for English Tweets. arXiv preprint arXiv:2005.10200","author":"Nguyen Dat Quoc","year":"2020","unstructured":"Dat Quoc Nguyen , Thanh Vu , and Anh Tuan Nguyen . 2020. BERTweet: A pre-trained language model for English Tweets. arXiv preprint arXiv:2005.10200 ( 2020 ). Dat Quoc Nguyen, Thanh Vu, and Anh Tuan Nguyen. 2020. BERTweet: A pre-trained language model for English Tweets. arXiv preprint arXiv:2005.10200 (2020)."},{"key":"e_1_3_2_1_26_1","volume-title":"Garnett (Eds.)","volume":"32","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke , Sam Gross , Francisco Massa , Adam Lerer , James Bradbury , Gregory Chanan , Trevor Killeen , Zeming Lin , Natalia Gimelshein , Luca Antiga , Alban Desmaison , Andreas Kopf , Edward Yang , Zachary DeVito , Martin Raison , Alykhan Tejani , Sasank Chilamkurthy , Benoit Steiner , Lu Fang , Junjie Bai , and Soumith Chintala . 2019 . PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alch\u00e9-Buc, E. Fox, and R . Garnett (Eds.) , Vol. 32 . Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/ 2019\/file\/bdbca288fee7f92f2bfa9f7012727740-Paper.pdf Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/bdbca288fee7f92f2bfa9f7012727740-Paper.pdf"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSM.2012.6405330"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11431-020-1647-3"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/SANER.2016.80"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/SANER.2015.7081855"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/2568225.2568305"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/1882362.1882435"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10664-019-09775-w"},{"key":"e_1_3_2_1_35_1","first-page":"693","article-title":"Dynamic Conditional Random Fields: Factorized Probabilistic Models for Labeling and Segmenting Sequence Data","volume":"8","author":"Sutton Charles","year":"2007","unstructured":"Charles Sutton , Andrew McCallum , and Khashayar Rohanimanesh . 2007 . Dynamic Conditional Random Fields: Factorized Probabilistic Models for Labeling and Segmenting Sequence Data . J. Mach. Learn. Res. 8 (2007), 693 -- 723 . http:\/\/dl.acm.org\/citation.cfm?id=1314523 Charles Sutton, Andrew McCallum, and Khashayar Rohanimanesh. 2007. Dynamic Conditional Random Fields: Factorized Probabilistic Models for Labeling and Segmenting Sequence Data. J. Mach. Learn. Res. 8 (2007), 693--723. http:\/\/dl.acm.org\/citation.cfm?id=1314523","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_1_36_1","volume-title":"Code and named entity recognition in stackoverflow. arXiv preprint arXiv:2005.01634","author":"Tabassum Jeniya","year":"2020","unstructured":"Jeniya Tabassum , Mounica Maddela , Wei Xu , and Alan Ritter . 2020. Code and named entity recognition in stackoverflow. arXiv preprint arXiv:2005.01634 ( 2020 ). Jeniya Tabassum, Mounica Maddela, Wei Xu, and Alan Ritter. 2020. Code and named entity recognition in stackoverflow. arXiv preprint arXiv:2005.01634 (2020)."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377811.3380403"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CSMR-WCRE.2014.6747213"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/2591062.2591071"},{"key":"e_1_3_2_1_40_1","volume-title":"A simple method for commonsense reasoning. arXiv preprint arXiv:1806.02847","author":"Trinh Trieu H","year":"2018","unstructured":"Trieu H Trinh and Quoc V Le. 2018. A simple method for commonsense reasoning. arXiv preprint arXiv:1806.02847 ( 2018 ). Trieu H Trinh and Quoc V Le. 2018. A simple method for commonsense reasoning. arXiv preprint arXiv:1806.02847 (2018)."},{"key":"e_1_3_2_1_41_1","volume-title":"Proceedings of the 48th annual meeting of the association for computational linguistics. 384--394","author":"Turian Joseph","year":"2010","unstructured":"Joseph Turian , Lev Ratinov , and Yoshua Bengio . 2010 . Word representations: a simple and general method for semi-supervised learning . In Proceedings of the 48th annual meeting of the association for computational linguistics. 384--394 . Joseph Turian, Lev Ratinov, and Yoshua Bengio. 2010. Word representations: a simple and general method for semi-supervised learning. In Proceedings of the 48th annual meeting of the association for computational linguistics. 384--394."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"e_1_3_2_1_43_1","volume-title":"Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32","author":"Yang Zhilin","year":"2019","unstructured":"Zhilin Yang , Zihang Dai , Yiming Yang , Jaime Carbonell , Russ R Salakhutdinov , and Quoc V Le . 2019 . Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32 (2019). Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ R Salakhutdinov, and Quoc V Le. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32 (2019)."},{"key":"e_1_3_2_1_44_1","volume-title":"Formal in the Informal: A Multi-Level Analysis of Core Python Developers' Tweets. In 2018 25th Australasian Software Engineering Conference (ASWEC). IEEE, 151--160","author":"Yasir Muhammad","year":"2018","unstructured":"Muhammad Yasir , Kevin Michael , Bastin Tony Roy Savarimuthu , and Sherlock A Licorish . 2018 . Formal in the Informal: A Multi-Level Analysis of Core Python Developers' Tweets. In 2018 25th Australasian Software Engineering Conference (ASWEC). IEEE, 151--160 . Muhammad Yasir, Kevin Michael, Bastin Tony Roy Savarimuthu, and Sherlock A Licorish. 2018. Formal in the Informal: A Multi-Level Analysis of Core Python Developers' Tweets. In 2018 25th Australasian Software Engineering Conference (ASWEC). IEEE, 151--160."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10664-018-9608-7"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3460319.3464819"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSME46990.2020.00017"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.11"}],"event":{"name":"ICPC '22: 30th International Conference on Program Comprehension","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering","IEEE CS"],"location":"Virtual Event","acronym":"ICPC '22"},"container-title":["Proceedings of the 30th IEEE\/ACM International Conference on Program Comprehension"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3524610.3527916","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3524610.3527916","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:09:53Z","timestamp":1750183793000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3524610.3527916"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,16]]},"references-count":48,"alternative-id":["10.1145\/3524610.3527916","10.1145\/3524610"],"URL":"https:\/\/doi.org\/10.1145\/3524610.3527916","relation":{},"subject":[],"published":{"date-parts":[[2022,5,16]]},"assertion":[{"value":"2022-10-20","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}