{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:15:08Z","timestamp":1750220108328,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":47,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,5,21]],"date-time":"2022-05-21T00:00:00Z","timestamp":1653091200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,5,21]]},"DOI":"10.1145\/3528588.3528652","type":"proceedings-article","created":{"date-parts":[[2023,2,1]],"date-time":"2023-02-01T23:05:55Z","timestamp":1675292755000},"page":"1-8","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":10,"title":["Unsupervised extreme multi label classification of stack overflow posts"],"prefix":"10.1145","author":[{"given":"Peter","family":"Devine","sequence":"first","affiliation":[{"name":"The University of Auckland, Auckland, New Zealand"}]},{"given":"Kelly","family":"Blincoe","sequence":"additional","affiliation":[{"name":"The University of Auckland, Auckland, New Zealand"}]}],"member":"320","published-online":{"date-parts":[[2023,2]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.infsof.2017.04.005"},{"key":"e_1_3_2_1_2_1","first-page":"e2309","article-title":"Conceptualising, extracting and analysing requirements arguments in users' forums: The CrowdRE-Arg framework","volume":"32","author":"Khan Javed Ali","year":"2020","unstructured":"Javed Ali Khan , Lin Liu , Lijie Wen , and Raian Ali . 2020 . Conceptualising, extracting and analysing requirements arguments in users' forums: The CrowdRE-Arg framework . Journal of Software: Evolution and Process 32 , 12 (2020), e2309 . Javed Ali Khan, Lin Liu, Lijie Wen, and Raian Ali. 2020. Conceptualising, extracting and analysing requirements arguments in users' forums: The CrowdRE-Arg framework. Journal of Software: Evolution and Process 32, 12 (2020), e2309.","journal-title":"Journal of Software: Evolution and Process"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSR.2019.00052"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10664-012-9231-y"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D15-1075"},{"key":"e_1_3_2_1_6_1","volume-title":"Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation. arXiv preprint arXiv:1708.00055","author":"Cer Daniel","year":"2017","unstructured":"Daniel Cer , Mona Diab , Eneko Agirre , Inigo Lopez-Gazpio , and Lucia Specia . 2017. Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation. arXiv preprint arXiv:1708.00055 ( 2017 ). Daniel Cer, Mona Diab, Eneko Agirre, Inigo Lopez-Gazpio, and Lucia Specia. 2017. Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation. arXiv preprint arXiv:1708.00055 (2017)."},{"key":"e_1_3_2_1_7_1","volume-title":"Noah Constant, Mario Guajardo-C\u00e9spedes, Steve Yuan, Chris Tar, et al.","author":"Cer Daniel","year":"2018","unstructured":"Daniel Cer , Yinfei Yang , Sheng-yi Kong, Nan Hua , Nicole Limtiaco , Rhomni St John , Noah Constant, Mario Guajardo-C\u00e9spedes, Steve Yuan, Chris Tar, et al. 2018 . Universal sentence encoder. arXiv preprint arXiv:1803.11175 (2018). Daniel Cer, Yinfei Yang, Sheng-yi Kong, Nan Hua, Nicole Limtiaco, Rhomni St John, Noah Constant, Mario Guajardo-C\u00e9spedes, Steve Yuan, Chris Tar, et al. 2018. Universal sentence encoder. arXiv preprint arXiv:1803.11175 (2018)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/2568225.2568263"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.207"},{"key":"e_1_3_2_1_10_1","volume-title":"Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies. 665--669","author":"Coster William","year":"2011","unstructured":"William Coster and David Kauchak . 2011 . Simple English Wikipedia: a new text simplification task . In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies. 665--669 . William Coster and David Kauchak. 2011. Simple English Wikipedia: a new text simplification task. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies. 665--669."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1007\/s42979-020-00427-1"},{"key":"e_1_3_2_1_12_1","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","volume":"1","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2019 . BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding . In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies , Volume 1 (Long and Short Papers). 4171--4186. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 4171--4186."},{"key":"e_1_3_2_1_13_1","volume-title":"A Transfer Learning Approach for Dialogue Act Classification of GitHub Issue Comments. arXiv preprint arXiv:2011.04867","author":"Enayet Ayesha","year":"2020","unstructured":"Ayesha Enayet and Gita Sukthankar . 2020. A Transfer Learning Approach for Dialogue Act Classification of GitHub Issue Comments. arXiv preprint arXiv:2011.04867 ( 2020 ). Ayesha Enayet and Gita Sukthankar. 2020. A Transfer Learning Approach for Dialogue Act Classification of GitHub Issue Comments. arXiv preprint arXiv:2011.04867 (2020)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/2623330.2623677"},{"key":"e_1_3_2_1_15_1","volume-title":"A Pre-trained Language Representation Model for Software Engineering. arXiv preprint arXiv:2112.00699","author":"De Bortoli F\u00e1vero Eliane Maria","year":"2021","unstructured":"Eliane Maria De Bortoli F\u00e1vero and Dalcimar Casanova . 2021. BERT_SE : A Pre-trained Language Representation Model for Software Engineering. arXiv preprint arXiv:2112.00699 ( 2021 ). Eliane Maria De Bortoli F\u00e1vero and Dalcimar Casanova. 2021. BERT_SE: A Pre-trained Language Representation Model for Software Engineering. arXiv preprint arXiv:2112.00699 (2021)."},{"key":"e_1_3_2_1_16_1","volume-title":"Language-agnostic bert sentence embedding. arXiv preprint arXiv:2007.01852","author":"Feng Fangxiaoyu","year":"2020","unstructured":"Fangxiaoyu Feng , Yinfei Yang , Daniel Cer , Naveen Arivazhagan , and Wei Wang . 2020. Language-agnostic bert sentence embedding. arXiv preprint arXiv:2007.01852 ( 2020 ). Fangxiaoyu Feng, Yinfei Yang, Daniel Cer, Naveen Arivazhagan, and Wei Wang. 2020. Language-agnostic bert sentence embedding. arXiv preprint arXiv:2007.01852 (2020)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.139"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ROPEC.2015.7395121"},{"key":"e_1_3_2_1_19_1","volume-title":"GraphCodeBERT: Pre-training Code Representations with Data Flow. In International Conference on Learning Representations.","author":"Guo Daya","year":"2020","unstructured":"Daya Guo , Shuo Ren , Shuai Lu , Zhangyin Feng , Duyu Tang , LIU Shujie , Long Zhou , Nan Duan , Alexey Svyatkovskiy , Shengyu Fu , 2020 . GraphCodeBERT: Pre-training Code Representations with Data Flow. In International Conference on Learning Representations. Daya Guo, Shuo Ren, Shuai Lu, Zhangyin Feng, Duyu Tang, LIU Shujie, Long Zhou, Nan Duan, Alexey Svyatkovskiy, Shengyu Fu, et al. 2020. GraphCodeBERT: Pre-training Code Representations with Data Flow. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467426"},{"key":"e_1_3_2_1_21_1","volume-title":"Keyword extraction and semantic tag prediction. unpublished(http:\/\/cs229.stanford.edu\/proj2013\/FangHong-Keyword%20Extraction%20and%20Semantic%20Tag%20Prediction.pdf)","author":"Hong James","year":"2013","unstructured":"James Hong and Michael Fang . 2013. Keyword extraction and semantic tag prediction. unpublished(http:\/\/cs229.stanford.edu\/proj2013\/FangHong-Keyword%20Extraction%20and%20Semantic%20Tag%20Prediction.pdf) ( 2013 ). James Hong and Michael Fang. 2013. Keyword extraction and semantic tag prediction. unpublished(http:\/\/cs229.stanford.edu\/proj2013\/FangHong-Keyword%20Extraction%20and%20Semantic%20Tag%20Prediction.pdf) (2013)."},{"key":"e_1_3_2_1_22_1","volume-title":"CodeSearchNet challenge: Evaluating the state of semantic code search. arXiv preprint arXiv:1909.09436","author":"Husain Hamel","year":"2019","unstructured":"Hamel Husain , Ho-Hsiang Wu , Tiferet Gazit , Miltiadis Allamanis , and Marc Brockschmidt . 2019. CodeSearchNet challenge: Evaluating the state of semantic code search. arXiv preprint arXiv:1909.09436 ( 2019 ). Hamel Husain, Ho-Hsiang Wu, Tiferet Gazit, Miltiadis Allamanis, and Marc Brockschmidt. 2019. CodeSearchNet challenge: Evaluating the state of semantic code search. arXiv preprint arXiv:1909.09436 (2019)."},{"volume-title":"8th ACM IKDD CODS and 26th COMAD. 228--232.","author":"Jain Yash","key":"e_1_3_2_1_23_1","unstructured":"Yash Jain and Anurag Roy . 2021. Distributed representation of tags for Active Zero Shot learning . In 8th ACM IKDD CODS and 26th COMAD. 228--232. Yash Jain and Anurag Roy. 2021. Distributed representation of tags for Active Zero Shot learning. In 8th ACM IKDD CODS and 26th COMAD. 228--232."},{"key":"e_1_3_2_1_24_1","unstructured":"Ryan Kiros Yukun Zhu Russ R Salakhutdinov Richard Zemel Raquel Urtasun Antonio Torralba and Sanja Fidler. 2015. Skip-thought vectors. In Advances in neural information processing systems. 3294--3302.  Ryan Kiros Yukun Zhu Russ R Salakhutdinov Richard Zemel Raquel Urtasun Antonio Torralba and Sanja Fidler. 2015. Skip-thought vectors. In Advances in neural information processing systems. 3294--3302."},{"key":"e_1_3_2_1_25_1","volume-title":"Technical report, EECS Department","author":"Kuo Darren","year":"2011","unstructured":"Darren Kuo . 2011. On word prediction methods. Technical report , Technical report, EECS Department ( 2011 ). Darren Kuo. 2011. On word prediction methods. Technical report, Technical report, EECS Department (2011)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/SANER.2018.8330202"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.447"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00766-016-0251-9"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/1978942.1979366"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSM.2012.6405249"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3380954"},{"key":"e_1_3_2_1_32_1","volume-title":"Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al.","author":"Radford Alec","year":"2021","unstructured":"Alec Radford , Jong Wook Kim , Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021 . Learning transferable visual models from natural language supervision. arXiv preprint arXiv:2103.00020 (2021). Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. arXiv preprint arXiv:2103.00020 (2021)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P18-2124"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1410"},{"key":"e_1_3_2_1_35_1","volume-title":"a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108","author":"Sanh Victor","year":"2019","unstructured":"Victor Sanh , Lysandre Debut , Julien Chaumond , and Thomas Wolf . 2019. DistilBERT , a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 ( 2019 ). Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019)."},{"key":"e_1_3_2_1_36_1","volume-title":"Proceedings of the LWDA 2017 Workshops: KDML, FGWM, IR, and FGDB.","author":"Schuster Sebastian","year":"2017","unstructured":"Sebastian Schuster , Wanying Zhu , and Yiying Cheng . 2017 . Predicting tags for stackoverflow questions . In Proceedings of the LWDA 2017 Workshops: KDML, FGWM, IR, and FGDB. Sebastian Schuster, Wanying Zhu, and Yiying Cheng. 2017. Predicting tags for stackoverflow questions. In Proceedings of the LWDA 2017 Workshops: KDML, FGWM, IR, and FGDB."},{"key":"e_1_3_2_1_37_1","volume-title":"Mpnet: Masked and permuted pre-training for language understanding. arXiv preprint arXiv:2004.09297","author":"Song Kaitao","year":"2020","unstructured":"Kaitao Song , Xu Tan , Tao Qin , Jianfeng Lu , and Tie-Yan Liu . 2020 . Mpnet: Masked and permuted pre-training for language understanding. arXiv preprint arXiv:2004.09297 (2020). Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, and Tie-Yan Liu. 2020. Mpnet: Masked and permuted pre-training for language understanding. arXiv preprint arXiv:2004.09297 (2020)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/RE.2019.00014"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/2597008.2597799"},{"key":"e_1_3_2_1_40_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all youneed. In Advances in neural information processing systems. 5998--6008.  Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all youneed. In Advances in neural information processing systems. 5998--6008."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2016.0060"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/2480362.2480557"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10664-017-9533-1"},{"key":"e_1_3_2_1_44_1","volume-title":"Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. arXiv preprint arXiv:2002.10957","author":"Wang Wenhui","year":"2020","unstructured":"Wenhui Wang , Furu Wei , Li Dong , Hangbo Bao , Nan Yang , and Ming Zhou . 2020 . Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. arXiv preprint arXiv:2002.10957 (2020). Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, and Ming Zhou. 2020. Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. arXiv preprint arXiv:2002.10957 (2020)."},{"key":"e_1_3_2_1_45_1","volume-title":"Post2vec: Learning distributed representations of Stack Overflow posts","author":"Xu Bowen","year":"2021","unstructured":"Bowen Xu , Thong Hoang , Abhishek Sharma , Chengran Yang , Xin Xia , and David Lo. 2021. Post2vec: Learning distributed representations of Stack Overflow posts . IEEE Transactions on Software Engineering ( 2021 ). Bowen Xu, Thong Hoang, Abhishek Sharma, Chengran Yang, Xin Xia, and David Lo. 2021. Post2vec: Learning distributed representations of Stack Overflow posts. IEEE Transactions on Software Engineering (2021)."},{"key":"e_1_3_2_1_46_1","volume-title":"Is deep learning better than traditional approaches in tag recommendation for software information sites? Information and software technology 109","author":"Zhou Pingyi","year":"2019","unstructured":"Pingyi Zhou , Jin Liu , Xiao Liu , Zijiang Yang , and John Grundy . 2019. Is deep learning better than traditional approaches in tag recommendation for software information sites? Information and software technology 109 ( 2019 ), 1--13. Pingyi Zhou, Jin Liu, Xiao Liu, Zijiang Yang, and John Grundy. 2019. Is deep learning better than traditional approaches in tag recommendation for software information sites? Information and software technology 109 (2019), 1--13."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.11"}],"event":{"name":"ICSE '22: 44th International Conference on Software Engineering","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering","IEEE CS"],"location":"Pittsburgh Pennsylvania","acronym":"ICSE '22"},"container-title":["Proceedings of the 1st International Workshop on Natural Language-based Software Engineering"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3528588.3528652","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3528588.3528652","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:10:34Z","timestamp":1750183834000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3528588.3528652"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,21]]},"references-count":47,"alternative-id":["10.1145\/3528588.3528652","10.1145\/3528588"],"URL":"https:\/\/doi.org\/10.1145\/3528588.3528652","relation":{},"subject":[],"published":{"date-parts":[[2022,5,21]]},"assertion":[{"value":"2023-02-01","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}