{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T14:10:09Z","timestamp":1755871809066,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":38,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,1,4]],"date-time":"2024-01-04T00:00:00Z","timestamp":1704326400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,1,4]]},"DOI":"10.1145\/3632410.3632454","type":"proceedings-article","created":{"date-parts":[[2024,1,3]],"date-time":"2024-01-03T18:15:16Z","timestamp":1704305716000},"page":"454-458","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Expertise Identification Using Transformers"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-1204-7521","authenticated-orcid":false,"given":"Sreekanth","family":"T M","sequence":"first","affiliation":[{"name":"Homi Bhabha National Institute, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7167-3435","authenticated-orcid":false,"given":"Ajith","family":"K John","sequence":"additional","affiliation":[{"name":"Bhabha Atomic Research Centre, India"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-0952-1093","authenticated-orcid":false,"given":"Rohitashva","family":"Sharma","sequence":"additional","affiliation":[{"name":"Bhabha Atomic Research Centre, India"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-2658-7472","authenticated-orcid":false,"given":"Prathamesh","family":"Berde","sequence":"additional","affiliation":[{"name":"Bhabha Atomic Research Centre, India"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-3419-3903","authenticated-orcid":false,"given":"C S R C","family":"Murthy","sequence":"additional","affiliation":[{"name":"Bhabha Atomic Research Centre, India"}]}],"member":"320","published-online":{"date-parts":[[2024,1,4]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"[n. d.]. Google Job Skills. https:\/\/www.kaggle.com\/datasets\/niyamatalmass\/google-job-skills"},{"key":"e_1_3_2_1_2_1","unstructured":"[n. d.]. HuggingFace. https:\/\/huggingface.co\/"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P15-1034"},{"key":"e_1_3_2_1_4_1","volume-title":"Simple unsupervised keyphrase extraction using sentence embeddings. arXiv preprint arXiv:1801.04470 abs\/1801.04470","author":"Bennani-Smires Kamil","year":"2018","unstructured":"Kamil Bennani-Smires, Claudiu Musat, Andreea Hossmann, Michael Baeriswyl, and Martin Jaggi. 2018. Simple unsupervised keyphrase extraction using sentence embeddings. arXiv preprint arXiv:1801.04470 abs\/1801.04470 (2018)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSA.2018.00018"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3440755"},{"key":"e_1_3_2_1_7_1","volume-title":"Expertseer: a keyphrase based expert recommender for digital libraries. arXiv preprint arXiv:1511.02058 abs\/1511.02058","author":"Chen Hung-Hsuan","year":"2015","unstructured":"Hung-Hsuan Chen, Alexander\u00a0G Ororbia\u00a0II, and C\u00a0Lee Giles. 2015. Expertseer: a keyphrase based expert recommender for digital libraries. arXiv preprint arXiv:1511.02058 abs\/1511.02058 (2015)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.3390\/app13137753"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3466933.3466984"},{"volume-title":"Working knowledge: How organizations manage what they know","author":"Davenport H","key":"e_1_3_2_1_10_1","unstructured":"Thomas\u00a0H Davenport and Laurence Prusak. 1998. Working knowledge: How organizations manage what they know. Harvard Business Press."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1423"},{"key":"e_1_3_2_1_12_1","volume-title":"Proceedings of the The 18th Annual Workshop of the Australasian Language Technology Association. 121\u2013126","author":"Ferraro Gabriela","year":"2020","unstructured":"Gabriela Ferraro and Hanna Suominen. 2020. Transformer semantic parsing. In Proceedings of the The 18th Annual Workshop of the Australasian Language Technology Association. 121\u2013126."},{"key":"e_1_3_2_1_13_1","volume-title":"An Introduction to Expertise at Work: Current and Emerging Trends. Identifying and Measuring Expertise in Organizations","author":"Grenier S","year":"2021","unstructured":"Robin\u00a0S Grenier and Marie-Line Germain. 2021. An Introduction to Expertise at Work: Current and Emerging Trends. Identifying and Measuring Expertise in Organizations (2021), 57\u201369."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/P14-1119"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447772"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.457"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1356"},{"key":"e_1_3_2_1_18_1","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics. Association for Computational Linguistics, 2284\u20132293","author":"Koncel-Kedziorski Rik","year":"2019","unstructured":"Rik Koncel-Kedziorski, Dhanush Bekal, Yi Luan, Maria Lapata, and Hannaneh Hajishirzi. 2019. Text Generation from Knowledge Graphs with Graph Transformers. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics. Association for Computational Linguistics, 2284\u20132293."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.23919\/PICMET.2017.8125418"},{"volume-title":"Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)","author":"Kovaleva Olga","key":"e_1_3_2_1_20_1","unstructured":"Olga Kovaleva, Alexey Romanov, Anna Rogers, and Anna Rumshisky. 2019. Revealing the Dark Secrets of BERT. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, 4365\u20134374. https:\/\/aclanthology.org\/D19-1445"},{"key":"e_1_3_2_1_21_1","volume-title":"Learning rich representation of keyphrases from text. arXiv preprint arXiv:2112.08547","author":"Kulkarni Mayank","year":"2021","unstructured":"Mayank Kulkarni, Debanjan Mahata, Ravneet Arora, and Rajarshi Bhowmik. 2021. Learning rich representation of keyphrases from text. arXiv preprint arXiv:2112.08547 (2021)."},{"key":"e_1_3_2_1_22_1","volume-title":"Neural Keyphrase Generation: Analysis and Evaluation. arXiv preprint arXiv:2304.13883","author":"Kundu Tuhin","year":"2023","unstructured":"Tuhin Kundu, Jishnu\u00a0Ray Chowdhury, and Cornelia Caragea. 2023. Neural Keyphrase Generation: Analysis and Evaluation. arXiv preprint arXiv:2304.13883 (2023)."},{"key":"e_1_3_2_1_23_1","volume-title":"Natural language to code using transformers. arXiv preprint arXiv:2202.00367","author":"Kusupati Uday","year":"2022","unstructured":"Uday Kusupati and Venkata Ravi\u00a0Teja Ailavarapu. 2022. Natural language to code using transformers. arXiv preprint arXiv:2202.00367 (2022)."},{"key":"e_1_3_2_1_24_1","volume-title":"Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461","author":"Lewis Mike","year":"2019","unstructured":"Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. 2019. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019)."},{"key":"e_1_3_2_1_25_1","volume-title":"Text summarization with pretrained encoders. arXiv preprint arXiv:1908.08345","author":"Liu Yang","year":"2019","unstructured":"Yang Liu and Mirella Lapata. 2019. Text summarization with pretrained encoders. arXiv preprint arXiv:1908.08345 (2019)."},{"key":"e_1_3_2_1_26_1","volume-title":"A unified approach to interpreting model predictions. Advances in neural information processing systems 30","author":"Lundberg M","year":"2017","unstructured":"Scott\u00a0M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_27_1","volume-title":"Proceedings of the 2004 conference on empirical methods in natural language processing. 404\u2013411","author":"Mihalcea Rada","year":"2004","unstructured":"Rada Mihalcea and Paul Tarau. 2004. Textrank: Bringing order into text. In Proceedings of the 2004 conference on empirical methods in natural language processing. 404\u2013411."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10489-022-04052-8"},{"key":"e_1_3_2_1_29_1","volume-title":"Transformer-based models for question answering on COVID19. arXiv preprint arXiv:2101.11432","author":"Ngai Hillary","year":"2021","unstructured":"Hillary Ngai, Yoona Park, John Chen, and Mahboobeh Parsapoor. 2021. Transformer-based models for question answering on COVID19. arXiv preprint arXiv:2101.11432 (2021)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i15.17605"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.748"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.5555\/3455716.3455856"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_34_1","volume-title":"Patternrank: leveraging pretrained language models and part of speech for unsupervised keyphrase extraction. arXiv preprint arXiv:2210.05245","author":"Schopf Tim","year":"2022","unstructured":"Tim Schopf, Simon Klimek, and Florian Matthes. 2022. Patternrank: leveraging pretrained language models and part of speech for unsupervised keyphrase extraction. arXiv preprint arXiv:2210.05245 (2022)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2007.01.010"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1207\/S15327744JOCE1301_1"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3361719"},{"key":"e_1_3_2_1_38_1","volume-title":"Bertscore: Evaluating text generation with bert. arXiv preprint arXiv:1904.09675","author":"Zhang Tianyi","year":"2019","unstructured":"Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian\u00a0Q Weinberger, and Yoav Artzi. 2019. Bertscore: Evaluating text generation with bert. arXiv preprint arXiv:1904.09675 (2019)."}],"event":{"name":"CODS-COMAD 2024: 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD)","acronym":"CODS-COMAD 2024","location":"Bangalore India"},"container-title":["Proceedings of the 7th Joint International Conference on Data Science &amp; Management of Data (11th ACM IKDD CODS and 29th COMAD)"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3632410.3632454","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3632410.3632454","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T13:36:41Z","timestamp":1755869801000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3632410.3632454"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,4]]},"references-count":38,"alternative-id":["10.1145\/3632410.3632454","10.1145\/3632410"],"URL":"https:\/\/doi.org\/10.1145\/3632410.3632454","relation":{},"subject":[],"published":{"date-parts":[[2024,1,4]]},"assertion":[{"value":"2024-01-04","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}