{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T17:40:11Z","timestamp":1755884411958,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":39,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,25]],"date-time":"2023-10-25T00:00:00Z","timestamp":1698192000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Indian Institute of Technology, Roorkee","award":["FIG-100874"],"award-info":[{"award-number":["FIG-100874"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,25]]},"DOI":"10.1145\/3639856.3639857","type":"proceedings-article","created":{"date-parts":[[2024,5,17]],"date-time":"2024-05-17T11:49:10Z","timestamp":1715946550000},"page":"1-9","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["LiBERTy: A Novel Model for Natural Language Understanding"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4511-1858","authenticated-orcid":false,"given":"Onkar","family":"Susladkar","sequence":"first","affiliation":[{"name":"Independent Researcher, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6442-0782","authenticated-orcid":false,"given":"Gayatri S","family":"Deshmukh","sequence":"additional","affiliation":[{"name":"Independent Researcher, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2908-993X","authenticated-orcid":false,"given":"Sparsh","family":"Mittal","sequence":"additional","affiliation":[{"name":"IIT Roorkee, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4433-7858","authenticated-orcid":false,"given":"Sai Chandra Teja","family":"R","sequence":"additional","affiliation":[{"name":"Independent Researcher, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3712-1784","authenticated-orcid":false,"given":"Rekha","family":"Singhal","sequence":"additional","affiliation":[{"name":"TCS Research, India"}]}],"member":"320","published-online":{"date-parts":[[2024,5,17]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2021. GoEmotions: A Dataset of Fine-Grained Emotions (Slide 26). https:\/\/bit.ly\/3LJdcnW."},{"key":"e_1_3_2_1_2_1","unstructured":"2022. BERT base model (uncased). https:\/\/huggingface.co\/bert-base-uncased."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1609\/icwsm.v15i1.18116"},{"key":"e_1_3_2_1_4_1","unstructured":"Dana Alon and Jeongwoo Ko. 2021. GoEmotions: A Dataset for Fine-Grained Emotion Classification. https:\/\/ai.googleblog.com\/2021\/10\/goemotions-dataset-for-fine-grained.html."},{"key":"e_1_3_2_1_5_1","volume-title":"COBERT: COVID-19 question answering system using BERT. Arabian journal for science and engineering","author":"Alzubi A","year":"2021","unstructured":"Jafar\u00a0A Alzubi, Rachna Jain, Anubhav Singh, Pritee Parwekar, and Meenu Gupta. 2021. COBERT: COVID-19 question answering system using BERT. Arabian journal for science and engineering (2021), 1\u201311."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2017.01.012"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W15-4319"},{"key":"e_1_3_2_1_8_1","first-page":"28","article-title":"Fully unsupervised word translation from cross-lingual word embeddings especially for healthcare professionals","volume":"13","author":"Chauhan Shweta","year":"2022","unstructured":"Shweta Chauhan, Shefali Saxena, and Philemon Daniel. 2022. Fully unsupervised word translation from cross-lingual word embeddings especially for healthcare professionals. International Journal of System Assurance Engineering and Management 13, 1 (2022), 28\u201337.","journal-title":"International Journal of System Assurance Engineering and Management"},{"key":"e_1_3_2_1_9_1","volume-title":"A Survey of Techniques for Optimizing Transformer Inference. Journal of Systems Architecture","author":"Krishna\u00a0Teja","year":"2023","unstructured":"Krishna\u00a0Teja Chitty-Venkata 2023. A Survey of Techniques for Optimizing Transformer Inference. Journal of Systems Architecture (2023)."},{"key":"e_1_3_2_1_10_1","volume-title":"DistillBERT, RoBERTa, XLNet and ELECTRA. arXiv preprint arXiv:2104.02041","author":"Cortiz Diogo","year":"2021","unstructured":"Diogo Cortiz. 2021. Exploring Transformers in Emotion Recognition: a comparison of BERT, DistillBERT, RoBERTa, XLNet and ELECTRA. arXiv preprint arXiv:2104.02041 (2021)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P18-3006"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.372"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00482"},{"key":"e_1_3_2_1_14_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ASRU.2013.6707742"},{"key":"e_1_3_2_1_16_1","volume-title":"BERT transformer model for detecting Arabic GPT2 auto-generated tweets. arXiv preprint arXiv:2101.09345","author":"Harrag Fouzi","year":"2021","unstructured":"Fouzi Harrag, Maria Debbah, Kareem Darwish, and Ahmed Abdelali. 2021. BERT transformer model for detecting Arabic GPT2 auto-generated tweets. arXiv preprint arXiv:2101.09345 (2021)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3041021.3054151"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00300"},{"key":"e_1_3_2_1_19_1","volume-title":"FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimedia tools and applications 80, 8","author":"Kaliyar Rohit\u00a0Kumar","year":"2021","unstructured":"Rohit\u00a0Kumar Kaliyar, Anurag Goswami, and Pratik Narang. 2021. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimedia tools and applications 80, 8 (2021), 11765\u201311788."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/2631775.2631824"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/R10-HTC47129.2019.9042443"},{"key":"e_1_3_2_1_22_1","first-page":"282","article-title":"Automated ontology generation from a plain text using statistical and NLP techniques","volume":"7","author":"Kumar Naresh","year":"2016","unstructured":"Naresh Kumar, Minakshi Kumar, and Manjeet Singh. 2016. Automated ontology generation from a plain text using statistical and NLP techniques. International Journal of System Assurance Engineering and Management 7, 1 (2016), 282\u2013293.","journal-title":"International Journal of System Assurance Engineering and Management"},{"key":"e_1_3_2_1_23_1","volume-title":"Albert: A lite BERT for self-supervised learning of language representations. ICLR","author":"Lan Zhenzhong","year":"2020","unstructured":"Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2020. Albert: A lite BERT for self-supervised learning of language representations. ICLR (2020)."},{"key":"e_1_3_2_1_24_1","volume-title":"RoBERTa: A robustly optimized bert pretraining approach. ICLR","author":"Liu Yinhan","year":"2020","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2020. RoBERTa: A robustly optimized bert pretraining approach. ICLR (2020)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICoICT49345.2020.9166407"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.sysarc.2020.101839"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1202"},{"key":"e_1_3_2_1_28_1","volume-title":"Named Entity Recognition in Tweets: An Experimental Study. Conference on Empirical Methods in Natural Language Processing","author":"Ritter Alan","year":"2011","unstructured":"Alan Ritter, Sam Clark, Mausam, and Oren Etzioni. 2011. Named Entity Recognition in Tweets: An Experimental Study. Conference on Empirical Methods in Natural Language Processing (2011), 1524\u20131534."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2806416.2806485"},{"key":"e_1_3_2_1_30_1","volume-title":"An analytical study of information extraction from unstructured and multidimensional big data. Artificial intelligence Proceedings of the National Assembly of the Society","author":"Sakaki Takeshi","year":"2019","unstructured":"Takeshi Sakaki, Kumiko Hara, Mitsuo Yoshida, Fujio Toriumi, Kosuke Shinoda, Satoshi Awahara, Kazuhiro Kazama, and Itsuki Noda. 2019. An analytical study of information extraction from unstructured and multidimensional big data. Artificial intelligence Proceedings of the National Assembly of the Society (2019)."},{"key":"e_1_3_2_1_31_1","volume-title":"Case Study of Text Mining on Tweets on Twitter -Application to Understanding the Current Situation in Disaster Areas in the Event of a Large-Scale Disaster","author":"Sakamaki Eiichi","year":"2014","unstructured":"Eiichi Sakamaki, Etsuko Kamei, Nobuhiro Uji, Naoki Yoshinaga, and Yu Kirenkawa. 2014-2015. Case Study of Text Mining on Tweets on Twitter -Application to Understanding the Current Situation in Disaster Areas in the Event of a Large-Scale Disaster. Japan Journal of Japan Industrial Management Association 65 (2014-2015), 39\u201350."},{"key":"e_1_3_2_1_32_1","volume-title":"NeurIPS workshop","author":"Sanh Victor","year":"2019","unstructured":"Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. NeurIPS workshop (2019)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.3390\/fi13070163"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6428"},{"key":"e_1_3_2_1_35_1","volume-title":"TPFNet: A Novel Text In-painting Transformer for Text Removal. In IEEE International Conference on Document Analysis and Recognition (ICDAR).","author":"Susladkar Onkar","year":"2023","unstructured":"Onkar Susladkar 2023. TPFNet: A Novel Text In-painting Transformer for Text Removal. In IEEE International Conference on Document Analysis and Recognition (ICDAR)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV56688.2023.00521"},{"key":"e_1_3_2_1_37_1","volume-title":"Breaking News Detection and Tracking in Twitter. 2010 IEEE\/WIC\/ACM International Conference on Web Intelligence and Intelligent Agent Technology 3","author":"Swit\u00a0Phuvipadawat Tsuyoshi\u00a0Murata","year":"2010","unstructured":"Tsuyoshi\u00a0Murata Swit\u00a0Phuvipadawat. 2010. Breaking News Detection and Tracking in Twitter. 2010 IEEE\/WIC\/ACM International Conference on Web Intelligence and Intelligent Agent Technology 3 (2010), 120\u2013123."},{"key":"e_1_3_2_1_38_1","volume-title":"TASTEset\u2013Recipe Dataset and Food Entities Recognition Benchmark. arXiv preprint arXiv:2204.07775","author":"Wr\u00f3blewska Ania","year":"2022","unstructured":"Ania Wr\u00f3blewska, Agnieszka Kaliska, Maciej Paw\u0142owski, Dawid Wi\u015bniewski, Witold Sosnowski, and Agnieszka \u0141awrynowicz. 2022. TASTEset\u2013Recipe Dataset and Food Entities Recognition Benchmark. arXiv preprint arXiv:2204.07775 (2022)."},{"key":"e_1_3_2_1_39_1","volume-title":"Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32","author":"Yang Zhilin","year":"2019","unstructured":"Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ\u00a0R Salakhutdinov, and Quoc\u00a0V Le. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32 (2019)."}],"event":{"name":"AIMLSystems 2023: The Third International Conference on Artificial Intelligence and Machine Learning Systems","acronym":"AIMLSystems 2023","location":"Bangalore India"},"container-title":["The Third International Conference on Artificial Intelligence and Machine Learning Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3639856.3639857","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3639856.3639857","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T16:59:12Z","timestamp":1755881952000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3639856.3639857"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,25]]},"references-count":39,"alternative-id":["10.1145\/3639856.3639857","10.1145\/3639856"],"URL":"https:\/\/doi.org\/10.1145\/3639856.3639857","relation":{},"subject":[],"published":{"date-parts":[[2023,10,25]]},"assertion":[{"value":"2024-05-17","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}