{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T16:50:08Z","timestamp":1772902208877,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":44,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,1,4]],"date-time":"2024-01-04T00:00:00Z","timestamp":1704326400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,1,4]]},"DOI":"10.1145\/3632410.3632463","type":"proceedings-article","created":{"date-parts":[[2024,1,3]],"date-time":"2024-01-03T18:15:16Z","timestamp":1704305716000},"page":"136-154","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["A Comprehensive Analysis of Adapter Efficiency"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-8672-6724","authenticated-orcid":false,"given":"Nandini","family":"Mundra","sequence":"first","affiliation":[{"name":"Indian Institute Of Technology, Madras, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4248-646X","authenticated-orcid":false,"given":"Sumanth","family":"Doddapaneni","sequence":"additional","affiliation":[{"name":"Indian Institute Of Technology, Madras, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0664-3421","authenticated-orcid":false,"given":"Raj","family":"Dabre","sequence":"additional","affiliation":[{"name":"National Institute of Information and Communications Technology, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-3143-9875","authenticated-orcid":false,"given":"Anoop","family":"Kunchukuttan","sequence":"additional","affiliation":[{"name":"Microsoft, India"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-0857-2842","authenticated-orcid":false,"given":"Ratish","family":"Puduppully","sequence":"additional","affiliation":[{"name":"Institute for Infocomm Research I2R, A*STAR, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-3687-9922","authenticated-orcid":false,"given":"Mitesh M.","family":"Khapra","sequence":"additional","affiliation":[{"name":"Indian Institute Of Technology, Madras, India"}]}],"member":"320","published-online":{"date-parts":[[2024,1,4]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1165"},{"key":"e_1_3_2_1_2_1","volume-title":"Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models. Transactions on Machine Learning Research","author":"BIG","year":"2023","unstructured":"BIG bench authors. 2023. Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models. Transactions on Machine Learning Research (2023). https:\/\/openreview.net\/forum?id=uyTL5Bvosj"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1016\/b978-1-55860-307-3.50012-5"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.168"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.747"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1423"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2212.05409"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1016\/S1364-6613(99)01294-2"},{"key":"e_1_3_2_1_9_1","volume-title":"SuperShaper: Task-Agnostic Super Pre-training of BERT Models with Variable Hidden Dimensions. CoRR abs\/2110.04711","author":"Ganesan Vinod","year":"2021","unstructured":"Vinod Ganesan, Gowtham Ramesh, and Pratyush Kumar. 2021. SuperShaper: Task-Agnostic Super Pre-training of BERT Models with Variable Hidden Dimensions. CoRR abs\/2110.04711 (2021). arXiv:2110.04711https:\/\/arxiv.org\/abs\/2110.04711"},{"key":"e_1_3_2_1_10_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=0RDcd5Axok","author":"He Junxian","year":"2022","unstructured":"Junxian He, Chunting Zhou, Xuezhe Ma, Taylor Berg-Kirkpatrick, and Graham Neubig. 2022. Towards a Unified View of Parameter-Efficient Transfer Learning. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=0RDcd5Axok"},{"key":"e_1_3_2_1_11_1","volume-title":"Proceedings of the 36th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a097)","author":"Houlsby Neil","year":"2019","unstructured":"Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin De\u00a0Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly. 2019. Parameter-Efficient Transfer Learning for NLP. In Proceedings of the 36th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol.\u00a097), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, 2790\u20132799. https:\/\/proceedings.mlr.press\/v97\/houlsby19a.html"},{"key":"e_1_3_2_1_12_1","volume-title":"LoRA: Low-Rank Adaptation of Large Language Models. In The Tenth International Conference on Learning Representations, ICLR 2022","author":"Hu J.","year":"2022","unstructured":"Edward\u00a0J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2022. LoRA: Low-Rank Adaptation of Large Language Models. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net. https:\/\/openreview.net\/forum?id=nZeVKeeFYf9"},{"key":"e_1_3_2_1_13_1","volume-title":"XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization. CoRR abs\/2003.11080","author":"Hu Junjie","year":"2020","unstructured":"Junjie Hu, Sebastian Ruder, Aditya Siddhant, Graham Neubig, Orhan Firat, and Melvin Johnson. 2020. XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization. CoRR abs\/2003.11080 (2020). arXiv:2003.11080https:\/\/arxiv.org\/abs\/2003.11080"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.445"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.369"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1611835114"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.coling-main.66"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2205.05638"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1441"},{"key":"e_1_3_2_1_21_1","volume-title":"RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR abs\/1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. CoRR abs\/1907.11692 (2019). arXiv:1907.11692http:\/\/arxiv.org\/abs\/1907.11692"},{"key":"e_1_3_2_1_22_1","volume-title":"Compacter: Efficient Low-Rank Hypercomplex Adapter Layers. In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021","author":"Mahabadi Rabeeh\u00a0Karimi","year":"2021","unstructured":"Rabeeh\u00a0Karimi Mahabadi, James Henderson, and Sebastian Ruder. 2021. Compacter: Efficient Low-Rank Hypercomplex Adapter Layers. In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, Marc\u2019Aurelio Ranzato, Alina Beygelzimer, Yann\u00a0N. Dauphin, Percy Liang, and Jennifer\u00a0Wortman Vaughan (Eds.). 1022\u20131035. https:\/\/proceedings.neurips.cc\/paper\/2021\/hash\/081be9fdff07f3bc808f935906ef70c0-Abstract.html"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1016\/S0079-7421(08)60536-8"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2212.10168"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"crossref","unstructured":"Niklas Muennighoff Thomas Wang Lintang Sutawika Adam Roberts Stella Biderman Teven\u00a0Le Scao M\u00a0Saiful Bari Sheng Shen Zheng-Xin Yong Hailey Schoelkopf Xiangru Tang Dragomir Radev Alham\u00a0Fikri Aji Khalid Almubarak Samuel Albanie Zaid Alyafeai Albert Webson Edward Raff and Colin Raffel. 2022. Crosslingual Generalization through Multitask Finetuning. arxiv:2211.01786\u00a0[cs.CL]","DOI":"10.18653\/v1\/2023.acl-long.891"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.mrl-1.11"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.eacl-main.39"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.7"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.617"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1264"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.626"},{"key":"e_1_3_2_1_32_1","volume-title":"An Overview of Multi-Task Learning in Deep Neural Networks. CoRR abs\/1706.05098","author":"Ruder Sebastian","year":"2017","unstructured":"Sebastian Ruder. 2017. An Overview of Multi-Task Learning in Deep Neural Networks. CoRR abs\/1706.05098 (2017). arXiv:1706.05098http:\/\/arxiv.org\/abs\/1706.05098"},{"key":"e_1_3_2_1_33_1","volume-title":"Multitask Prompted Training Enables Zero-Shot Task Generalization. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=9Vrb9D0WI4","author":"Sanh Victor","year":"2022","unstructured":"Victor Sanh, Albert Webson, Colin Raffel, Stephen Bach, Lintang Sutawika, Zaid Alyafeai, Antoine Chaffin, Arnaud Stiegler, Arun Raja, Manan Dey, M\u00a0Saiful Bari, Canwen Xu, Urmish Thakker, Shanya\u00a0Sharma Sharma, Eliza Szczechla, Taewoon Kim, Gunjan Chhablani, Nihal Nayak, Debajyoti Datta, Jonathan Chang, Mike Tian-Jian Jiang, Han Wang, Matteo Manica, Sheng Shen, Zheng\u00a0Xin Yong, Harshit Pandey, Rachel Bawden, Thomas Wang, Trishala Neeraj, Jos Rozen, Abheesht Sharma, Andrea Santilli, Thibault Fevry, Jason\u00a0Alan Fries, Ryan Teehan, Teven\u00a0Le Scao, Stella Biderman, Leo Gao, Thomas Wolf, and Alexander\u00a0M Rush. 2022. Multitask Prompted Training Enables Zero-Shot Task Generalization. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=9Vrb9D0WI4"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1454"},{"key":"e_1_3_2_1_35_1","first-page":"03","volume-title":"Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL","author":"F.","year":"2003","unstructured":"Erik\u00a0F. Tjong Kim\u00a0Sang and Fien De\u00a0Meulder. 2003. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition. In Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL 2003. 142\u2013147. https:\/\/aclanthology.org\/W03-0419"},{"key":"e_1_3_2_1_36_1","volume-title":"Advances in Neural Information Processing Systems, I.\u00a0Guyon, U.\u00a0Von Luxburg, S.\u00a0Bengio, H.\u00a0Wallach, R.\u00a0Fergus, S.\u00a0Vishwanathan, and R.\u00a0Garnett (Eds.). Vol.\u00a030. Curran Associates","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan\u00a0N Gomez, \u0141\u00a0ukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In Advances in Neural Information Processing Systems, I.\u00a0Guyon, U.\u00a0Von Luxburg, S.\u00a0Bengio, H.\u00a0Wallach, R.\u00a0Fergus, S.\u00a0Vishwanathan, and R.\u00a0Garnett (Eds.). Vol.\u00a030. Curran Associates, Inc.https:\/\/proceedings.neurips.cc\/paper\/2017\/file\/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2205.12647"},{"key":"e_1_3_2_1_38_1","unstructured":"Jason Wei Maarten Bosma Vincent\u00a0Y. Zhao Kelvin Guu Adams\u00a0Wei Yu Brian Lester Nan Du Andrew\u00a0M. Dai and Quoc\u00a0V. Le. 2021. Finetuned Language Models Are Zero-Shot Learners. arxiv:2109.01652\u00a0[cs.CL]"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N18-1101"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1382"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2212.09535"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2204.03508"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1163"}],"event":{"name":"CODS-COMAD 2024: 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD)","location":"Bangalore India","acronym":"CODS-COMAD 2024"},"container-title":["Proceedings of the 7th Joint International Conference on Data Science &amp; Management of Data (11th ACM IKDD CODS and 29th COMAD)"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3632410.3632463","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3632410.3632463","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T13:37:53Z","timestamp":1755869873000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3632410.3632463"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,4]]},"references-count":44,"alternative-id":["10.1145\/3632410.3632463","10.1145\/3632410"],"URL":"https:\/\/doi.org\/10.1145\/3632410.3632463","relation":{},"subject":[],"published":{"date-parts":[[2024,1,4]]},"assertion":[{"value":"2024-01-04","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}