{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T18:26:21Z","timestamp":1775067981548,"version":"3.50.1"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,9,1]],"date-time":"2024-09-01T00:00:00Z","timestamp":1725148800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T00:00:00Z","timestamp":1725840000000},"content-version":"vor","delay-in-days":8,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001554","name":"Massey University","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100001554","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["World Wide Web"],"published-print":{"date-parts":[[2024,9]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>\n                    Code-mixing and code-switching (CMCS) are prevalent phenomena observed in social media conversations and various other modes of communication. When developing applications such as sentiment analysers and hate-speech detectors that operate on this social media data, CMCS text poses challenges. Recent studies have demonstrated that prompt-based learning of pre-trained language models outperforms full fine-tuning across various tasks. Despite the growing interest in classifying CMCS text, the effectiveness of prompt-based learning for the task remains unexplored. This paper presents an extensive exploration of prompt-based learning for CMCS text classification and the first comprehensive analysis of the impact of the script on classifying CMCS text. Our study reveals that the performance in classifying CMCS text is significantly influenced by the inclusion of multiple scripts and the intensity of code-mixing. In response, we introduce a novel method,\n                    <jats:italic>Dynamic+AdapterPrompt<\/jats:italic>\n                    , which employs distinct models for each script, integrated with adapters. While DynamicPrompt captures the script-specific representation of the text, AdapterPrompt emphasizes capturing the task-oriented functionality. Our experiments on Sinhala-English, Kannada-English, and Hindi-English datasets for sentiment classification, hate-speech detection, and humour detection tasks show that our method outperforms strong fine-tuning baselines and basic prompting strategies.\n                  <\/jats:p>","DOI":"10.1007\/s11280-024-01302-2","type":"journal-article","created":{"date-parts":[[2024,9,9]],"date-time":"2024-09-09T06:03:08Z","timestamp":1725861788000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Use of prompt-based learning for code-mixed and code-switched text classification"],"prefix":"10.1007","volume":"27","author":[{"given":"Pasindu","family":"Udawatta","sequence":"first","affiliation":[]},{"given":"Indunil","family":"Udayangana","sequence":"additional","affiliation":[]},{"given":"Chathulanka","family":"Gamage","sequence":"additional","affiliation":[]},{"given":"Ravi","family":"Shekhar","sequence":"additional","affiliation":[]},{"given":"Surangika","family":"Ranathunga","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,9]]},"reference":[{"key":"1302_CR1","doi-asserted-by":"publisher","unstructured":"Bali, K., Sharma, J., Choudhury, M., Vyas, Y.: I am borrowing ya mixing? an analysis of English-Hindi code mixing in Facebook. In: Diab, M., Hirschberg, J., Fung, P., Solorio, T. (eds.) Proceedings of the First Workshop on Computational Approaches to Code Switching, pp. 116\u2013126. Association for Computational Linguistics, Doha, Qatar (2014). https:\/\/doi.org\/10.3115\/v1\/W14-3914 . https:\/\/aclanthology.org\/W14-3914","DOI":"10.3115\/v1\/W14-3914"},{"key":"1302_CR2","unstructured":"Gundapu, S., Mamidi, R.: Word level language identification in English Telugu code mixed data. In: Proceedings of the 32nd Pacific Asia Conference on Language, Information and Computation. Association for Computational Linguistics, Hong Kong (2018). https:\/\/aclanthology.org\/Y18-1021"},{"key":"1302_CR3","unstructured":"Zirker, K.A.H.: Intrasentential vs. intersentential code switching in early and late bilinguals (2007). https:\/\/api.semanticscholar.org\/CorpusID:60154198"},{"key":"1302_CR4","doi-asserted-by":"publisher","unstructured":"Hande, A., Puranik, K., Yasaswini, K., Priyadharshini, R., Thavareesan, S., Sampath, A., Shanmugavadivel, K., Thenmozhi, D., Chakravarthi, B.R.: Offensive Language Identification in Low-resourced Code-mixed Dravidian languages using Pseudo-labeling (2021). https:\/\/doi.org\/10.48550\/arXiv.2108.12177","DOI":"10.48550\/arXiv.2108.12177"},{"key":"1302_CR5","doi-asserted-by":"publisher","unstructured":"Srivastava, V., Singh, M.: Code-mixed nlg: Resources, metrics, and challenges. In: 5th Joint international conference on data science & management of data (9th ACM IKDD CODS and 27th COMAD). CODS-COMAD 2022, pp. 328\u2013332. Association for Computing Machinery, New York, NY, USA (2022). https:\/\/doi.org\/10.1145\/3493700.3493766","DOI":"10.1145\/3493700.3493766"},{"key":"1302_CR6","doi-asserted-by":"publisher","first-page":"1937","DOI":"10.1007\/s10115-022-01698-1","volume":"64","author":"H Rathnayake","year":"2022","unstructured":"Rathnayake, H., Sumanapala, J., Rukshani, R., Ranathunga, S.: Adapter-based fine-tuning of pre-trained multilingual language models for code-mixed and code-switched text classification. Knowl. Inf. Syst. 64, 1937\u20131966 (2022). https:\/\/doi.org\/10.1007\/s10115-022-01698-1","journal-title":"Knowl. Inf. Syst."},{"key":"1302_CR7","doi-asserted-by":"publisher","unstructured":"Krishnan, J., Anastasopoulos, A., Purohit, H., Rangwala, H.: Cross-lingual text classification of transliterated hindi and malayalam. In: 2022 IEEE International Conference on Big Data (Big Data), pp. 1850\u20131857. IEEE Computer Society, Los Alamitos, CA, USA (2022). https:\/\/doi.org\/10.1109\/BigData55660.2022.10021079","DOI":"10.1109\/BigData55660.2022.10021079"},{"key":"1302_CR8","doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171\u20134186. Association for Computational Linguistics, Minneapolis, Minnesota (2019). https:\/\/doi.org\/10.18653\/v1\/N19-1423. https:\/\/aclanthology.org\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"1302_CR9","doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171\u20134186. Association for Computational Linguistics, Minneapolis, Minnesota (2019). https:\/\/doi.org\/10.18653\/v1\/N19-1423. https:\/\/aclanthology.org\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"1302_CR10","doi-asserted-by":"publisher","first-page":"182","DOI":"10.1016\/j.aiopen.2022.11.003","volume":"3","author":"X Han","year":"2022","unstructured":"Han, X., Zhao, W., Ding, N., Liu, Z., Sun, M.: Ptr: Prompt tuning with rules for text classification. AI Open 3, 182\u2013192 (2022). https:\/\/doi.org\/10.1016\/j.aiopen.2022.11.003","journal-title":"AI Open"},{"key":"1302_CR11","doi-asserted-by":"publisher","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., Neubig, G.: Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9) (2023). https:\/\/doi.org\/10.1145\/3560815","DOI":"10.1145\/3560815"},{"key":"1302_CR12","doi-asserted-by":"publisher","unstructured":"Hu, S., Ding, N., Wang, H., Liu, Z., Wang, J., Li, J., Wu, W., Sun, M.: Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2225\u20132240. Association for Computational Linguistics, Dublin, Ireland (2022). https:\/\/doi.org\/10.18653\/v1\/2022.acl-long.158. https:\/\/aclanthology.org\/2022.acl-long.158","DOI":"10.18653\/v1\/2022.acl-long.158"},{"key":"1302_CR13","doi-asserted-by":"crossref","unstructured":"Tu, L., Xiong, C., Zhou, Y.: Prompt-tuning can be much better than fine-tuning on cross-lingual understanding with multilingual language models. In: Findings of the Association for Computational Linguistics: EMNLP 2022, pp. 5478\u20135485. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates (2022). https:\/\/aclanthology.org\/2022.findings-emnlp.401","DOI":"10.18653\/v1\/2022.findings-emnlp.401"},{"key":"1302_CR14","doi-asserted-by":"publisher","unstructured":"Zhao, M., Sch\u00fctze, H.: Discrete and soft prompting for multilingual models. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 8547\u20138555. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic (2021). https:\/\/doi.org\/10.18653\/v1\/2021.emnlp-main.672. https:\/\/aclanthology.org\/2021.emnlp-main.672","DOI":"10.18653\/v1\/2021.emnlp-main.672"},{"key":"1302_CR15","doi-asserted-by":"publisher","unstructured":"Karimi\u00a0Mahabadi, R., Zettlemoyer, L., Henderson, J., Mathias, L., Saeidi, M., Stoyanov, V., Yazdani, M.: Prompt-free and efficient few-shot learning with language models. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 3638\u20133652. Association for Computational Linguistics, Dublin, Ireland (2022). https:\/\/doi.org\/10.18653\/v1\/2022.acl-long.254. https:\/\/aclanthology.org\/2022.acl-long.254","DOI":"10.18653\/v1\/2022.acl-long.254"},{"key":"1302_CR16","doi-asserted-by":"crossref","unstructured":"Huang, L., Ma, S., Zhang, D., Wei, F., Wang, H.: Zero-shot cross-lingual transfer of prompt-based tuning with a unified multilingual prompt. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 11488\u201311497. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates (2022). https:\/\/aclanthology.org\/2022.emnlp-main.790","DOI":"10.18653\/v1\/2022.emnlp-main.790"},{"key":"1302_CR17","doi-asserted-by":"crossref","unstructured":"Fu, J., Ng, S.-K., Liu, P.: Polyglot prompt: Multilingual multitask prompt training. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 9919\u20139935. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates (2022). https:\/\/aclanthology.org\/2022.emnlp-main.674","DOI":"10.18653\/v1\/2022.emnlp-main.674"},{"key":"1302_CR18","doi-asserted-by":"publisher","unstructured":"Winata, G.I., Madotto, A., Lin, Z., Liu, R., Yosinski, J., Fung, P.: Language models are few-shot multilingual learners. In: Proceedings of the 1st Workshop on Multilingual Representation Learning, pp. 1\u201315. Association for Computational Linguistics, Punta Cana, Dominican Republic (2021). https:\/\/doi.org\/10.18653\/v1\/2021.mrl-1.1. https:\/\/aclanthology.org\/2021.mrl-1.1","DOI":"10.18653\/v1\/2021.mrl-1.1"},{"key":"1302_CR19","doi-asserted-by":"publisher","unstructured":"Chakravarthi, B.R., Priyadharshini, R., Muralidaran, V., Jose, N., Suryawanshi, S., Sherly, E., McCrae, J.P.: DravidianCodeMix: sentiment analysis and offensive language identification dataset for dravidian languages in code-mixed text 56(3), 765\u2013806 (2022). https:\/\/doi.org\/10.1007\/s10579-022-09583-7","DOI":"10.1007\/s10579-022-09583-7"},{"key":"1302_CR20","doi-asserted-by":"publisher","unstructured":"Joshi, P., Santy, S., Budhiraja, A., Bali, K., Choudhury, M.: The state and fate of linguistic diversity and inclusion in the NLP world, pp. 6282\u20136293. Association for Computational Linguistics, Online (2020). https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.560. https:\/\/aclanthology.org\/2020.acl-main.560","DOI":"10.18653\/v1\/2020.acl-main.560"},{"key":"1302_CR21","doi-asserted-by":"publisher","unstructured":"Howard, J., Ruder, S.: Universal language model fine-tuning for text classification. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 328\u2013339. Association for Computational Linguistics, Melbourne, Australia (2018). https:\/\/doi.org\/10.18653\/v1\/P18-1031. https:\/\/aclanthology.org\/P18-1031","DOI":"10.18653\/v1\/P18-1031"},{"key":"1302_CR22","doi-asserted-by":"publisher","unstructured":"Li, X.L., Liang, P.: Prefix-tuning: Optimizing continuous prompts for generation. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 4582\u20134597. Association for Computational Linguistics, Online (2021). https:\/\/doi.org\/10.18653\/v1\/2021.acl-long.353 . https:\/\/aclanthology.org\/2021.acl-long.353","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"1302_CR23","doi-asserted-by":"publisher","unstructured":"Liu, X., Zheng, Y., Du, Z., Ding, M., Qian, Y., Yang, Z., Tang, J.: Gpt understands, too. AI Open (2023). https:\/\/doi.org\/10.1016\/j.aiopen.2023.08.012","DOI":"10.1016\/j.aiopen.2023.08.012"},{"key":"1302_CR24","doi-asserted-by":"publisher","unstructured":"Liu, X., Ji, K., Fu, Y., Tam, W., Du, Z., Yang, Z., Tang, J.: P-tuning: Prompt tuning can be comparable to fine-tuning across scales and tasks. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 61\u201368. Association for Computational Linguistics, Dublin, Ireland (2022). https:\/\/doi.org\/10.18653\/v1\/2022.acl-short.8. https:\/\/aclanthology.org\/2022.acl-short.8","DOI":"10.18653\/v1\/2022.acl-short.8"},{"key":"1302_CR25","doi-asserted-by":"publisher","unstructured":"Qin, G., Eisner, J.: Learning how to ask: Querying LMs with mixtures of soft prompts. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 5203\u20135212. Association for Computational Linguistics, Online (2021). https:\/\/doi.org\/10.18653\/v1\/2021.naacl-main.410. https:\/\/aclanthology.org\/2021.naacl-main.410","DOI":"10.18653\/v1\/2021.naacl-main.410"},{"key":"1302_CR26","doi-asserted-by":"publisher","unstructured":"Ding, N., Hu, S., Zhao, W., Chen, Y., Liu, Z., Zheng, H., Sun, M.: OpenPrompt: An open-source framework for prompt-learning. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 105\u2013113. Association for Computational Linguistics, Dublin, Ireland (2022). https:\/\/doi.org\/10.18653\/v1\/2022.acl-demo.10. https:\/\/aclanthology.org\/2022.acl-demo.10","DOI":"10.18653\/v1\/2022.acl-demo.10"},{"key":"1302_CR27","doi-asserted-by":"publisher","unstructured":"Hambardzumyan, K., Khachatrian, H., May, J.: WARP: Word-level Adversarial ReProgramming. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 4921\u20134933. Association for Computational Linguistics, Online (2021). https:\/\/doi.org\/10.18653\/v1\/2021.acl-long.381. https:\/\/aclanthology.org\/2021.acl-long.381","DOI":"10.18653\/v1\/2021.acl-long.381"},{"key":"1302_CR28","unstructured":"Zhao, T., Wallace, E., Feng, S., Klein, D., Singh, S.: Calibrate before use: Improving few-shot performance of language models. In: International Conference on Machine Learning (2021). https:\/\/api.semanticscholar.org\/CorpusID:231979430"},{"key":"1302_CR29","unstructured":"Houlsby, N., Giurgiu, A., Jastrzebski, S., Morrone, B., De\u00a0Laroussilhe, Q., Gesmundo, A., Attariyan, M., Gelly, S.: Parameter-efficient transfer learning for NLP. In: Proceedings of the 36th International Conference on Machine Learning. PMLR, vol. 97, pp. 2790\u20132799 (2019). https:\/\/proceedings.mlr.press\/v97\/houlsby19a.html"},{"key":"1302_CR30","doi-asserted-by":"publisher","unstructured":"Pfeiffer, J., Vuli\u0107, I., Gurevych, I., Ruder, S.: Mad-x: An adapter-based framework for multi-task cross-lingual transfer. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) (2020). https:\/\/doi.org\/10.18653\/v1\/2020.emnlp-main.617","DOI":"10.18653\/v1\/2020.emnlp-main.617"},{"key":"1302_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2023.107239","volume":"127","author":"H Rathnayake","year":"2024","unstructured":"Rathnayake, H., Sumanapala, J., Rukshani, R., Ranathunga, S.: Adapterfusion-based multi-task learning for code-mixed and code-switched text classification. Eng. Appl. Artif. Intell. 127, 107239 (2024). https:\/\/doi.org\/10.1016\/j.engappai.2023.107239","journal-title":"Eng. Appl. Artif. Intell."},{"key":"1302_CR32","doi-asserted-by":"publisher","unstructured":"R\u00fcckl\u00e9, A., Geigle, G., Glockner, M., Beck, T., Pfeiffer, J., Reimers, N., Gurevych, I.: AdapterDrop: On the efficiency of adapters in transformers. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 7930\u20137946. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic (2021). https:\/\/doi.org\/10.18653\/v1\/2021.emnlp-main.626. https:\/\/aclanthology.org\/2021.emnlp-main.626","DOI":"10.18653\/v1\/2021.emnlp-main.626"},{"key":"1302_CR33","doi-asserted-by":"crossref","unstructured":"Shah, A., Thapa, S., Jain, A., Huang, L.: ADEPT: Adapter-based efficient prompt tuning approach for language models. In: Proceedings of The Fourth Workshop on Simple and Efficient Natural Language Processing (SustaiNLP), pp. 121\u2013128. Association for Computational Linguistics, Toronto, Canada (Hybrid) (2023). https:\/\/aclanthology.org\/2023.sustainlp-1.8","DOI":"10.18653\/v1\/2023.sustainlp-1.8"},{"key":"1302_CR34","doi-asserted-by":"publisher","unstructured":"Reynolds, L., McDonell, K.: Prompt programming for large language models: Beyond the few-shot paradigm. In: Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems. CHI EA \u201921. Association for Computing Machinery, New York, NY, USA (2021). https:\/\/doi.org\/10.1145\/3411763.3451760. https:\/\/doi.org\/10.1145\/3411763.3451760","DOI":"10.1145\/3411763.3451760"},{"key":"1302_CR35","doi-asserted-by":"publisher","unstructured":"Bohra, A., Vijay, D., Singh, V., Akhtar, S.S., Shrivastava, M.: A dataset of Hindi-English code-mixed social media text for hate speech detection. In: Proceedings of the Second Workshop on Computational Modeling of People\u2019s Opinions, Personality, and Emotions in Social Media, pp. 36\u201341. Association for Computational Linguistics, New Orleans, Louisiana, USA (2018). https:\/\/doi.org\/10.18653\/v1\/W18-1105. https:\/\/aclanthology.org\/W18-1105","DOI":"10.18653\/v1\/W18-1105"},{"key":"1302_CR36","unstructured":"Vilares, D., Alonso, M.A., G\u00f3mez-Rodr\u00edguez, C.: EN-ES-CS: An English-Spanish code-switching Twitter corpus for multilingual sentiment analysis. In: Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC\u201916), pp. 4149\u20134153. European Language Resources Association (ELRA), Portoro\u017e, Slovenia (2016). https:\/\/aclanthology.org\/L16-1655"},{"key":"1302_CR37","doi-asserted-by":"crossref","unstructured":"Chathuranga, S., Ranathunga, S.: Classification of code-mixed text using capsule networks. In: Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP 2021), pp. 256\u2013263. INCOMA Ltd., Held Online (2021). https:\/\/aclanthology.org\/2021.ranlp-1.30","DOI":"10.26615\/978-954-452-072-4_030"},{"key":"1302_CR38","unstructured":"Kamble, S., Joshi, A.: Hate speech detection from code-mixed Hindi-English tweets using deep learning models. In: Proceedings of the 15th International Conference on Natural Language Processing, pp. 150\u2013155. NLP Association of India, International Institute of Information Technology, Hyderabad, India (2018). https:\/\/aclanthology.org\/2018.icon-1.22"},{"key":"1302_CR39","doi-asserted-by":"publisher","unstructured":"Tatariya, K., Lent, H., De\u00a0Lhoneux, M.: Transfer learning for code-mixed data: Do pretraining languages matter? In: Proceedings of the 13th Workshop on Computational Approaches to Subjectivity, Sentiment, & Social Media Analysis, pp. 365\u2013378. Association for Computational Linguistics, Toronto, Canada (2023). https:\/\/doi.org\/10.18653\/v1\/2023.wassa-1.32. https:\/\/aclanthology.org\/2023.wassa-1.32","DOI":"10.18653\/v1\/2023.wassa-1.32"},{"key":"1302_CR40","doi-asserted-by":"publisher","DOI":"10.1016\/j.nlp.2023.100042","volume":"5","author":"G Takawane","year":"2023","unstructured":"Takawane, G., Phaltankar, A., Patwardhan, V., Patil, A., Joshi, R., Takalikar, M.: Language augmentation approach for code-mixed text classification. Natural Language Processing Journal 5, 100042 (2023). https:\/\/doi.org\/10.1016\/j.nlp.2023.100042","journal-title":"Natural Language Processing Journal"},{"key":"1302_CR41","unstructured":"Laureano De\u00a0Leon, F.A., Tayyar\u00a0Madabushi, H., Lee, M.: Code-mixed probes show how pre-trained models generalise on code-switched text. In: Calzolari, N., Kan, M.-Y., Hoste, V., Lenci, A., Sakti, S., Xue, N. (eds.) Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pp. 3457\u20133468. ELRA and ICCL, Torino, Italia (2024). https:\/\/aclanthology.org\/2024.lrec-main.307"},{"key":"1302_CR42","doi-asserted-by":"publisher","unstructured":"Winata, G., Cahyawijaya, S., Liu, Z., Lin, Z., Madotto, A., Fung, P.: Are multilingual models effective in code-switching?, pp. 142\u2013153 (2021). https:\/\/doi.org\/10.18653\/v1\/2021.calcs-1.20","DOI":"10.18653\/v1\/2021.calcs-1.20"},{"key":"1302_CR43","doi-asserted-by":"publisher","first-page":"118837","DOI":"10.1109\/ACCESS.2021.3104106","volume":"9","author":"S Thara","year":"2021","unstructured":"Thara, S., Poornachandran, P.: Transformer based language identification for malayalam-english code-mixed text. IEEE Access 9, 118837\u2013118850 (2021). https:\/\/doi.org\/10.1109\/ACCESS.2021.3104106","journal-title":"IEEE Access"},{"key":"1302_CR44","doi-asserted-by":"publisher","unstructured":"Zhang, R., Cahyawijaya, S., Cruz, J.C.B., Winata, G., Aji, A.: Multilingual large language models are not (yet) code-switchers. In: Bouamor, H., Pino, J., Bali, K. (eds.) Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp. 12567\u201312582. Association for Computational Linguistics, Singapore (2023). https:\/\/doi.org\/10.18653\/v1\/2023.emnlp-main.774. https:\/\/aclanthology.org\/2023.emnlp-main.774","DOI":"10.18653\/v1\/2023.emnlp-main.774"},{"key":"1302_CR45","doi-asserted-by":"publisher","unstructured":"Qin, L., Ni, M., Zhang, Y., Che, W.: Cosda-ml: Multi-lingual code-switching data augmentation for zero-shot cross-lingual nlp. In: Bessiere, C. (ed.) Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20, pp. 3853\u20133860. International Joint Conferences on Artificial Intelligence Organization, ??? (2020). https:\/\/doi.org\/10.24963\/ijcai.2020\/533 . Main track. https:\/\/doi.org\/10.24963\/ijcai.2020\/533","DOI":"10.24963\/ijcai.2020\/533"},{"key":"1302_CR46","doi-asserted-by":"crossref","unstructured":"Hande, A., Hegde, S.U., Priyadharshini, R., Ponnusamy, R., Kumaresan, P.K., Thavareesan, S., Chakravarthi, B.R.: Benchmarking multi-task learning for sentiment analysis and offensive language identification in under-resourced dravidian languages. CoRR abs\/2108.03867 (2021) arXiv:2108.03867","DOI":"10.21203\/rs.3.rs-546149\/v1"}],"container-title":["World Wide Web"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11280-024-01302-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11280-024-01302-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11280-024-01302-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,20]],"date-time":"2024-09-20T10:16:20Z","timestamp":1726827380000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11280-024-01302-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9]]},"references-count":46,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,9]]}},"alternative-id":["1302"],"URL":"https:\/\/doi.org\/10.1007\/s11280-024-01302-2","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-4248891\/v1","asserted-by":"object"}]},"ISSN":["1386-145X","1573-1413"],"issn-type":[{"value":"1386-145X","type":"print"},{"value":"1573-1413","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,9]]},"assertion":[{"value":"10 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 August 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 August 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 September 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"We confirm that this work is original and has not been published elsewhere, nor is it currently under consideration for publication elsewhere. All authors consent to the publication of this manuscript.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for Publication"}},{"value":"Ethical approval and consent to participate are not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval and Consent to Participate"}}],"article-number":"63"}}