{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,21]],"date-time":"2026-04-21T20:09:04Z","timestamp":1776802144898,"version":"3.51.2"},"reference-count":72,"publisher":"MDPI AG","issue":"2","license":[{"start":{"date-parts":[[2025,2,8]],"date-time":"2025-02-08T00:00:00Z","timestamp":1738972800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["BDCC"],"abstract":"<jats:p>Stance detection for low-resource languages, such as the Kuwaiti dialect, poses a significant challenge in natural language processing (NLP) due to the scarcity of annotated datasets and specialized tools. This study addresses these limitations by evaluating the effectiveness of open large language models (LLMs) in automating stance detection through zero-shot and few-shot prompt engineering, with a focus on the potential of open-source models to achieve performance levels comparable to those of closed-source alternatives. We also highlight the critical distinctions between zero- and few-shot learning, emphasizing their significance for addressing the challenges posed by low-resource languages. Our evaluation involved testing 11 LLMs on a manually labeled dataset of social media posts, including GPT-4o, Gemini Pro 1.5, Mistral-Large, Jais-30B, and AYA-23. As expected, closed-source models such as GPT-4o, Gemini Pro 1.5, and Mistral-Large demonstrated superior performance, achieving maximum F1 scores of 95.4%, 95.0%, and 93.2%, respectively, in few-shot scenarios with English as the prompt template language. However, open-source models such as Jais-30B and AYA-23 achieved competitive results, with maximum F1 scores of 93.0% and 93.1%, respectively, under the same conditions. Furthermore, statistical analysis using ANOVA and Tukey\u2019s HSD post hoc tests revealed no significant differences in overall performance among GPT-4o, Gemini Pro 1.5, Mistral-Large, Jais-30B, and AYA-23. This finding underscores the potential of open-source LLMs as cost-effective and privacy-preserving alternatives for low-resource language annotation. This is the first study comparing LLMs for stance detection in the Kuwaiti dialect. Our findings highlight the importance of prompt design and model consistency in improving the quality of annotations and pave the way for NLP solutions for under-represented Arabic dialects.<\/jats:p>","DOI":"10.3390\/bdcc9020033","type":"journal-article","created":{"date-parts":[[2025,2,10]],"date-time":"2025-02-10T06:43:07Z","timestamp":1739169787000},"page":"33","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Large Language Models as Kuwaiti Annotators"],"prefix":"10.3390","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5167-1558","authenticated-orcid":false,"given":"Hana","family":"Alostad","sequence":"first","affiliation":[{"name":"Systems and Software Development Department, Institute for Scientific Research, P.O. Box 24885, Safat 13109, Kuwait"}]}],"member":"1968","published-online":{"date-parts":[[2025,2,8]]},"reference":[{"key":"ref_1","first-page":"1","article-title":"Stance Detection: A Survey","volume":"53","author":"Can","year":"2020","journal-title":"ACM Comput. Surv."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.4018\/IJMDEM.2018070101","article-title":"Efficient Large-Scale Stance Detection in Tweets","volume":"9","author":"Shyu","year":"2018","journal-title":"Int. J. Multimed. Data Eng. Manag."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Burnham, M. (2024). Stance detection: A practical guide to classifying political beliefs in text. Political Sci. Res. Methods, 1\u201318.","DOI":"10.1017\/psrm.2024.35"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Kuo, K.H., Wang, M.H., Kao, H.Y., and Dai, Y.C. (2024, January 13\u201317). Advancing Stance Detection of Political Fan Pages: A Multimodal Approach. Proceedings of the Companion Proceedings of the ACM Web Conference 2024, WWW \u201924, New York, NY, USA.","DOI":"10.1145\/3589335.3651467"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Mets, M., Karjus, A., Ibrus, I., and Schich, M. (2024). Automated stance detection in complex topics and small languages: The challenging case of immigration in polarizing news media. PLoS ONE, 19.","DOI":"10.1371\/journal.pone.0302380"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"44","DOI":"10.1145\/3610409.3610412","article-title":"Identifying Behavioral Factors Leading to Differential Polarization Effects of Adversarial Botnets","volume":"23","author":"Lee","year":"2023","journal-title":"SIGAPP Appl. Comput. Rev."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Lee, Y., Alostad, H., and Davulcu, H. (2024). Quantifying Variations in Controversial Discussions within Kuwaiti Social Networks. Big Data Cogn. Comput., 8.","DOI":"10.3390\/bdcc8060060"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"1668","DOI":"10.1093\/jamia\/ocac114","article-title":"Using Twitter Data to Understand Public Perceptions of Approved Versus Off-Label Use for COVID-19-related Medications","volume":"29","author":"Hua","year":"2022","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"101454","DOI":"10.1016\/j.eclinm.2022.101454","article-title":"Social Media and Attitudes Towards a COVID-19 Vaccination: A Systematic Review of the Literature","volume":"48","author":"Cascini","year":"2022","journal-title":"Eclinicalmedicine"},{"key":"ref_10","unstructured":"(2024, December 16). List of Countries and Territories Where Arabic Is an Official Language. Available online: https:\/\/en.wikipedia.org\/wiki\/List_of_countries_and_territories_where_Arabic_is_an_official_language."},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Alostad, H., Dawiek, S., and Davulcu, H. (2023). Q8VaxStance: Dataset Labeling System for Stance Detection towards Vaccines in Kuwaiti Dialect. Big Data Cogn. Comput., 7.","DOI":"10.20944\/preprints202308.0954.v1"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Alhindi, T., Alabdulkarim, A., Alshehri, A., Abdul-Mageed, M., and Nakov, P. (2021). AraStance: A Multi-Country and Multi-Domain Dataset of Arabic Stance Detection for Fact Checking. Proceedings of the Fourth Workshop on NLP for Internet Freedom: Censorship, Disinformation, and Propaganda, Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.nlp4if-1.9"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Alturayeif, N.S., Luqman, H.A., and Ahmed, M.A.K. (2022). MAWQIF: A Multi-label Arabic Dataset for Target-specific Stance Detection. Proceedings of the Seventh Arabic Natural Language Processing Workshop (WANLP), Association for Computational Linguistics.","DOI":"10.18653\/v1\/2022.wanlp-1.16"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Haouari, F., and Elsayed, T. (2023). Detecting stance of authorities towards rumors in Arabic tweets: A preliminary study. Proceedings of the 45th European Conference on Information Retrieval, Springer.","DOI":"10.1007\/978-3-031-28238-6_33"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Hamad, O., Hamdi, A., Hamdi, S., and Shaban, K. (2022). StEduCov: An Explored and Benchmarked Dataset on Stance Detection in Tweets towards Online Education during COVID-19 Pandemic. Big Data Cogn. Comput., 6.","DOI":"10.3390\/bdcc6030088"},{"key":"ref_16","first-page":"10729","article-title":"Few-Shot Cross-Lingual Stance Detection With Sentiment-Based Pre-Training","volume":"36","author":"Hardalov","year":"2022","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"ref_17","unstructured":"Aletras, N., and De Clercq, O. (2024, January 17\u201322). MEGAnno+: A Human-LLM Collaborative Annotation System. Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations, St. Julian\u2019s, Malta."},{"key":"ref_18","unstructured":"Yang, Y., Davani, A., Sil, A., and Kumar, A. (2024, January 16\u201321). AnnoLLM: Making Large Language Models to Be Better Crowdsourced Annotators. Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 6: Industry Track), Mexico City, Mexico."},{"key":"ref_19","unstructured":"Bouamor, H., Pino, J., and Bali, K. (2023, January 6\u201310). LLMaAA: Making Large Language Models as Active Annotators. Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, Singapore."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Liu, R., Lin, Z., Tan, Y., and Wang, W. (2021). Enhancing zero-shot and few-shot stance detection with commonsense knowledge graph. Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Association for Computational Linguistics.","DOI":"10.18653\/v1\/2021.findings-acl.278"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"81","DOI":"10.18653\/v1\/2022.acl-long.7","article-title":"Jointcl: A joint contrastive learning framework for zero-shot stance detection","volume":"Volume 1","author":"Liang","year":"2022","journal-title":"Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Liew, X.Y., Hameed, N., Clos, J., and Fischer, J.E. (2023, January 11\u201312). Predicting Stance to Detect Misinformation in Few-shot Learning. Proceedings of the First International Symposium on Trustworthy Autonomous Systems, TAS \u201923, New York, NY, USA.","DOI":"10.1145\/3597512.3599717"},{"key":"ref_23","unstructured":"Calzolari, N., Kan, M.Y., Hoste, V., Lenci, A., Sakti, S., and Xue, N. (2024, January 20\u201325). Can GPT-4 Identify Propaganda? Annotation and Detection of Propaganda Spans in News Articles. Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), Torino, Italy."},{"key":"ref_24","unstructured":"Graham, Y., and Purver, M. (2024, January 17\u201322). LAraBench: Benchmarking Arabic AI with Large Language Models. Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), St. Julian\u2019s, Malta."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"27709","DOI":"10.1109\/ACCESS.2024.3364367","article-title":"Bridging the Kuwaiti Dialect Gap in Natural Language Processing","volume":"12","author":"Husain","year":"2024","journal-title":"IEEE Access"},{"key":"ref_26","unstructured":"(2024, September 22). OpenAI. GPT-4o System Card. Available online: https:\/\/openai.com\/index\/gpt-4o-system-card\/."},{"key":"ref_27","unstructured":"(2024, December 16). AI Models. Premier Models. Available online: https:\/\/mistral.ai\/technology\/#models."},{"key":"ref_28","unstructured":"Team, G., Anil, R., Borgeaud, S., Wu, Y., Alayrac, J.B., Yu, J., Soricut, R., Schalkwyk, J., Dai, A.M., and Hauth, A. (2023). Gemini: A family of highly capable multimodal models. arXiv."},{"key":"ref_29","unstructured":"Sengupta, N., Sahu, S.K., Jia, B., Katipomu, S., Li, H., Koto, F., Marshall, W., Gosal, G., Liu, C., and Chen, Z. (2023). Jais and Jais-chat: Arabic-Centric Foundation and Instruction-Tuned Open Generative Large Language Models. arXiv."},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Muennighoff, N., Wang, T., Sutawika, L., Roberts, A., Biderman, S., Scao, T.L., Bari, M.S., Shen, S., Yong, Z.X., and Schoelkopf, H. (2023, January 9\u201314). Crosslingual Generalization through Multitask Finetuning. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada.","DOI":"10.18653\/v1\/2023.acl-long.891"},{"key":"ref_31","unstructured":"Aryabumi, V., Dang, J., Talupuru, D., Dash, S., Cairuz, D., Lin, H., Venkitesh, B., Smith, M., Campos, J.A., and Tan, Y.C. (2024). Aya 23: Open Weight Releases to Further Multilingual Progress. arXiv."},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"\u00dcst\u00fcn, A., Aryabumi, V., Yong, Z.X., Ko, W.Y., D\u2019souza, D., Onilude, G., Bhandari, N., Singh, S., Ooi, H.L., and Kayid, A. (2024). Aya model: An instruction finetuned open-access multilingual language model. arXiv.","DOI":"10.18653\/v1\/2024.acl-long.845"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Huang, H., Yu, F., Zhu, J., Sun, X., Cheng, H., Song, D., Chen, Z., Alharthi, A., An, B., and He, J. (2024). AceGPT, Localizing Large Language Models in Arabic. arXiv.","DOI":"10.18653\/v1\/2024.naacl-long.450"},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"68675","DOI":"10.1109\/ACCESS.2021.3077350","article-title":"The NLP Cookbook: Modern Recipes for Transformer Based Deep Learning Architectures","volume":"9","author":"Singh","year":"2021","journal-title":"IEEE Access"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Li, K., Liu, Z., He, T., Huang, H., Peng, F., Povey, D., and Khudanpur, S. (2020, January 4\u20138). An Empirical Study of Transformer-Based Neural Language Model Adaptation. Proceedings of the ICASSP 2020\u20142020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain.","DOI":"10.1109\/ICASSP40776.2020.9053399"},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"2584","DOI":"10.1109\/JSAC.2022.3191326","article-title":"Task-oriented multi-user semantic communications","volume":"40","author":"Xie","year":"2022","journal-title":"IEEE J. Sel. Areas Commun."},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Choi, H., Kim, J., Joe, S., and Gwon, Y. (2021, January 10\u201315). Evaluation of BERT and ALBERT Sentence Embedding Performance on Downstream NLP Tasks. Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy.","DOI":"10.1109\/ICPR48806.2021.9412102"},{"key":"ref_38","unstructured":"Devlin, J., Chang, M.W., Lee, K., and Toutanova, K. (2019). BERT Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv."},{"key":"ref_39","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Beheitt, M.E.G., and Ben Haj Hmida, M. (2022, January 3\u20135). Automatic Arabic Poem Generation with GPT-2. Proceedings of the 14th International Conference on Agents and Artificial Intelligence, Online.","DOI":"10.5220\/0010847100003116"},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"100160","DOI":"10.1016\/j.caeai.2023.100160","article-title":"To GPT or not GPT? Empowering our students to learn with AI","volume":"5","author":"Steele","year":"2023","journal-title":"Comput. Educ. Artif. Intell."},{"key":"ref_42","unstructured":"(2024, December 16). Google AI. Gemini Models. Available online: https:\/\/ai.google.dev\/gemini-api\/docs\/models\/gemini."},{"key":"ref_43","unstructured":"(2024, December 16). Google AI. Pricing Models. Available online: https:\/\/ai.google.dev\/pricing."},{"key":"ref_44","unstructured":"(2024, December 16). Google AI. Billing. Available online: https:\/\/ai.google.dev\/gemini-api\/docs\/billing."},{"key":"ref_45","unstructured":"(2024, December 16). Core42. Core42\u2019s Bilingual AI for Arabic Speakers. Available online: https:\/\/www.core42.ai\/jais.html."},{"key":"ref_46","unstructured":"Zong, C., Xia, F., Li, W., and Navigli, R. (2021, January 1\u20136). ARBERT & MARBERT: Deep Bidirectional Transformers for Arabic. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online."},{"key":"ref_47","unstructured":"Kmainasi, M., Shahroor, A., Hasanain, M., Laskar, S., Hassan, N., and Alam, F. (2024). LlamaLens: Specialized Multilingual LLM for Analyzing News and Social Media Content. arXiv."},{"key":"ref_48","doi-asserted-by":"crossref","unstructured":"Alturayeif, N., Luqman, H., Alyafeai, Z., and Yamani, A. (2024, January 16). StanceEval 2024: The First Arabic Stance Detection Shared Task. Proceedings of the Second Arabic Natural Language Processing Conference, Bangkok, Thailand.","DOI":"10.18653\/v1\/2024.arabicnlp-1.88"},{"key":"ref_49","unstructured":"Habash, N., Bouamor, H., Eskander, R., Tomeh, N., Abu Farha, I., Abdelali, A., Touileb, S., Hamed, I., Onaizan, Y., and Alhafni, B. (2024, January 16). AlexUNLP-BH at StanceEval2024: Multiple Contrastive Losses Ensemble Strategy with Multi-Task Learning For Stance Detection in Arabic. Proceedings of the Second Arabic Natural Language Processing Conference, Bangkok, Thailand."},{"key":"ref_50","unstructured":"Habash, N., Bouamor, H., Eskander, R., Tomeh, N., Abu Farha, I., Abdelali, A., Touileb, S., Hamed, I., Onaizan, Y., and Alhafni, B. (2024, January 16). MGKM at StanceEval2024 Fine-Tuning Large Language Models for Arabic Stance Detection. Proceedings of the Second Arabic Natural Language Processing Conference, Bangkok, Thailand."},{"key":"ref_51","unstructured":"Habash, N., Bouamor, H., Eskander, R., Tomeh, N., Abu Farha, I., Abdelali, A., Touileb, S., Hamed, I., Onaizan, Y., and Alhafni, B. (2024, January 16). StanceCrafters at StanceEval2024: Multi-task Stance Detection using BERT Ensemble with Attention Based Aggregation. Proceedings of the Second Arabic Natural Language Processing Conference, Bangkok, Thailand."},{"key":"ref_52","unstructured":"Salamah, J.B., and Elkhlifi, A. (2014, January 9). Microblogging opinion mining approach for kuwaiti dialect. Proceedings of the the International Conference on Computing Technology and Information Management (ICCTIM), Dubai, United Arab Emirates. Society of Digital Information and Wireless Communication (SDIWC)."},{"key":"ref_53","unstructured":"Quinlan, J.R. (1993). C4.5: Programs for Machine Learning, Morgan Kaufmann Publishers Inc."},{"key":"ref_54","unstructured":"Freund, Y., and Mason, L. (1999, January 27\u201330). The Alternating Decision Tree Learning Algorithm. Proceedings of the Sixteenth International Conference on Machine Learning, ICML \u201999, San Francisco, CA, USA."},{"key":"ref_55","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1023\/A:1010933404324","article-title":"Random Forests","volume":"45","author":"Breiman","year":"2001","journal-title":"Mach. Learn."},{"key":"ref_56","doi-asserted-by":"crossref","unstructured":"Boser, B.E., Guyon, I.M., and Vapnik, V.N. (1992, January 27\u201329). A training algorithm for optimal margin classifiers. Proceedings of the Fifth Annual Workshop on Computational Learning Theory, Pittsburgh, PA, USA.","DOI":"10.1145\/130385.130401"},{"key":"ref_57","unstructured":"Al-Khalifa, H., Magdy, W., Darwish, K., Elsayed, T., and Mubarak, H. (2020, January 11\u201316). AraBERT: Transformer-based Model for Arabic Language Understanding. Proceedings of the 4th Workshop on Open-Source Arabic Corpora and Processing Tools, with a Shared Task on Offensive Language Detection, Marseille, France."},{"key":"ref_58","unstructured":"Wang, W., Wei, F., Dong, L., Bao, H., Yang, N., and Zhou, M. (2020, January 6\u201312). MINILM: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS \u201920, Vancouver, BC, Canada."},{"key":"ref_59","unstructured":"Habash, N., Bouamor, H., Hajj, H., Magdy, W., Zaghouani, W., Bougares, F., Tomeh, N., Abu Farha, I., and Touileb, S. (2021, January 19). The Interplay of Variant, Size, and Task Type in Arabic Pre-trained Language Models. Proceedings of the Sixth Arabic Natural Language Processing Workshop, Kyiv, Ukraine (Virtual)."},{"key":"ref_60","unstructured":"Al-Onaizan, Y., Bansal, M., and Chen, Y.N. (2024, January 12\u201316). Large Language Models for Data Annotation and Synthesis: A Survey. Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Miami, FL, USA."},{"key":"ref_61","unstructured":"Kojima, T., Gu, S.S., Reid, M., Matsuo, Y., and Iwasawa, Y. (December, January 28). Large language models are zero-shot reasoners. Proceedings of the 36th International Conference on Neural Information Processing Systems, NIPS \u201922, Red Hook, NY, USA."},{"key":"ref_62","unstructured":"Lee, H.Y., Mohtarami, M., Li, S.W., Jin, D., Korpusik, M., Dong, S., Vu, N.T., and Hakkani-Tur, D. (2021, January 5). Semi-supervised Meta-learning for Cross-domain Few-shot Intent Classification. Proceedings of the 1st Workshop on Meta Learning and Its Applications to Natural Language Processing, Online."},{"key":"ref_63","unstructured":"Wei, J., Bosma, M., Zhao, V.Y., Guu, K., Yu, A.W., Lester, B., Du, N., Dai, A.M., and Le, Q.V. (2022, January 25\u201329). Finetuned Language Models are Zero-Shot Learners. Proceedings of the the Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event."},{"key":"ref_64","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., and Askell, A. (2020). Language Models are Few-Shot Learners. Proceedings of the 34th International Conference on Neural Information Processing Systems, Curran Associates, Inc."},{"key":"ref_65","unstructured":"Zong, C., Xia, F., Li, W., and Navigli, R. (2021, January 5). Making Pre-trained Language Models Better Few-shot Learners. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online."},{"key":"ref_66","first-page":"82","article-title":"Llms accelerate annotation for medical information extraction","volume":"Volume 225","author":"Goel","year":"2023","journal-title":"Proceedings of the Machine Learning for Health (ML4H), PMLR"},{"key":"ref_67","unstructured":"Calzolari, N., Kan, M.Y., Hoste, V., Lenci, A., Sakti, S., and Xue, N. (2024, January 20\u201325). Large Language Models as Financial Data Annotators: A Study on Effectiveness and Efficiency. Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), Torino, Italy."},{"key":"ref_68","doi-asserted-by":"crossref","first-page":"367","DOI":"10.3390\/make6010018","article-title":"Prompt Engineering or Fine-Tuning? A Case Study on Phishing Detection with Large Language Models","volume":"6","author":"Trad","year":"2024","journal-title":"Mach. Learn. Knowl. Extr."},{"key":"ref_69","doi-asserted-by":"crossref","unstructured":"Kholodna, N., Julka, S., Khodadadi, M., Gumus, M.N., and Granitzer, M. (2024, January 9\u201313). LLMs in the loop: Leveraging large language model annotations for active learning in low-resource languages. Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Vilnius, Lithuania.","DOI":"10.1007\/978-3-031-70381-2_25"},{"key":"ref_70","unstructured":"FreedomIntelligence (2024, December 16). FreedomIntelligence\/AceGPT-v2-32B. Available online: https:\/\/huggingface.co\/FreedomIntelligence\/AceGPT-v2-32B."},{"key":"ref_71","unstructured":"Wikipedia (2024, December 16). Variance. Available online: https:\/\/en.wikipedia.org\/wiki\/Variance."},{"key":"ref_72","unstructured":"RunPod (2024, December 16). RunPod\u2014The Cloud Built for AI. Available online: https:\/\/www.runpod.io\/."}],"container-title":["Big Data and Cognitive Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2504-2289\/9\/2\/33\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T16:29:26Z","timestamp":1760027366000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2504-2289\/9\/2\/33"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,8]]},"references-count":72,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2025,2]]}},"alternative-id":["bdcc9020033"],"URL":"https:\/\/doi.org\/10.3390\/bdcc9020033","relation":{},"ISSN":["2504-2289"],"issn-type":[{"value":"2504-2289","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,8]]}}}