{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:01:07Z","timestamp":1775815267344,"version":"3.50.1"},"reference-count":85,"publisher":"Association for Computing Machinery (ACM)","issue":"5","funder":[{"name":"(Huawei Innovation Research Program), Research Impact Fund","award":["R1015-23"],"award-info":[{"award-number":["R1015-23"]}]},{"name":"APRC - CityU New Research Initiatives","award":["9610565"],"award-info":[{"award-number":["9610565"]}]},{"name":"CityU - HKIDS Early Career Research Grant","award":["9360163"],"award-info":[{"award-number":["9360163"]}]},{"name":"Hong Kong ITC Innovation and Technology Fund Midstream Research Programme for Universities Project","award":["ITS\/034\/22MS"],"award-info":[{"award-number":["ITS\/034\/22MS"]}]},{"name":"Hong Kong Environmental and Conservation Fund","award":["88\/2022"],"award-info":[{"award-number":["88\/2022"]}]},{"name":"SIRG - CityU Strategic Interdisciplinary Research Grant","award":["7020046"],"award-info":[{"award-number":["7020046"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Inf. Syst."],"published-print":{"date-parts":[[2025,9,30]]},"abstract":"<jats:p>Multi-Domain Click-Through Rate (MDCTR) prediction is crucial for online recommendation platforms, which involves providing personalized recommendation services to users in different domains. However, current MDCTR models are confronted with the following limitations. Firstly, due to varying data sparsity in different domains, models can easily be dominated by some specific domains, which leads to significant performance degradation in other domains (i.e., the \u201cseesaw phenomenon\u201d). Secondly, when new domain emerges, the scalability of existing methods is limited, making it difficult to adapt to the dynamic growth of the domain. Traditional MDCTR models usually use one-hot encoding for semantic information such as product titles, thus losing rich semantic information and leading to insufficient generalization of the model. In this article, we propose a novel solution Uni-CTR to address these challenges. Uni-CTR leverages Large Language Model (LLM) to extract layer-wise semantic representations that capture domain commonalities, mitigating the seesaw phenomenon and enhancing generalization. Besides, it incorporates a pluggable domain-specific network to capture domain characteristics, ensuring scalability to dynamic domain growth. Experimental results on public datasets and industrial scenarios show that Uni-CTR significantly outperforms state-of-the-art (SOTA) models. In addition, Uni-CTR shows significant results in zero shot prediction. Code is available at Applied Machine Learning Lab (Pytorch), GitHub (Pytorch) and Gitee (MindSpore).<\/jats:p>","DOI":"10.1145\/3698878","type":"journal-article","created":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T10:10:31Z","timestamp":1728900631000},"page":"1-33","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":16,"title":["A Unified Framework for Multi-Domain CTR Prediction via Large Language Models"],"prefix":"10.1145","volume":"43","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1698-4342","authenticated-orcid":false,"given":"Zichuan","family":"Fu","sequence":"first","affiliation":[{"name":"City University of Hong Kong, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2862-0239","authenticated-orcid":false,"given":"Xiangyang","family":"Li","sequence":"additional","affiliation":[{"name":"Huawei Noah\u2019s Ark Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5730-8792","authenticated-orcid":false,"given":"Chuhan","family":"Wu","sequence":"additional","affiliation":[{"name":"Huawei Noah\u2019s Ark Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7053-8269","authenticated-orcid":false,"given":"Yichao","family":"Wang","sequence":"additional","affiliation":[{"name":"Huawei Noah\u2019s Ark Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5564-0641","authenticated-orcid":false,"given":"Kuicai","family":"Dong","sequence":"additional","affiliation":[{"name":"Huawei Noah\u2019s Ark Lab, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2926-4416","authenticated-orcid":false,"given":"Xiangyu","family":"Zhao","sequence":"additional","affiliation":[{"name":"City University of Hong Kong, Hong Kong, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-2382-7053","authenticated-orcid":false,"given":"Mengchen","family":"Zhao","sequence":"additional","affiliation":[{"name":"Huawei Noah\u2019s Ark Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7393-8994","authenticated-orcid":false,"given":"Huifeng","family":"Guo","sequence":"additional","affiliation":[{"name":"Huawei Noah\u2019s Ark Lab, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9224-2431","authenticated-orcid":false,"given":"Ruiming","family":"Tang","sequence":"additional","affiliation":[{"name":"Huawei Noah\u2019s Ark Lab, Shenzhen, China"}]}],"member":"320","published-online":{"date-parts":[[2025,7,10]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1145\/3604915.3608857"},{"key":"e_1_3_2_3_2","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared D. Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et al. 2020. Language Models are Few-Shot Learners. In Advances in Neural Information Processing Systems. H. Larochelle M. Ranzato R. Hadsell M.F. Balcan and H. Lin (Eds.) Vol. 33. Curran Associates Inc. 1877\u20131901. Retrieved from https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2020\/file\/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1023\/A:1007379606734"},{"key":"e_1_3_2_5_2","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599884"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1007\/S10791-009-9109-9"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2964325"},{"key":"e_1_3_2_8_2","unstructured":"Xu Chen Zida Cheng Shuai Xiao Xiaoyi Zeng and Weilin Huang. 2023. Cross-domain augmentation networks for click-through rate prediction. arXiv:2305.03953. Retrieved from https:\/\/arxiv.org\/abs\/2305.03953"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1111\/j.2517-6161.1958.tb00292.x"},{"key":"e_1_3_2_10_2","unstructured":"Zeyu Cui Jianxin Ma Chang Zhou Jingren Zhou and Hongxia Yang. 2022. M6-Rec: Generative Pretrained Language Models are Open-Ended Recommender Systems. arXiv:2205.08084. Retrieved from https:\/\/arxiv.org\/abs\/2205.08084"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210071"},{"key":"e_1_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.18653\/V1\/N19-1423"},{"key":"e_1_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.18653\/V1\/2022.ACL-LONG.26"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.3389\/FRAI.2020.00004"},{"key":"e_1_3_2_15_2","doi-asserted-by":"publisher","unstructured":"Kun Gai Xiaoqiang Zhu Han Li Kai Liu and Zhe Wang. 2017. Learning piece-wise linear models from large scale data for ad click prediction. arXiv:1704.05194. Retrieved from 10.48550\/arXiv.1704.05194","DOI":"10.48550\/arXiv.1704.05194"},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.indmarman.2020.05.011"},{"key":"e_1_3_2_17_2","unstructured":"Rohit Girdhar and Deva Ramanan. 2017. Attentional Pooling for Action Recognition. In Advances in Neural Information Processing Systems. I. Guyon U. Von Luxburg S. Bengio H. Wallach R. Fergus S. Vishwanathan and R. Garnett (Eds.) Vol. 30 Curran Associates Inc. Retrieved from https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2017\/file\/67c6a1e7ce56d3d6fa748ab6d9af3fd7-Paper.pdf"},{"key":"e_1_3_2_18_2","unstructured":"Rafael Glauber and Angelo C. Loula. 2019. Collaborative Filtering vs. Content-Based Filtering: differences and similarities. arXiv:1912.08932. Retrieved from http:\/\/arXiv.org\/abs\/1912.08932"},{"key":"e_1_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3614657"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.24963\/IJCAI.2017\/239"},{"key":"e_1_3_2_21_2","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3615469"},{"key":"e_1_3_2_22_2","unstructured":"Junyou He Guibao Mei Feng Xing Xiaorui Yang Yongjun Bao and Weipeng Yan. 2020. DADNN: Multi-Scene CTR Prediction via Domain-Aware Deep Neural Network. arXiv:2011.11938. Retrieved from https:\/\/arxiv.org\/abs\/2011.11938"},{"key":"e_1_3_2_23_2","volume-title":"Proceedings of the 11th International Conference on Learning Representations (ICLR \u201923)","author":"He Pengcheng","year":"2023","unstructured":"Pengcheng He, Jianfeng Gao, and Weizhu Chen. 2023. DeBERTaV3: Improving DeBERTa Using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing. In Proceedings of the 11th International Conference on Learning Representations (ICLR \u201923). OpenReview.net. Retrieved from https:\/\/openreview.net\/pdf?id=sE7-XhLxHA"},{"key":"e_1_3_2_24_2","volume-title":"Proceedings of the 9th International Conference on Learning Representations (ICLR \u201921)","author":"He Pengcheng","year":"2021","unstructured":"Pengcheng He, Xiaodong Liu, Jianfeng Gao, and Weizhu Chen. 2021. Deberta: Decoding-Enhanced Bert with Disentangled Attention. In Proceedings of the 9th International Conference on Learning Representations (ICLR \u201921). OpenReview.net. Retrieved from https:\/\/openreview.net\/forum?id=XPZIaotutsD"},{"key":"e_1_3_2_25_2","unstructured":"Edward J. Hu Yelong Shen Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang and Weizhu Chen. 2022. LoRA: Low-Rank Adaptation of Large Language Models. In Proceedings of the 10th International Conference on Learning Representations (ICLR \u201922). OpenReview.net. Retrieved from https:\/\/openreview.net\/forum?id=nZeVKeeFYf9"},{"key":"e_1_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.1145\/3269206.3271684"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.1145\/3298689.3347043"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.18653\/V1\/P19-1356"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.372"},{"key":"e_1_3_2_30_2","doi-asserted-by":"publisher","DOI":"10.1145\/2959100.2959134"},{"key":"e_1_3_2_31_2","doi-asserted-by":"publisher","unstructured":"Jared Kaplan Sam McCandlish Tom Henighan Tom B. Brown Benjamin Chess Rewon Child Scott Gray Alec Radford Jeffrey Wu and Dario Amodei. 2020. Scaling laws for neural language models. arXiv:2001.08361. Retrieved from 10.48550\/arXiv.2001.08361","DOI":"10.48550\/arXiv.2001.08361"},{"key":"e_1_3_2_32_2","volume-title":"Proceedings of the 3rd International Conference on Learning Representations (ICLR \u201915).","author":"Kingma Diederik P.","year":"2015","unstructured":"Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR \u201915). Yoshua Bengio and Yann LeCun (Eds.), OpenReview.net. Retrieved from http:\/\/arXiv.org\/abs\/1412.6980"},{"key":"e_1_3_2_33_2","unstructured":"Anders Krogh and John Hertz. 1991. A Simple Weight Decay Can Improve Generalization. In Advances in Neural Information Processing Systems. J. Moody S. Hanson and R. P. Lippmann (Eds.) Vol. 4 Morgan-Kaufmann. Retrieved from https:\/\/proceedings.neurips.cc\/paper_files\/paper\/1991\/file\/8eefcfdf5990e441f0fb6f3fad709e21-Paper.pdf"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1109\/IADCC.2015.7154880"},{"key":"e_1_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1007\/S10489-019-01571-9"},{"key":"e_1_3_2_36_2","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3412713"},{"key":"e_1_3_2_37_2","doi-asserted-by":"publisher","DOI":"10.1145\/3511808.3557072"},{"key":"e_1_3_2_38_2","unstructured":"Xiangyang Li Bo Chen Lu Hou and Ruiming Tang. 2023. CTRL: Connect collaborative and language model for CTR prediction. arXiv:2306.02841. Retrieved from https:\/\/arxiv.org\/abs\/2306.02841"},{"key":"e_1_3_2_39_2","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3615137"},{"key":"e_1_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220023"},{"key":"e_1_3_2_41_2","doi-asserted-by":"publisher","DOI":"10.1145\/2806416.2806603"},{"key":"e_1_3_2_42_2","unstructured":"Yinhan Liu Myle Ott Naman Goyal Jingfei Du Mandar Joshi Danqi Chen Omer Levy Mike Lewis Luke Zettlemoyer and Veselin Stoyanov. 2019. RoBERTa: A robustly optimized BERT pretraining approach. arXiv:1907.11692. Retrieved from https:\/\/arxiv.org\/abs\/1907.11692"},{"key":"e_1_3_2_43_2","unstructured":"Yuhao Luo Shiwei Ma Mingjun Nie Changping Peng Zhangang Lin Jingping Shao and Qianfang Xu. 2024. Domain-aware cross-attention for cross-domain recommendation. arXiv:2401.11705. Retrieved from https:\/\/arxiv.org\/abs\/2401.11705"},{"key":"e_1_3_2_44_2","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220007"},{"key":"e_1_3_2_45_2","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210104"},{"key":"e_1_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.24963\/IJCAI.2017\/343"},{"key":"e_1_3_2_47_2","first-page":"2204","volume-title":"Proceedings of the 27th International Conference on Neural Information Processing Systems","author":"Mnih Volodymyr","year":"2014","unstructured":"Volodymyr Mnih, Nicolas Heess, Alex Graves, and Koray Kavukcuoglu. 2014. Recurrent Models of Visual Attention. In Proceedings of the 27th International Conference on Neural Information Processing Systems. Zoubin Ghahramani, Max Welling, Corinna Cortes, Neil D. Lawrence, and Kilian Q. Weinberger (Eds.), MIT Press, 2204\u20132212. Retrieved from https:\/\/proceedings.neurips.cc\/paper\/2014\/hash\/09c6c3783b4a70054da74f2538ed47c6-Abstract.html"},{"key":"e_1_3_2_48_2","volume-title":"Proceedings of the NeurIPS Efficient Natural Language and Speech Processing Workshop","author":"Muhamed Aashiq","year":"2021","unstructured":"Aashiq Muhamed, Iman Keivanloo, Sujan Perera, James Mracek, Yi Xu, Qingjun Cui, Santosh Rajagopalan, Belinda Zeng, and Trishul Chilimbi. 2021. CTR-BERT: Cost-Effective Knowledge Distillation for Billion-Parameter Teacher Models. In Proceedings of the NeurIPS Efficient Natural Language and Speech Processing Workshop."},{"key":"e_1_3_2_49_2","doi-asserted-by":"publisher","DOI":"10.1145\/3487553.3524206"},{"key":"e_1_3_2_50_2","doi-asserted-by":"publisher","DOI":"10.1016\/0925-2312(91)90023-5"},{"key":"e_1_3_2_51_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1018"},{"key":"e_1_3_2_52_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11431-020-1647-3"},{"key":"e_1_3_2_53_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2016.0151"},{"key":"e_1_3_2_54_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2010.127"},{"key":"e_1_3_2_55_2","first-page":"452","volume-title":"Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence","author":"Rendle Steffen","year":"2009","unstructured":"Steffen Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. 2009. BPR: Bayesian Personalized Ranking from Implicit Feedback. In Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence. Jeff A. Bilmes and Andrew Y. Ng (Eds.), AUAI Press, 452\u2013461. Retrieved from https:\/\/www.auai.org\/uai2009\/papers\/UAI2009_0139_48141db02b9f0b02bc7158819ebfa2c7.pdf"},{"key":"e_1_3_2_56_2","doi-asserted-by":"publisher","DOI":"10.1145\/1242572.1242643"},{"key":"e_1_3_2_57_2","doi-asserted-by":"crossref","unstructured":"Anna Rogers Olga Kovaleva and Anna Rumshisky. 2021. A Primer in BERTology: What We Know about How BERT Works. Transactions of the Association for Computational Linguistics 8 (2021) 842\u2013866.","DOI":"10.1162\/tacl_a_00349"},{"key":"e_1_3_2_58_2","doi-asserted-by":"publisher","DOI":"10.1145\/371920.372071"},{"key":"e_1_3_2_59_2","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3481948"},{"key":"e_1_3_2_60_2","doi-asserted-by":"publisher","DOI":"10.1145\/3459637.3481941"},{"key":"e_1_3_2_61_2","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357925"},{"key":"e_1_3_2_62_2","doi-asserted-by":"publisher","DOI":"10.1145\/3357384.3357925"},{"key":"e_1_3_2_63_2","doi-asserted-by":"publisher","DOI":"10.5555\/2627435"},{"key":"e_1_3_2_64_2","volume-title":"Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS \u201922)","author":"Sung Yi-Lin","year":"2022","unstructured":"Yi-Lin Sung, Jaemin Cho, and Mohit Bansal. 2022. LST: Ladder Side-Tuning for Parameter and Memory Efficient Transfer Learning. In Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS \u201922). Sanmi Koyejo, S. Mohamed, A. Agarwal, Danielle Belgrave, K. Cho, and A. Oh (Eds.), Curran Associates Inc. Retrieved from http:\/\/papers.nips.cc\/paper_files\/paper\/2022\/hash\/54801e196796134a2b0ae5e8adef502f-Abstract-Conference.html"},{"key":"e_1_3_2_65_2","doi-asserted-by":"publisher","DOI":"10.1109\/BIGDATA52589.2021.9671920"},{"key":"e_1_3_2_66_2","doi-asserted-by":"publisher","DOI":"10.1145\/3383313.3412236"},{"key":"e_1_3_2_67_2","unstructured":"Rohan Taori Ishaan Gulrajani Tianyi Zhang Yann Dubois Xuechen Li Carlos Guestrin Percy Liang and Tatsunori B. Hashimoto. 2023. Stanford Alpaca: An Instruction-Following LLaMA Model. Retrieved from https:\/\/github.com\/tatsu-lab\/stanford_alpaca"},{"key":"e_1_3_2_68_2","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard Xavier Martinet Marie-Anne Lachaux Timoth\u00e9e Lacroix Baptiste Rozi\u00e8re Naman Goyal Eric Hambro Faisal Azhar et al. 2023. LLaMA: Open and efficient foundation language models. arXiv:2302.13971. Retrieved from https:\/\/arxiv.org\/abs\/2302.13971"},{"issue":"86","key":"e_1_3_2_69_2","first-page":"2579","article-title":"Visualizing Data Using t-SNE","volume":"9","author":"Maaten Laurens van der","year":"2008","unstructured":"Laurens van der Maaten and Geoffrey Hinton. 2008. Visualizing Data Using t-SNE. Journal of Machine Learning Research 9, 86 (2008), 2579\u20132605. Retrieved from http:\/\/jmlr.org\/papers\/v9\/vandermaaten08a.html","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_70_2","first-page":"5998","volume-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention Is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems. Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (Eds.), Curran Associates Inc., 5998\u20136008. Retrieved from https:\/\/proceedings.neurips.cc\/paper\/2017\/hash\/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html"},{"key":"e_1_3_2_71_2","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599780"},{"key":"e_1_3_2_72_2","unstructured":"Hangyu Wang Jianghao Lin Xiangyang Li Bo Chen Chenxu Zhu Ruiming Tang Weinan Zhang and Yong Yu. 2023. FLIP: Towards fine-grained alignment between ID-based models and pretrained language models for CTR prediction. arXiv:2310.19453. Retrieved from https:\/\/arxiv.org\/abs\/2310.19453"},{"key":"e_1_3_2_73_2","doi-asserted-by":"publisher","DOI":"10.1145\/3124749.3124754"},{"key":"e_1_3_2_74_2","doi-asserted-by":"publisher","DOI":"10.1016\/J.ESWA.2016.09.040"},{"key":"e_1_3_2_75_2","unstructured":"Yunjia Xi Weiwen Liu Jianghao Lin Xiaoling Cai Hong Zhu Jieming Zhu Bo Chen Ruiming Tang Weinan Zhang Rui Zhang and Yong Yu. 2023. Towards open-world recommendation with knowledge augmentation from large language models. arXiv:2306.10933. Retrieved from https:\/\/arxiv.org\/abs\/2306.10933"},{"key":"e_1_3_2_76_2","unstructured":"Mengzhou Xia Tianyu Gao Zhiyuan Zeng and Danqi Chen. 2023. Sheared LLaMA: Accelerating language model pre-training via structured pruning. arXiv:2310.06694. Retrieved from https:\/\/arxiv.org\/abs\/2310.06694"},{"key":"e_1_3_2_77_2","doi-asserted-by":"publisher","DOI":"10.1080\/02681102.2021.1971831"},{"key":"e_1_3_2_78_2","first-page":"802","volume-title":"Proceedings of the 31th International Conference on Machine Learning (ICML \u201914)","volume":"32","author":"Yan Ling","year":"2014","unstructured":"Ling Yan, Wu-Jun Li, Gui-Rong Xue, and Dingyi Han. 2014. Coupled Group Lasso for Web-Scale CTR Prediction in Display Advertising. In Proceedings of the 31th International Conference on Machine Learning (ICML \u201914), JMLR Workshop and Conference Proceedings, Vol. 32, JMLR.org, 802\u2013810. DOI: http:\/\/proceedings.mlr.press\/v32\/yan14.html"},{"key":"e_1_3_2_79_2","doi-asserted-by":"publisher","DOI":"10.1145\/3511808.3557541"},{"key":"e_1_3_2_80_2","doi-asserted-by":"publisher","DOI":"10.1016\/J.IPM.2021.102853"},{"key":"e_1_3_2_81_2","doi-asserted-by":"publisher","DOI":"10.1145\/3488560.3498479"},{"key":"e_1_3_2_82_2","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599758"},{"key":"e_1_3_2_83_2","unstructured":"Wayne Xin Zhao Kun Zhou Junyi Li Tianyi Tang Xiaolei Wang Yupeng Hou Yingqian Min Beichen Zhang Junjie Zhang Zican Dong et al. 2023. A survey of large language models. arXiv:2303.18223. Retrieved from https:\/\/arxiv.org\/abs\/2303.18223"},{"key":"e_1_3_2_84_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33015941"},{"key":"e_1_3_2_85_2","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219823"},{"key":"e_1_3_2_86_2","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3531942"}],"container-title":["ACM Transactions on Information Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3698878","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,15]],"date-time":"2025-07-15T17:11:58Z","timestamp":1752599518000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3698878"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,10]]},"references-count":85,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,9,30]]}},"alternative-id":["10.1145\/3698878"],"URL":"https:\/\/doi.org\/10.1145\/3698878","relation":{},"ISSN":["1046-8188","1558-2868"],"issn-type":[{"value":"1046-8188","type":"print"},{"value":"1558-2868","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,7,10]]},"assertion":[{"value":"2024-02-01","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-09-24","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-07-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}