{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,31]],"date-time":"2025-12-31T12:20:16Z","timestamp":1767183616062,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":41,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819617098"},{"type":"electronic","value":"9789819617104"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-1710-4_1","type":"book-chapter","created":{"date-parts":[[2025,1,31]],"date-time":"2025-01-31T09:54:51Z","timestamp":1738317291000},"page":"1-13","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Play to\u00a0Your Strengths: Collaborative Intelligence of\u00a0Conventional Recommender Models and\u00a0Large Language Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6883-881X","authenticated-orcid":false,"given":"Yunjia","family":"Xi","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9148-3997","authenticated-orcid":false,"given":"Weiwen","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8953-3203","authenticated-orcid":false,"given":"Jianghao","family":"Lin","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5730-8792","authenticated-orcid":false,"given":"Chuhan","family":"Wu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3750-2533","authenticated-orcid":false,"given":"Bo","family":"Chen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9224-2431","authenticated-orcid":false,"given":"Ruiming","family":"Tang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0127-2425","authenticated-orcid":false,"given":"Weinan","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0281-8271","authenticated-orcid":false,"given":"Yong","family":"Yu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,1]]},"reference":[{"key":"1_CR1","unstructured":"Mindspore (2020). https:\/\/www.mindspore.cn\/"},{"key":"1_CR2","doi-asserted-by":"crossref","unstructured":"Bao, K., Zhang, J., Zhang, Y., Wang, W., Feng, F., He, X.: TALLREC: an effective and efficient tuning framework to align large language model with recommendation. arXiv preprint arXiv:2305.00447 (2023)","DOI":"10.1145\/3604915.3608857"},{"key":"1_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"411","DOI":"10.1007\/978-3-642-36973-5_35","volume-title":"Advances in Information Retrieval","author":"S Cleger-Tamayo","year":"2013","unstructured":"Cleger-Tamayo, S., Fern\u00e1ndez-Luna, J.M., Huete, J.F., Tintarev, N.: Being confident about the quality of the predictions in recommender systems. In: Serdyukov, P., et al. (eds.) ECIR 2013. LNCS, vol. 7814, pp. 411\u2013422. Springer, Heidelberg (2013). https:\/\/doi.org\/10.1007\/978-3-642-36973-5_35"},{"key":"1_CR4","unstructured":"Cui, Z., Ma, J., Zhou, C., Zhou, J., Yang, H.: M6-REC: generative pretrained language models are open-ended recommender systems. arXiv preprint arXiv:2205.08084 (2022)"},{"key":"1_CR5","doi-asserted-by":"crossref","unstructured":"Du, Z., et al.: GLM: general language model pretraining with autoregressive blank infilling. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 320\u2013335 (2022)","DOI":"10.18653\/v1\/2022.acl-long.26"},{"key":"1_CR6","unstructured":"Friedman, L., et al.: Leveraging large language models in conversational recommender systems. arXiv preprint arXiv:2305.07961 (2023)"},{"key":"1_CR7","unstructured":"Gao, Y., Sheng, T., Xiang, Y., Xiong, Y., Wang, H., Zhang, J.: Chat-REC: towards interactive and explainable LLMs-augmented recommender system. arXiv preprint arXiv:2303.14524 (2023)"},{"key":"1_CR8","doi-asserted-by":"crossref","unstructured":"Geng, S., Liu, S., Fu, Z., Ge, Y., Zhang, Y.: Recommendation as language processing (RLP): a unified pretrain, personalized prompt and predict paradigm (p5). In: RecSys, pp. 299\u2013315 (2022)","DOI":"10.1145\/3523227.3546767"},{"key":"1_CR9","doi-asserted-by":"crossref","unstructured":"Gong, Y., Ding, X., Su, Y., Shen, K., Liu, Z., Zhang, G.: An unified search and recommendation foundation model for cold-start scenario. In: Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. CIKM \u201923, pp. 4595\u20134601 (2023)","DOI":"10.1145\/3583780.3614657"},{"issue":"5","key":"1_CR10","doi-asserted-by":"publisher","first-page":"419","DOI":"10.1037\/h0029810","volume":"77","author":"PM Groves","year":"1970","unstructured":"Groves, P.M., Thompson, R.F.: Habituation: a dual-process theory. Psychol. Rev. 77(5), 419 (1970)","journal-title":"Psychol. Rev."},{"key":"1_CR11","unstructured":"Hou, Y., et al.: Large language models are zero-shot rankers for recommender systems. arXiv preprint arXiv:2305.08845 (2023)"},{"key":"1_CR12","unstructured":"Hu, E.J., et al.: Lora: low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)"},{"key":"1_CR13","doi-asserted-by":"crossref","unstructured":"Huang, T., Zhang, Z., Zhang, J.: Fibinet: combining feature importance and bilinear feature interaction for click-through rate prediction. In: RecSys, pp. 169\u2013177 (2019)","DOI":"10.1145\/3298689.3347043"},{"key":"1_CR14","doi-asserted-by":"publisher","first-page":"242","DOI":"10.1016\/j.ins.2022.03.068","volume":"601","author":"N Joorabloo","year":"2022","unstructured":"Joorabloo, N., Jalili, M., Ren, Y.: Improved recommender systems by denoising ratings in highly sparse datasets through individual rating confidence. Inf. Sci. 601, 242\u2013254 (2022)","journal-title":"Inf. Sci."},{"key":"1_CR15","unstructured":"Li, X., Chen, B., Hou, L., Tang, R.: CTRL: connect tabular and language model for CTR prediction. arXiv preprint arXiv:2306.02841 (2023)"},{"key":"1_CR16","doi-asserted-by":"crossref","unstructured":"Li, Z., Cui, Z., Wu, S., Zhang, X., Wang, L.: Fi-GNN: modeling feature interactions via graph neural networks for CTR prediction. In: CIKM, pp. 539\u2013548 (2019)","DOI":"10.1145\/3357384.3357951"},{"key":"1_CR17","doi-asserted-by":"crossref","unstructured":"Lian, J., Zhou, X., Zhang, F., Chen, Z., Xie, X., Sun, G.: XdeepFM: combining explicit and implicit feature interactions for recommender systems. In: KDD, pp. 1754\u20131763 (2018)","DOI":"10.1145\/3219819.3220023"},{"key":"1_CR18","doi-asserted-by":"crossref","unstructured":"Lin, J., et al.: Clickprompt: CTR models are strong prompt generators for adapting language models to CTR prediction. arXiv preprint arXiv:2310.09234 (2023)","DOI":"10.1145\/3589334.3645396"},{"key":"1_CR19","unstructured":"Lin, J., et al.: How can recommender systems benefit from large language models: a survey. arXiv preprint arXiv:2306.05817 (2023)"},{"key":"1_CR20","doi-asserted-by":"crossref","unstructured":"Lin, J., et al.: Map: a modelagnostic pretraining framework for click-through rate prediction. In: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 1384\u20131395 (2023)","DOI":"10.1145\/3580305.3599422"},{"key":"1_CR21","doi-asserted-by":"crossref","unstructured":"Lin, J., et al.: ReLLa: retrieval-enhanced large language models for lifelong sequential behavior comprehension in recommendation. arXiv preprint arXiv:2308.11131 (2023)","DOI":"10.1145\/3589334.3645467"},{"key":"1_CR22","unstructured":"Liu, J., Liu, C., Lv, R., Zhou, K., Zhang, Y.: Is ChatGPT a good recommender? A preliminary study. arXiv preprint arXiv:2304.10149 (2023)"},{"key":"1_CR23","doi-asserted-by":"crossref","unstructured":"Luo, S., et al.: Recranker: instruction tuning large language model as ranker for top-k recommendation. arXiv preprint arXiv:2312.16018 (2023)","DOI":"10.1145\/3705728"},{"key":"1_CR24","doi-asserted-by":"crossref","unstructured":"Lyu, H., Jiang, S., Zeng, H., Xia, Y., Luo, J.: LLM-REC: personalized recommendation via prompting large language models. arXiv preprint arXiv:2307.15780 (2023)","DOI":"10.18653\/v1\/2024.findings-naacl.39"},{"key":"1_CR25","doi-asserted-by":"crossref","unstructured":"Ni, J., Li, J., McAuley, J.: Justifying recommendations using distantly-labeled reviews and fine-grained aspects. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 188-197 (2019)","DOI":"10.18653\/v1\/D19-1018"},{"key":"1_CR26","doi-asserted-by":"crossref","unstructured":"Park, L.A.F., Simoff, S.: Using entropy as a measure of acceptance for multi-label classification. In: XIV (2015)","DOI":"10.1007\/978-3-319-24465-5_19"},{"key":"1_CR27","doi-asserted-by":"crossref","unstructured":"Qin, J., Zhang, W., Wu, X., Jin, J., Fang, Y., Yu, Y.: User behavior retrieval for click-through rate prediction. In: Proceedings of SIGIR, pp. 2347-2356 (2020)","DOI":"10.1145\/3397271.3401440"},{"key":"1_CR28","doi-asserted-by":"crossref","unstructured":"Sanner, S., Balog, K., Radlinski, F.,Wedin, B., Dixon, L.: Large language models are competitive near cold-start recommenders for language-and item-based preferences. In: Proceedings of the 17th ACM Conference on Recommender Systems, pp. 890\u2013896 (2023)","DOI":"10.1145\/3604915.3608845"},{"key":"1_CR29","doi-asserted-by":"crossref","unstructured":"Song, W., et al.: Autoint: automatic feature interaction learning via self-attentive neural networks. In: CIKM, pp. 1161\u20131170 (2019)","DOI":"10.1145\/3357384.3357925"},{"key":"1_CR30","unstructured":"Touvron, H., et al.: Llama: open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)"},{"key":"1_CR31","unstructured":"Touvron, H., et al.: Llama 2: open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)"},{"key":"1_CR32","doi-asserted-by":"crossref","unstructured":"Wang, H., et al.: Flip: towards fine-grained alignment between id-based models and pretrained language models for CTR prediction. arXiv e-prints pp. arXiv-2310 (2023)","DOI":"10.1145\/3640457.3688106"},{"key":"1_CR33","doi-asserted-by":"crossref","unstructured":"Wang, R., et al.: DCN V2: improved deep & cross network and practical lessons for web-scale learning to rank systems. In: WWW, pp. 1785\u20131797 (2021)","DOI":"10.1145\/3442381.3450078"},{"key":"1_CR34","unstructured":"Wang, Z., et al.: Augmentation with projection: towards an effective and efficient data augmentation paradigm for distillation. In: ICML (2023)"},{"issue":"6","key":"1_CR35","doi-asserted-by":"publisher","DOI":"10.1007\/s11704-023-2689-5","volume":"17","author":"M Wen","year":"2023","unstructured":"Wen, M., et al.: Large sequence models for sequential decision-making: a survey. Front. Comput. Sci. 17(6), 176349 (2023)","journal-title":"Front. Comput. Sci."},{"key":"1_CR36","doi-asserted-by":"crossref","unstructured":"Xi, Y., et al.: MemoCRS: memoryenhanced sequential conversational recommender systems with large language models. arXiv preprint arXiv:2407.04960 (2024)","DOI":"10.1145\/3627673.3679599"},{"key":"1_CR37","unstructured":"Xi, Y., et al.: Efficient and deployable knowledge infusion for open-world recommendations via large language models. arXiv preprint arXiv:2408.10520 (2024)"},{"key":"1_CR38","unstructured":"Xi, Y., et al.: Towards open-world recommendation with knowledge augmentation from large language models. arXiv preprint arXiv:2306.10933 (2023)"},{"key":"1_CR39","unstructured":"Xi, Y., et al.: A decoding acceleration framework for industrial deployable LLM-based recommender systems. arXiv preprint arXiv:2408.05676 (2024)"},{"key":"1_CR40","unstructured":"Zhang, Y., et al.: Language models as recommender systems: evaluations and limitations. In: I (Still) Can\u2019t Believe It\u2019s Not Better! NeurIPS 2021 Workshop (2021)"},{"key":"1_CR41","doi-asserted-by":"crossref","unstructured":"Zhou, G., et al.: Deep interest network for click-through rate prediction. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1059\u20131068 (2018)","DOI":"10.1145\/3219819.3219823"}],"container-title":["Lecture Notes in Computer Science","Information Retrieval"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-1710-4_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,31]],"date-time":"2025-01-31T09:55:21Z","timestamp":1738317321000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-1710-4_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819617098","9789819617104"],"references-count":41,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-1710-4_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"1 February 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"CCIR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China Conference on Information Retrieval","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Wuhan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccir2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.cips-ir.org.cn\/CCIR2024\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}