{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T12:02:45Z","timestamp":1743076965411,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":21,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819756650"},{"type":"electronic","value":"9789819756667"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-5666-7_31","type":"book-chapter","created":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T20:37:45Z","timestamp":1722544665000},"page":"368-379","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Lowering Costs and Increasing Benefits Through the Ensemble of LLMs and Machine Learning Models"],"prefix":"10.1007","author":[{"given":"Zhilin","family":"Liu","sequence":"first","affiliation":[]},{"given":"Qun","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Jianjian","family":"Zou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,1]]},"reference":[{"key":"31_CR1","doi-asserted-by":"publisher","unstructured":"Dasarathy, B.V., Sheela, B.V.: A composite classifier system design: concepts and methodology. In: Proceedings of the IEEE, pp. 708\u2013713 (1979). https:\/\/doi.org\/10.1109\/proc.1979.11321","DOI":"10.1109\/proc.1979.11321"},{"key":"31_CR2","doi-asserted-by":"publisher","unstructured":"Wolpert, D.H.: Stacked generalization. Neural Netw. 5(2), 241\u2013259 (1992). https:\/\/doi.org\/10.1016\/s0893-6080(05)80023-1","DOI":"10.1016\/s0893-6080(05)80023-1"},{"key":"31_CR3","unstructured":"Xiao, G., Lin, J., Seznec, M., Wu, H., Demouth, J., Han, S.: SmoothQuant: accurate and \nefficient post-training quantization for large language models. In: 40th International \nConference on Machine Learning, ICML 2023, pp. 38087\u201338099"},{"key":"31_CR4","unstructured":"Frantar, E., Ashkboos, S., Hoefler, T., Alistarh, D.: GPTQ: accurate post-training quantization for generative pre-trained transformers. arXiv (2022)"},{"key":"31_CR5","unstructured":"Hinton, G.E., Vinyals, O., Dean, J.M.: Distilling the knowledge in a neural network. arXiv: Machine Learning (2015)"},{"key":"31_CR6","unstructured":"Gu, Y., Dong, L., Wei, F., Huang, M.: MiniLLM: Knowledge Distillation of Large Language Models. arXiv (2023)"},{"key":"31_CR7","unstructured":"Agarwal, R., Vieillard, N., Stanczyk, P., Ramos, S., Geist, M., Bachem, O.J.a.p.a.: GKD: Generalized Knowledge Distillation for Auto-regressive Sequence Models. arXiv (2023)"},{"key":"31_CR8","unstructured":"Li, S., et al.: Explanations from Large Language Models Make Small Reasoners Better. arXiv (2022)"},{"key":"31_CR9","doi-asserted-by":"crossref","unstructured":"Ho, N., Schmid, L., Yun, S.-Y.: Large language models are reasoning teachers. In: 61st \nAnnual Meeting of the Association for Computational Linguistics, ACL 2023, pp. 14852\u201314882","DOI":"10.18653\/v1\/2023.acl-long.830"},{"key":"31_CR10","doi-asserted-by":"crossref","unstructured":"Cho, J.H., Hariharan, B.: On the Efficacy of Knowledge Distillation. arXiv (2019)","DOI":"10.1109\/ICCV.2019.00489"},{"key":"31_CR11","doi-asserted-by":"crossref","unstructured":"Mirzadeh, S.I., Farajtabar, M., Li, A., Levine, N., Matsukawa, A., Ghasemzadeh, H.: Im proved knowledge distillation via teacher assistant. In: 34th AAAI Conference on Artificial \nIntelligence, AAAI 2020,  pp. 5191\u20135198","DOI":"10.1609\/aaai.v34i04.5963"},{"key":"31_CR12","doi-asserted-by":"publisher","unstructured":"Liu, Y., Yao, X.: Ensemble learning via negative correlation. Neural Netw. 12(10), 1399\u20131404 (1999). https:\/\/doi.org\/10.1016\/s0893-6080(99)00073-8","DOI":"10.1016\/s0893-6080(99)00073-8"},{"key":"31_CR13","doi-asserted-by":"crossref","unstructured":"Ju, C., Bibaut, A., van der Laan, M.J.: The relative performance of ensemble methods with deep convolutional neural networks for image classification. arXiv (2017)","DOI":"10.1080\/02664763.2018.1441383"},{"key":"31_CR14","doi-asserted-by":"publisher","unstructured":"Jain, A., Kumar, A., Susan, S.: Evaluating deep neural network ensembles by majority voting cum meta-learning scheme. In: Reddy, V.S., Prasad, V.K., Wang, J., Reddy, K.T.V. (eds.) Soft Computing and Signal Processing. AISC, vol. 1340, pp. 29\u201337. Springer, Singapore (2022). https:\/\/doi.org\/10.1007\/978-981-16-1249-7_4","DOI":"10.1007\/978-981-16-1249-7_4"},{"key":"31_CR15","doi-asserted-by":"crossref","unstructured":"Qureshi, A.S., Roos, T.: Transfer learning with ensembles of deep neural networks for \nskin cancer detection in imbalanced data sets. Neural Process. Lett. 55, 4461\u20134479 \n(2023)","DOI":"10.1007\/s11063-022-11049-4"},{"key":"31_CR16","doi-asserted-by":"crossref","unstructured":"Jiang, D., Ren, X., Lin, B.Y.: LLM-BLENDER: ensembling large language models with \npairwise ranking and generative fusion. In: 61st Annual Meeting of the Association for \nComputational Linguistics, ACL 2023,  pp. 14165\u201314178","DOI":"10.18653\/v1\/2023.acl-long.792"},{"key":"31_CR17","doi-asserted-by":"crossref","unstructured":"Lu, K., et al.: Routing to the Expert: Efficient Reward-guided Ensemble of Large Language Models. arXiv (2023)","DOI":"10.18653\/v1\/2024.naacl-long.109"},{"key":"31_CR18","unstructured":"Lin, C.-Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74\u201381. Association for Computational Linguistics (2004)"},{"key":"31_CR19","doi-asserted-by":"crossref","unstructured":"Keswani, V., Lease, M., Kenthapadi, K.: Towards unbiased and accurate deferral to multiple experts. In: 4th AAAI\/ACM Conference on Artificial Intelligence, Ethics, and Society, \nAIES 2021, pp. 154\u2013165","DOI":"10.1145\/3461702.3462516"},{"key":"31_CR20","doi-asserted-by":"crossref","unstructured":"Pennington, J., Socher, R., Manning, C.D.: GloVe: global vectors for word representation. \nIn: 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, \npp. 1532\u20131543","DOI":"10.3115\/v1\/D14-1162"},{"key":"31_CR21","unstructured":"Zhao, W.X., et al.: A survey of large language models. arXiv (2023)"}],"container-title":["Lecture Notes in Computer Science","Advanced Intelligent Computing Technology and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-5666-7_31","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,1]],"date-time":"2024-08-01T20:46:53Z","timestamp":1722545213000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-5666-7_31"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819756650","9789819756667"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-5666-7_31","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"1 August 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tianjin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 August 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 August 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icic2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.ic-icc.cn\/2024\/index.htm","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}