{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T11:40:11Z","timestamp":1750160411409,"version":"3.41.0"},"publisher-location":"Singapore","reference-count":28,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819681853","type":"print"},{"value":"9789819681860","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-8186-0_2","type":"book-chapter","created":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T10:17:41Z","timestamp":1750155461000},"page":"16-27","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DoTA: Weight-Decomposed Tensor Adaptation for\u00a0Large Language Models"],"prefix":"10.1007","author":[{"given":"Xiaolin","family":"Hu","sequence":"first","affiliation":[]},{"given":"Xiang","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Peiyu","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Luan","sequence":"additional","affiliation":[]},{"given":"Bin","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,18]]},"reference":[{"key":"2_CR1","unstructured":"Bershatsky, D., Cherniuk, D., Daulbaev, T., Oseledets, I.: Lotr: Low tensor rank weight adaptation. arXiv preprint arXiv:2402.01376 (2024)"},{"key":"2_CR2","unstructured":"Bommasani, R., Hudson, D.A., Adeli, E., Altman, e.a.: On the Opportunities and Risks of Foundation Models (Jul 2022), arXiv:2108.07258"},{"key":"2_CR3","unstructured":"Chen, Z., Dangovski, R., Loh, C., Dugan, O., Luo, D., Solja\u010di\u0107, M.: Quanta: Efficient high-rank fine-tuning of llms with quantum-informed tensor adaptation. arXiv preprint arXiv:2406.00132 (2024)"},{"key":"2_CR4","unstructured":"Cobbe, K., Kosaraju, V., Bavarian, M., Chen, M., Jun, H., Kaiser, L., Plappert, M., Tworek, J., Hilton, J., Nakano, R., et\u00a0al.: Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168 (2021)"},{"issue":"3","key":"2_CR5","doi-asserted-by":"publisher","first-page":"1084","DOI":"10.1137\/06066518X","volume":"30","author":"V De Silva","year":"2008","unstructured":"De Silva, V., Lim, L.H.: Tensor rank and the ill-posedness of the best low-rank approximation problem. SIAM J. Matrix Anal. Appl. 30(3), 1084\u20131127 (2008)","journal-title":"SIAM J. Matrix Anal. Appl."},{"key":"2_CR6","unstructured":"Dettmers, T., Pagnoni, A., Holtzman, A., Zettlemoyer, L.: Qlora: Efficient finetuning of quantized llms. Advances in Neural Information Processing Systems 36 (2024)"},{"key":"2_CR7","unstructured":"Dubey, A., Jauhri, A., Pandey, A., Kadian, A., Al-Dahle, A., Letman, A., Mathur, A., Schelten, A., Yang, A., Fan, A., et\u00a0al.: The llama 3 herd of models. arXiv preprint arXiv:2407.21783 (2024)"},{"issue":"2","key":"2_CR8","doi-asserted-by":"publisher","DOI":"10.1103\/PhysRevResearch.2.023300","volume":"2","author":"ZF Gao","year":"2020","unstructured":"Gao, Z.F., Cheng, S., He, R.Q., Xie, Z., Zhao, H.H., Lu, Z.Y., Xiang, T.: Compressing deep neural networks by matrix product operators. Physical Review Research 2(2), 023300 (2020)","journal-title":"Physical Review Research"},{"key":"2_CR9","doi-asserted-by":"publisher","first-page":"621","DOI":"10.1007\/978-3-031-53963-3_43","volume-title":"Advances in information and communication","author":"ZF Gao","year":"2024","unstructured":"Gao, Z.F., Liu, P., Zhao, W.X., Xie, Z.Y., Wen, J.R., Lu, Z.Y.: Compression image dataset based on multiple matrix product states. In: Arai, K. (ed.) Advances in information and communication, pp. 621\u2013638. Springer Nature Switzerland, Cham (2024)"},{"key":"2_CR10","unstructured":"Hayou, S., Ghosh, N., Yu, B.: LoRA+: Efficient Low Rank Adaptation of Large Models (Jul 2024), arXiv:2402.12354"},{"key":"2_CR11","unstructured":"He, J., Zhou, C., Ma, X., Berg-Kirkpatrick, T., Neubig, G.: Towards a unified view of parameter-efficient transfer learning. arXiv preprint arXiv:2110.04366 (2021)"},{"key":"2_CR12","unstructured":"Hu, E.J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., Chen, W.: Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)"},{"key":"2_CR13","doi-asserted-by":"crossref","unstructured":"Hu, Z., Wang, L., Lan, Y., Xu, W., Lim, E.P., Bing, L., Xu, X., Poria, S., Lee, R.K.W.: Llm-adapters: An adapter family for parameter-efficient fine-tuning of large language models. arXiv preprint arXiv:2304.01933 (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.319"},{"key":"2_CR14","doi-asserted-by":"crossref","unstructured":"Huang, M., Shen, A., Li, K., Peng, H., Li, B., Yu, H.: EdgeLLM: A Highly Efficient CPU-FPGA Heterogeneous Edge Accelerator for Large Language Models (Jul 2024), arXiv:2407.21325","DOI":"10.1109\/TCSI.2025.3546256"},{"key":"2_CR15","unstructured":"Kalajdzievski, D.: A rank stabilization scaling factor for fine-tuning with lora. arXiv preprint arXiv:2312.03732 (2023)"},{"key":"2_CR16","unstructured":"Li, C., Farkhoor, H., Liu, R., Yosinski, J.: Measuring the intrinsic dimension of objective landscapes. arXiv preprint arXiv:1804.08838 (2018)"},{"key":"2_CR17","unstructured":"Liu, S.Y., Wang, C.Y., Yin, H., Molchanov, P., Wang, Y.C.F., Cheng, K.T., Chen, M.H.: Dora: Weight-decomposed low-rank adaptation. arXiv preprint arXiv:2402.09353 (2024)"},{"key":"2_CR18","unstructured":"Meng, F., Wang, Z., Zhang, M.: Pissa: Principal singular values and singular vectors adaptation of large language models. arXiv preprint arXiv:2404.02948 (2024)"},{"issue":"2","key":"2_CR19","doi-asserted-by":"publisher","DOI":"10.1088\/1367-2630\/12\/2\/025012","volume":"12","author":"B Pirvu","year":"2010","unstructured":"Pirvu, B., Murg, V., Cirac, J.I., Verstraete, F.: Matrix product operator representations. New J. Phys. 12(2), 025012 (2010)","journal-title":"New J. Phys."},{"key":"2_CR20","unstructured":"Qiu, S., Potapczynski, A., Finzi, M., Goldblum, M., Wilson, A.G.: Compute Better Spent: Replacing Dense Layers with Structured Matrices (Jun 2024), arXiv:2406.06248"},{"key":"2_CR21","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language Models are Unsupervised Multitask Learners"},{"key":"2_CR22","unstructured":"Si, C., Wang, X., Yang, X., Xu, Z., Li, Q., Dai, J., Qiao, Y., Yang, X., Shen, W.: FLoRA: Low-Rank Core Space for N-dimension (May 2024), arXiv:2405.14739"},{"key":"2_CR23","unstructured":"Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., et\u00a0al.: Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)"},{"key":"2_CR24","unstructured":"Wang, S., Yu, L., Li, J.: LoRA-GA: Low-Rank Adaptation with Gradient Approximation (Jul 2024), arXiv:2407.05000"},{"key":"2_CR25","doi-asserted-by":"crossref","unstructured":"Yang, Y., Zhou, J., Wong, N., Zhang, Z.: LoRETTA: Low-Rank Economic Tensor-Train Adaptation for Ultra-Low-Parameter Fine-Tuning of Large Language Models (Feb 2024), arXiv:2402.11417","DOI":"10.18653\/v1\/2024.naacl-long.174"},{"key":"2_CR26","unstructured":"Yin, C., Acun, B., Liu, X., Wu, C.J.: TT-Rec: Tensor Train Compression for Deep Learning Recommendation Models (Jan 2021), arXiv:2101.11714"},{"key":"2_CR27","unstructured":"Yu, L., Jiang, W., Shi, H., Yu, J., Liu, Z., Zhang, Y., Kwok, J.T., Li, Z., Weller, A., Liu, W.: Metamath: Bootstrap your own mathematical questions for large language models. arXiv preprint arXiv:2309.12284 (2023)"},{"key":"2_CR28","unstructured":"Zhang, Q., Chen, M., Bukharin, A., Karampatziakis, N., He, P., Cheng, Y., Chen, W., Zhao, T.: AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (Mar 2023)"}],"container-title":["Lecture Notes in Computer Science","Advances in Knowledge Discovery and Data Mining"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-8186-0_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T11:03:01Z","timestamp":1750158181000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-8186-0_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819681853","9789819681860"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-8186-0_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"18 June 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PAKDD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific-Asia Conference on Knowledge Discovery and Data Mining","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Sydney, NSW","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Australia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10 June 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 June 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"pakdd2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/pakdd2025.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}