{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T12:51:54Z","timestamp":1774702314437,"version":"3.50.1"},"reference-count":41,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U24B20171"],"award-info":[{"award-number":["U24B20171"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62394332"],"award-info":[{"award-number":["62394332"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62272023"],"award-info":[{"award-number":["62272023"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Knowledge-Based Systems"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1016\/j.knosys.2026.115742","type":"journal-article","created":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T17:03:36Z","timestamp":1773335016000},"page":"115742","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Asymdapter: Asymmetric adapter architecture for efficient fine-tuning of segment-based trajectory representation models"],"prefix":"10.1016","volume":"340","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9684-611X","authenticated-orcid":false,"given":"Tianxi","family":"Liao","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0000-1407-4818","authenticated-orcid":false,"given":"Xuxiang","family":"Ta","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1989-8231","authenticated-orcid":false,"given":"Liangzhe","family":"Han","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5259-2875","authenticated-orcid":false,"given":"Yi","family":"Xu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0157-1716","authenticated-orcid":false,"given":"Leilei","family":"Sun","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7061-9509","authenticated-orcid":false,"given":"Weifeng","family":"Lv","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.knosys.2026.115742_bib0001","series-title":"2023IEEE 39th International Conference on Data Engineering (ICDE)","first-page":"843","article-title":"Self-supervised trajectory representation learning with temporal regularities and travel semantics","author":"Jiang","year":"2023"},{"key":"10.1016\/j.knosys.2026.115742_bib0002","series-title":"2020IEEE 36th International Conference on Data Engineering (ICDE)","first-page":"2006","article-title":"Learning to rank paths in spatial networks","author":"Yang","year":"2020"},{"key":"10.1016\/j.knosys.2026.115742_bib0003","series-title":"Proceedings of the 31st ACM International Conference on Information & Knowledge Management","first-page":"1501","article-title":"Jointly contrastive representation learning on road network and trajectory","author":"Mao","year":"2022"},{"key":"10.1016\/j.knosys.2026.115742_bib0004","series-title":"Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21","first-page":"3286","article-title":"Unsupervised path representation learning with curriculum negative sampling","author":"Yang","year":"2021"},{"issue":"2","key":"10.1016\/j.knosys.2026.115742_bib0005","doi-asserted-by":"crossref","first-page":"325","DOI":"10.1007\/s00778-022-00748-y","article-title":"Unified route representation learning for multi-modal transportation recommendation with spatiotemporal pre-training","volume":"32","author":"Liu","year":"2023","journal-title":"VLDB J."},{"key":"10.1016\/j.knosys.2026.115742_bib0006","series-title":"Proceedings of the 30th ACM International Conference on Information & Knowledge Management","first-page":"211","article-title":"Robust road network representation learning: when traffic patterns meet traveling semantics","author":"Chen","year":"2021"},{"issue":"9","key":"10.1016\/j.knosys.2026.115742_bib0007","doi-asserted-by":"crossref","first-page":"11782","DOI":"10.1109\/TITS.2024.3371071","article-title":"Multi-faceted route representation learning for travel time estimation","volume":"25","author":"Liao","year":"2024","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"10.1016\/j.knosys.2026.115742_bib0008","first-page":"1","article-title":"Attention is all you need","volume":"30","author":"Vaswani","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115742_bib0009","article-title":"Making foundation models adaptable: a review of parameter-efficient fine-tuning approaches","author":"Chang","year":"2025","journal-title":"TechRxiv"},{"key":"10.1016\/j.knosys.2026.115742_bib0010","unstructured":"Y. Xin, J. Yang, S. Luo, H. Zhou, J. Du, X. Liu, Y. Fan, Q. Li, Y. Du, Parameter-efficient fine-tuning for pre-trained vision models: a survey, arXiv preprint arXiv: 2402.02242 (2024)."},{"key":"10.1016\/j.knosys.2026.115742_bib0011","series-title":"International Conference on Machine Learning","first-page":"2790","article-title":"Parameter-efficient transfer learning for NLP","author":"Houlsby","year":"2019"},{"key":"10.1016\/j.knosys.2026.115742_bib0012","series-title":"16th Conference of the European Chapter of the Associationfor Computational Linguistics, EACL 2021","first-page":"487","article-title":"AdapterFusion: non-destructive task composition for transfer learning","author":"Pfeiffer","year":"2021"},{"key":"10.1016\/j.knosys.2026.115742_bib0013","series-title":"Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing","first-page":"3045","article-title":"The power of scale for parameter-Efficient prompt tuning","author":"Lester","year":"2021"},{"key":"10.1016\/j.knosys.2026.115742_bib0014","series-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)","first-page":"4582","article-title":"Prefix-Tuning: optimizing continuous prompts for generation","author":"Li","year":"2021"},{"key":"10.1016\/j.knosys.2026.115742_bib0015","series-title":"Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing","first-page":"10617","article-title":"An empirical study on the transferability of transformer modules in parameter-efficient fine-tuning","author":"AkbarTajari","year":"2022"},{"key":"10.1016\/j.knosys.2026.115742_bib0016","series-title":"Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)","first-page":"1","article-title":"BitFit: simple parameter-efficient fine-tuning for transformer-based masked language-models","author":"Zaken","year":"2022"},{"key":"10.1016\/j.knosys.2026.115742_bib0017","series-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)","first-page":"7319","article-title":"Intrinsic dimensionality explains the effectiveness of language model fine-tuning","author":"Aghajanyan","year":"2021"},{"issue":"2","key":"10.1016\/j.knosys.2026.115742_bib0018","first-page":"3","article-title":"LORA: low-rank adaptation of large language models","volume":"1","author":"Hu","year":"2022","journal-title":"ICLR"},{"key":"10.1016\/j.knosys.2026.115742_bib0019","series-title":"The Thirteenth International Conference on Learning Representations","article-title":"HiRA: parameter-efficient hadamard high-rank adaptation for large language models","author":"Huang","year":"2025"},{"key":"10.1016\/j.knosys.2026.115742_bib0020","series-title":"Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","first-page":"11662","article-title":"DoRA: enhancing parameter-efficient fine-tuning with dynamic rank distribution","author":"Mao","year":"2024"},{"key":"10.1016\/j.knosys.2026.115742_bib0021","first-page":"1022","article-title":"Compacter: efficient low-rank hypercomplex adapter layers","volume":"34","author":"Karimi Mahabadi","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115742_bib0022","series-title":"12th International Conference on Learning Representations, ICLR 2024","article-title":"DePT: decomposed prompt tuning for parameter-efficient fine-tuning","volume":"2024","author":"Shi","year":"2024"},{"key":"10.1016\/j.knosys.2026.115742_bib0023","series-title":"The Thirteenth International Conference on Learning Representations","article-title":"ADePT: adaptive decomposed prompt tuning for parameter-Efficient fine-tuning","author":"Tang","year":"2025"},{"key":"10.1016\/j.knosys.2026.115742_bib0024","series-title":"The Twelfth International Conference on Learning Representations","article-title":"Bayesian low-rank adaptation for large language models","author":"Yang","year":"2024"},{"issue":"86","key":"10.1016\/j.knosys.2026.115742_bib0025","first-page":"2579","article-title":"Visualizing data using t-SNE","volume":"9","author":"Van der Maaten","year":"2008","journal-title":"J. Mach. Learn. Res."},{"key":"10.1016\/j.knosys.2026.115742_bib0026","first-page":"200","article-title":"Multimodal few-shot learning with frozen language models","volume":"34","author":"Tsimpoukelli","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115742_bib0027","unstructured":"J. Kaplan, S. McCandlish, T. Henighan, T.B. Brown, B. Chess, R. Child, S. Gray, A. Radford, J. Wu, D. Amodei, Scaling laws for neural language models, arXiv preprint arXiv: 2001.08361 (2020)."},{"key":"10.1016\/j.knosys.2026.115742_bib0028","series-title":"International Conference on Learning Representations","article-title":"Outrageously large neural networks: the sparsely-Gated mixture-of-experts layer","author":"Shazeer","year":"2017"},{"key":"10.1016\/j.knosys.2026.115742_bib0029","series-title":"The Thirteenth International Conference on Learning Representations","article-title":"Efficient dictionary learning with switch sparse autoencoders","author":"Mudide","year":"2025"},{"key":"10.1016\/j.knosys.2026.115742_bib0030","series-title":"International Conference on Learning Representations","article-title":"{GS}Hard: scaling giant models with conditional computation and automatic sharding","author":"Lepikhin","year":"2021"},{"key":"10.1016\/j.knosys.2026.115742_bib0031","first-page":"1","article-title":"Switch transformers: scaling to trillion parameter models with simple and efficient sparsity","volume":"23","author":"Fedus","year":"2022","journal-title":"J. Mach. Learn. Res."},{"key":"10.1016\/j.knosys.2026.115742_bib0032","unstructured":"D.P. Kingma, J. Ba, Adam: a method for stochastic optimization, arXiv preprint arXiv: 1412.6980 (2014)."},{"key":"10.1016\/j.knosys.2026.115742_bib0033","doi-asserted-by":"crossref","first-page":"126","DOI":"10.1016\/j.compenvurbsys.2017.05.004","article-title":"OSMnx: New methods for acquiring, constructing, analyzing, and visualizing complex street networks","volume":"65","author":"Boeing","year":"2017","journal-title":"Comput. Environ. Urban Syst."},{"issue":"3","key":"10.1016\/j.knosys.2026.115742_bib0034","doi-asserted-by":"crossref","first-page":"547","DOI":"10.1080\/13658816.2017.1400548","article-title":"Fast map matching, an algorithm integrating hidden Markov model with precomputation","volume":"32","author":"Yang","year":"2018","journal-title":"Int. J. Geogr. Inf. Sci."},{"issue":"1","key":"10.1016\/j.knosys.2026.115742_bib0035","doi-asserted-by":"crossref","first-page":"400","DOI":"10.1109\/TITS.2025.3629800","article-title":"SimPRL: a simple contrastive learning for path representation learning by joint GPS trajectories and road paths","volume":"27","author":"Liao","year":"2026","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"10.1016\/j.knosys.2026.115742_bib0036","unstructured":"V. Lialin, V. Deshpande, A. Rumshisky, Scaling down to scale up: a guide to parameter-efficient fine-tuning, arXiv preprint arXiv: 2303.15647 (2023)."},{"key":"10.1016\/j.knosys.2026.115742_bib0037","first-page":"1950","article-title":"Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning","volume":"35","author":"Liu","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115742_bib0038","series-title":"Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics","first-page":"3274","article-title":"DyloRA: parameter-efficient tuning of pre-trained models using dynamic search-free low-rank adaptation","author":"Valipour","year":"2023"},{"key":"10.1016\/j.knosys.2026.115742_bib0039","series-title":"International Conference on Learning Representations (ICLR 2021)","article-title":"Beyond fully-connected layers with quaternions: parameterization of hypercomplex multiplications with 1\/n parameters","author":"Zhang","year":"2021"},{"key":"10.1016\/j.knosys.2026.115742_bib0040","series-title":"The Twelfth International Conference on Learning Representations","article-title":"When do prompting and prefix-tuning work? a theory of capabilities and limitations","author":"Petrov","year":"2024"},{"key":"10.1016\/j.knosys.2026.115742_bib0041","unstructured":"F.A.I. Research, fvcore: Facebook Vision Core Library, 2024, (https:\/\/github.com\/facebookresearch\/fvcore). Apache License 2.0."}],"container-title":["Knowledge-Based Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126004752?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126004752?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T12:15:27Z","timestamp":1774700127000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0950705126004752"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,5]]},"references-count":41,"alternative-id":["S0950705126004752"],"URL":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115742","relation":{},"ISSN":["0950-7051"],"issn-type":[{"value":"0950-7051","type":"print"}],"subject":[],"published":{"date-parts":[[2026,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Asymdapter: Asymmetric adapter architecture for efficient fine-tuning of segment-based trajectory representation models","name":"articletitle","label":"Article Title"},{"value":"Knowledge-Based Systems","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115742","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"115742"}}