{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T23:12:05Z","timestamp":1772925125157,"version":"3.50.1"},"reference-count":65,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2025,9,16]],"date-time":"2025-09-16T00:00:00Z","timestamp":1757980800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,9,16]],"date-time":"2025-09-16T00:00:00Z","timestamp":1757980800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/100007567","name":"City University of Hong Kong","doi-asserted-by":"publisher","award":["Strategic Research Grant 7200747"],"award-info":[{"award-number":["Strategic Research Grant 7200747"]}],"id":[{"id":"10.13039\/100007567","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Nat Comput Sci"],"DOI":"10.1038\/s43588-025-00863-0","type":"journal-article","created":{"date-parts":[[2025,9,16]],"date-time":"2025-09-16T09:03:10Z","timestamp":1758013390000},"page":"1080-1090","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Increasing alignment of large language models with language processing in the human brain"],"prefix":"10.1038","volume":"5","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-0763-8854","authenticated-orcid":false,"given":"Changjiang","family":"Gao","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-8523-8639","authenticated-orcid":false,"given":"Zhengwu","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Jiajun","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Ping","family":"Li","sequence":"additional","affiliation":[]},{"given":"Shujian","family":"Huang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5210-6224","authenticated-orcid":false,"given":"Jixing","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,16]]},"reference":[{"key":"863_CR1","doi-asserted-by":"publisher","first-page":"134","DOI":"10.1038\/s42003-022-03036-1","volume":"5","author":"C Caucheteux","year":"2022","unstructured":"Caucheteux, C. & King, J.-R. Brains and algorithms partially converge in natural language processing. Commun. Biol. 5, 134 (2022).","journal-title":"Commun. Biol."},{"key":"863_CR2","doi-asserted-by":"publisher","first-page":"369","DOI":"10.1038\/s41593-022-01026-4","volume":"25","author":"A Goldstein","year":"2022","unstructured":"Goldstein, A. et al. Shared computational principles for language processing in humans and deep language models. Nat. Neurosci. 25, 369\u2013380 (2022).","journal-title":"Nat. Neurosci."},{"key":"863_CR3","doi-asserted-by":"publisher","first-page":"453","DOI":"10.1038\/nature17637","volume":"532","author":"AG Huth","year":"2016","unstructured":"Huth, A. G., de Heer, W. A., Griffiths, T. L., Theunissen, F. E. & Gallant, J. L. Natural speech reveals the semantic maps that tile human cerebral cortex. Nature 532, 453\u2013458 (2016).","journal-title":"Nature"},{"key":"863_CR4","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2105646118","volume":"118","author":"M Schrimpf","year":"2021","unstructured":"Schrimpf, M. et al. The neural architecture of language: integrative modeling converges on predictive processing. Proc. Natl Acad. Sci. USA 118, e2105646118 (2021).","journal-title":"Proc. Natl Acad. Sci. USA"},{"key":"863_CR5","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2307876121","volume":"121","author":"C Shain","year":"2024","unstructured":"Shain, C., Meister, C., Pimentel, T., Cotterell, R. & Levy, R. Large-scale evidence for logarithmic effects of word predictability on reading time. Proc. Natl Acad. Sci. USA 121, e2307876121 (2024).","journal-title":"Proc. Natl Acad. Sci. USA"},{"key":"863_CR6","doi-asserted-by":"publisher","first-page":"eadn7744","DOI":"10.1126\/sciadv.adn7744","volume":"10","author":"S Yu","year":"2024","unstructured":"Yu, S., Gu, C., Huang, K. & Li, P. Predicting the next sentence (not word) in large language models: what model\u2013brain alignment tells us about discourse comprehension. Sci. Adv. 10, eadn7744 (2024).","journal-title":"Sci. Adv."},{"key":"863_CR7","doi-asserted-by":"publisher","first-page":"301","DOI":"10.1016\/j.tics.2004.05.003","volume":"8","author":"JL Elman","year":"2004","unstructured":"Elman, J. L. An alternative view of the mental lexicon. Trends Cogn. Sci. 8, 301\u2013306 (2004).","journal-title":"Trends Cogn. Sci."},{"key":"863_CR8","doi-asserted-by":"publisher","first-page":"416","DOI":"10.1016\/j.neuron.2019.12.002","volume":"105","author":"U Hasson","year":"2020","unstructured":"Hasson, U., Nastase, S. A. & Goldstein, A. Direct fit to nature: an evolutionary perspective on biological and artificial neural networks. Neuron 105, 416\u2013434 (2020).","journal-title":"Neuron"},{"key":"863_CR9","unstructured":"Millet, J. et al. Toward a realistic model of speech processing in the brain with self-supervised learning. In 36th Conference on Neural Information Processing Systems 33428\u201333443 (NeurIPS, 2022)."},{"key":"863_CR10","doi-asserted-by":"publisher","first-page":"858","DOI":"10.1038\/s41593-023-01304-9","volume":"26","author":"J Tang","year":"2023","unstructured":"Tang, J., LeBel, A., Jain, S. & Huth, A. G. Semantic reconstruction of continuous language from non-invasive brain recordings. Nat. Neurosci. 26, 858\u2013866 (2023).","journal-title":"Nat. Neurosci."},{"key":"863_CR11","unstructured":"Radford, A. et al. Language Models are Unsupervised Multitask Learners (OpenAI, 2019)."},{"key":"863_CR12","doi-asserted-by":"publisher","unstructured":"Devlin, J., Chang, M.-W., Lee, K. & Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proc. 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers) 4171\u20134186 (Association for Computational Linguistics, 2019); https:\/\/doi.org\/10.18653\/v1\/N19-1423","DOI":"10.18653\/v1\/N19-1423"},{"key":"863_CR13","doi-asserted-by":"publisher","unstructured":"OpenAI et al. GPT-4 Technical Report. Preprint at https:\/\/doi.org\/10.48550\/arXiv.2303.08774 (2024).","DOI":"10.48550\/arXiv.2303.08774"},{"key":"863_CR14","doi-asserted-by":"publisher","unstructured":"Touvron, H. et al. LLaMA: open and efficient foundation language models. Preprint at https:\/\/doi.org\/10.48550\/arXiv.2302.13971 (2023).","DOI":"10.48550\/arXiv.2302.13971"},{"key":"863_CR15","doi-asserted-by":"publisher","unstructured":"Henighan, T. et al. Scaling laws for autoregressive generative modeling. Preprint at https:\/\/doi.org\/10.48550\/arXiv.2010.14701 (2020).","DOI":"10.48550\/arXiv.2010.14701"},{"key":"863_CR16","doi-asserted-by":"publisher","unstructured":"Hestness, J. et al. Deep learning scaling is predictable, empirically. Preprint at https:\/\/doi.org\/10.48550\/arXiv.1712.00409 (2017).","DOI":"10.48550\/arXiv.1712.00409"},{"key":"863_CR17","doi-asserted-by":"publisher","unstructured":"Kaplan, J. et al. Scaling laws for neural language models. Preprint at https:\/\/doi.org\/10.48550\/arXiv.2001.08361 (2020).","DOI":"10.48550\/arXiv.2001.08361"},{"key":"863_CR18","doi-asserted-by":"crossref","unstructured":"Gao, C., Li, J., Chen, J. & Huang, S. Measuring meaning composition in the human brain with composition scores from large language models. In Proc. 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (eds Ku, L.-W., Martins, A. & Srikumar, V.) 11295\u201311308 (Association for Computational Linguistics, 2024).","DOI":"10.18653\/v1\/2024.acl-long.609"},{"key":"863_CR19","doi-asserted-by":"publisher","unstructured":"Xu, Q. et al. Large language models without grounding recover non-sensorimotor but not sensorimotor features of human concepts. Nat. Hum. Behav. https:\/\/doi.org\/10.1038\/s41562-025-02203-8 (2025).","DOI":"10.1038\/s41562-025-02203-8"},{"key":"863_CR20","unstructured":"Antonello, R., Vaidya, A. & Huth, A. Scaling laws for language encoding models in fMRI. In 37th Conference on Neural Information Processing Systems 21895\u201321907 (NeurIPS 2023)."},{"key":"863_CR21","doi-asserted-by":"publisher","unstructured":"Hong, Z. et al. Scale matters: large language models with billions (rather than millions) of parameters better match neural representations of natural language. eLife https:\/\/doi.org\/10.7554\/eLife.101204.1.sa4 (2024).","DOI":"10.7554\/eLife.101204.1.sa4"},{"key":"863_CR22","first-page":"336","volume":"11","author":"B-D Oh","year":"2023","unstructured":"Oh, B.-D. & Schuler, W. Why does surprisal from larger transformer-based language models provide a poorer fit to human reading times? Trans. Assoc. Comput. Ling. 11, 336\u2013350 (2023).","journal-title":"Trans. Assoc. Comput. Ling."},{"key":"863_CR23","doi-asserted-by":"publisher","unstructured":"Kuribayashi, T. et al. Lower perplexity is not always human-like. In Proc. 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) 5203\u20135217 (Association for Computational Linguistics, Online, 2021); https:\/\/doi.org\/10.18653\/v1\/2021.acl-long.405","DOI":"10.18653\/v1\/2021.acl-long.405"},{"key":"863_CR24","doi-asserted-by":"publisher","unstructured":"Oh, B.-D. & Schuler, W. Transformer-based language model surprisal predicts human reading times best with about two billion training tokens. In Findings of the Association for Computational Linguistics: EMNLP 2023 1915\u20131921 (Association for Computational Linguistics, 2023); https:\/\/doi.org\/10.18653\/v1\/2023.findings-emnlp.128","DOI":"10.18653\/v1\/2023.findings-emnlp.128"},{"key":"863_CR25","doi-asserted-by":"publisher","unstructured":"Kuribayashi, T., Oseki, Y., Brassard, A. & Inui, K. Context limitations make neural language models more human-like. In Proc. 2022 Conference on Empirical Methods in Natural Language Processing (eds Goldberg, Y., Kozareva, Z. & Zhang, Y.) 10421\u201310436 (Association for Computational Linguistics, 2022); https:\/\/doi.org\/10.18653\/v1\/2022.emnlp-main.712","DOI":"10.18653\/v1\/2022.emnlp-main.712"},{"key":"863_CR26","first-page":"3381","volume":"25","author":"HW Chung","year":"2024","unstructured":"Chung, H. W. et al. Scaling instruction-finetuned language models. J. Mach. Learn. Res. 25, 3381\u20133433 (2024).","journal-title":"J. Mach. Learn. Res."},{"key":"863_CR27","unstructured":"Ouyang, L. et al. Training language models to follow instructions with human feedback. In 36th Conference on Neural Information Processing Systems 27730\u201327744 (NeurIPS, 2022)."},{"key":"863_CR28","unstructured":"Sanh, V. et al. Multitask prompted training enables zero-shot task generalization. In International Conference on Learning Representations (2022)."},{"key":"863_CR29","unstructured":"Wei, J. et al. Finetuned language models are zero-shot learners. Preprint at https:\/\/arxiv.org\/abs\/2109.01652 (2021)."},{"key":"863_CR30","unstructured":"Christiano, P. F. et al. Deep reinforcement learning from human preferences. In 31st Conference on Neural Information Processing Systems (NeurIPS, 2017)."},{"key":"863_CR31","unstructured":"Stiennon, N. et al. Learning to summarize with human feedback. In 34th Conference on Neural Information Processing Systems 3008\u20133021 (NeurIPS, 2020)."},{"key":"863_CR32","doi-asserted-by":"publisher","unstructured":"DeepSeek-AI et al. DeepSeek-R1: incentivizing reasoning capability in LLMs via reinforcement learning. Preprint at https:\/\/doi.org\/10.48550\/arXiv.2501.12948 (2025).","DOI":"10.48550\/arXiv.2501.12948"},{"key":"863_CR33","doi-asserted-by":"publisher","unstructured":"Li, P. et al. The Reading Brain project L1 adults. OpenNeuro https:\/\/doi.org\/10.18112\/openneuro.ds003974.v3.0.0 (2022).","DOI":"10.18112\/openneuro.ds003974.v3.0.0"},{"key":"863_CR34","doi-asserted-by":"crossref","unstructured":"Kuribayashi, T., Oseki, Y. & Baldwin, T. Psychometric predictive power of large language models. In Findings of the Association for Computational Linguistics: NAACL 2024 (eds Duh, K. et al.) 1983\u20132005 (Association for Computational Linguistics, 2024).","DOI":"10.18653\/v1\/2024.findings-naacl.129"},{"key":"863_CR35","unstructured":"Chiang, W.-L. et al. Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality (LMSYS, 2023)."},{"key":"863_CR36","unstructured":"Taori, R. et al. Stanford Alpaca: An Instruction-Following LLaMA Model (GitHub, 2023)."},{"key":"863_CR37","doi-asserted-by":"publisher","unstructured":"Vig, J. & Belinkov, Y. Analyzing the structure of attention in a transformer language model. In Proc. 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP (eds Linzen, T., Chrupa\u0142a, G., Belinkov, Y. & Hupkes, D.) 63\u201376 (Association for Computational Linguistics, 2019); https:\/\/doi.org\/10.18653\/v1\/W19-4808","DOI":"10.18653\/v1\/W19-4808"},{"key":"863_CR38","doi-asserted-by":"publisher","unstructured":"Clark, K., Khandelwal, U., Levy, O. & Manning, C. D. What does BERT look at? An analysis of BERT\u2019s attention. In Proc. 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP (eds Linzen, T., Chrupa\u0142a, G., Belinkov, Y. & Hupkes, D.) 276\u2013286 (Association for Computational Linguistics, 2019); https:\/\/doi.org\/10.18653\/v1\/W19-4828","DOI":"10.18653\/v1\/W19-4828"},{"key":"863_CR39","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1016\/S1364-6613(99)01418-7","volume":"4","author":"SP Liversedge","year":"2000","unstructured":"Liversedge, S. P. & Findlay, J. M. Saccadic eye movements and cognition. Trends Cogn. Sci. 4, 6\u201314 (2000).","journal-title":"Trends Cogn. Sci."},{"key":"863_CR40","doi-asserted-by":"publisher","first-page":"46","DOI":"10.1038\/s41539-025-00337-y","volume":"10","author":"C Gu","year":"2025","unstructured":"Gu, C., Nastase, S. A., Zada, Z. & Li, P. Reading comprehension in L1 and L2 readers: neurocomputational mechanisms revealed through large language models. npj Sci. Learn. 10, 46 (2025).","journal-title":"npj Sci. Learn."},{"key":"863_CR41","doi-asserted-by":"publisher","first-page":"E6256","DOI":"10.1073\/pnas.1612132113","volume":"113","author":"E Fedorenko","year":"2016","unstructured":"Fedorenko, E. et al. Neural correlate of the construction of sentence meaning. Proc. Natl Acad. Sci. USA 113, E6256\u2013E6262 (2016).","journal-title":"Proc. Natl Acad. Sci. USA"},{"key":"863_CR42","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-018-03068-4","volume":"9","author":"F Pereira","year":"2018","unstructured":"Pereira, F. et al. Toward a universal decoder of linguistic meaning from brain activation. Nat. Commun. 9, 963 (2018).","journal-title":"Nat. Commun."},{"key":"863_CR43","doi-asserted-by":"publisher","first-page":"1032","DOI":"10.1016\/j.tics.2023.08.003","volume":"27","author":"R Ryskin","year":"2023","unstructured":"Ryskin, R. & Nieuwland, M. S. Prediction during language comprehension: what is next? Trends Cogn. Sci. 27, 1032\u20131052 (2023).","journal-title":"Trends Cogn. Sci."},{"key":"863_CR44","unstructured":"Aw, K. L. & Toneva, M. Training language models to summarize narratives improves brain alignment. In International Conference on Learning Representations (2023)."},{"key":"863_CR45","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1007\/s11145-017-9781-x","volume":"31","author":"DJ Follmer","year":"2018","unstructured":"Follmer, D. J., Fang, S.-Y., Clariana, R. B., Meyer, B. J. F. & Li, P. What predicts adult readers\u2019 understanding of STEM texts? Read. Writ. 31, 185\u2013214 (2018).","journal-title":"Read. Writ."},{"key":"863_CR46","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1016\/j.neuroimage.2013.04.127","volume":"80","author":"MF Glasser","year":"2013","unstructured":"Glasser, M. F. et al. The minimal preprocessing pipelines for the Human Connectome Project. Neuroimage 80, 105\u2013124 (2013).","journal-title":"Neuroimage"},{"key":"863_CR47","doi-asserted-by":"publisher","first-page":"2186","DOI":"10.1038\/s41596-020-0327-3","volume":"15","author":"O Esteban","year":"2020","unstructured":"Esteban, O. et al. Analysis of task-based functional MRI data preprocessed with fMRIPrep. Nat. Protoc. 15, 2186\u20132202 (2020).","journal-title":"Nat. Protoc."},{"key":"863_CR48","unstructured":"EyeLink 1000 Plus Long Range MRI-Compatible Eye-Tracker (SR Research EyeLink, 2016)."},{"key":"863_CR49","doi-asserted-by":"publisher","unstructured":"Gemma Team. Gemma: Open models based on Gemini research and technology. Preprint at https:\/\/doi.org\/10.48550\/arXiv.2403.08295 (2024).","DOI":"10.48550\/arXiv.2403.08295"},{"key":"863_CR50","doi-asserted-by":"publisher","unstructured":"Jiang, A. Q. et al. Mistral 7B. Preprint at https:\/\/doi.org\/10.48550\/arXiv.2310.06825 (2023).","DOI":"10.48550\/arXiv.2310.06825"},{"key":"863_CR51","unstructured":"Brown, T. et al. Language models are few-shot learners. In 34th Conference on Neural Information Processing Systems 1877\u20131901 (NeurIPS, 2020)."},{"key":"863_CR52","doi-asserted-by":"publisher","unstructured":"Wang, Y. et al. Self-instruct: aligning language models with self-generated instructions. In Proc. 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (eds Rogers, A., Boyd-Graber, J. & Okazaki, N.) 13484\u201313508 (Association for Computational Linguistics, 2023); https:\/\/doi.org\/10.18653\/v1\/2023.acl-long.754","DOI":"10.18653\/v1\/2023.acl-long.754"},{"key":"863_CR53","unstructured":"Hendrycks, D. et al. Measuring massive multitask language understanding. In International Conference on Learning Representations (2021)."},{"key":"863_CR54","doi-asserted-by":"crossref","unstructured":"Kudo, T. & Richardson, J. SentencePiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. In Proc. 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations 66\u201371 (Association for Computational Linguistics, 2018).","DOI":"10.18653\/v1\/D18-2012"},{"key":"863_CR55","doi-asserted-by":"publisher","first-page":"30046","DOI":"10.1073\/pnas.1907367117","volume":"117","author":"CD Manning","year":"2020","unstructured":"Manning, C. D., Clark, K., Hewitt, J., Khandelwal, U. & Levy, O. Emergent linguistic structure in artificial neural networks trained by self-supervision. Proc. Natl Acad. Sci. USA 117, 30046\u201330054 (2020).","journal-title":"Proc. Natl Acad. Sci. USA"},{"key":"863_CR56","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1214\/aoms\/1177729694","volume":"22","author":"S Kullback","year":"1951","unstructured":"Kullback, S. & Leibler, R. A. On information and sufficiency. Ann. Math. Stat. 22, 79\u201386 (1951).","journal-title":"Ann. Math. Stat."},{"key":"863_CR57","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1016\/j.jneumeth.2007.03.024","volume":"164","author":"E Maris","year":"2007","unstructured":"Maris, E. & Oostenveld, R. Nonparametric statistical testing of EEG- and MEG-data. J. Neurosci. Methods 164, 177\u2013190 (2007).","journal-title":"J. Neurosci. Methods"},{"key":"863_CR58","unstructured":"Paszke, A. et al. PyTorch: an imperative style, high-performance deep learning library. In Proc. 33rd International Conference on Neural Information Processing Systems 8026\u20138037 (Curran Associates, 2019)."},{"key":"863_CR59","doi-asserted-by":"publisher","first-page":"446","DOI":"10.1016\/j.neuroimage.2013.10.027","volume":"86","author":"A Gramfort","year":"2014","unstructured":"Gramfort, A. et al. MNE software for processing MEG and EEG data. NeuroImage 86, 446\u2013460 (2014).","journal-title":"NeuroImage"},{"key":"863_CR60","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1038\/s41592-019-0686-2","volume":"17","author":"P Virtanen","year":"2020","unstructured":"Virtanen, P. et al. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nat. Methods 17, 261\u2013272 (2020).","journal-title":"Nat. Methods"},{"key":"863_CR61","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-024-49173-5","volume":"15","author":"S Kumar","year":"2024","unstructured":"Kumar, S. et al. Shared functional specialization in transformer-based language models and the human brain. Nat. Commun. 15, 5523 (2024).","journal-title":"Nat. Commun."},{"key":"863_CR62","unstructured":"Ma, Z. & Gao, C. RiverGao\/scaling_finetuning (GitHub, 2025); https:\/\/github.com\/RiverGao\/scaling_finetuning"},{"key":"863_CR63","doi-asserted-by":"publisher","unstructured":"Gao, C. & Ma, Z. RiverGao\/scaling_finetuning: scaling_finetuning_v0.0.0. Zenodo https:\/\/doi.org\/10.5281\/zenodo.15788717 (2025).","DOI":"10.5281\/zenodo.15788717"},{"key":"863_CR64","doi-asserted-by":"publisher","DOI":"10.1038\/s41597-025-05158-7","volume":"12","author":"Q Wang","year":"2025","unstructured":"Wang, Q. et al. Le Petit Prince (LPP) multi-talker: naturalistic 7 T fMRI and EEG dataset. Sci Data 12, 829 (2025).","journal-title":"Sci Data"},{"key":"863_CR65","unstructured":"manu\/mistral-7B-v0.1 (HuggingFace, 2025); https:\/\/huggingface.co\/manu\/mistral-7B-v0.1#training-data"}],"container-title":["Nature Computational Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s43588-025-00863-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s43588-025-00863-0","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s43588-025-00863-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,21]],"date-time":"2025-11-21T04:02:25Z","timestamp":1763697745000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s43588-025-00863-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,16]]},"references-count":65,"journal-issue":{"issue":"11","published-online":{"date-parts":[[2025,11]]}},"alternative-id":["863"],"URL":"https:\/\/doi.org\/10.1038\/s43588-025-00863-0","relation":{},"ISSN":["2662-8457"],"issn-type":[{"value":"2662-8457","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,9,16]]},"assertion":[{"value":"20 September 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 August 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 September 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}