{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T20:58:28Z","timestamp":1771880308103,"version":"3.50.1"},"reference-count":57,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neural Networks"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1016\/j.neunet.2025.108496","type":"journal-article","created":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T17:09:16Z","timestamp":1766164156000},"page":"108496","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["State-space models are accurate and efficient neural operators for dynamical systems"],"prefix":"10.1016","volume":"197","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0116-9874","authenticated-orcid":false,"given":"Zheyuan","family":"Hu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2485-8987","authenticated-orcid":false,"given":"Nazanin Ahmadi","family":"Daryakenari","sequence":"additional","affiliation":[]},{"given":"Qianli","family":"Shen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1839-7504","authenticated-orcid":false,"given":"Kenji","family":"Kawaguchi","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9713-7120","authenticated-orcid":false,"given":"George Em","family":"Karniadakis","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neunet.2025.108496_bib0001","unstructured":"Ba, J.L., Kiros, J.R., & Hinton, G.E. (2016). Layer normalization. arXiv: 1607.06450."},{"issue":"2","key":"10.1016\/j.neunet.2025.108496_bib0002","doi-asserted-by":"crossref","first-page":"157","DOI":"10.1109\/72.279181","article-title":"Learning long-term dependencies with gradient descent is difficult","volume":"5","author":"Bengio","year":"1994","journal-title":"IEEE Transactions on Neural Networks"},{"key":"10.1016\/j.neunet.2025.108496_bib0003","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in Neural Information Processing Systems"},{"key":"10.1016\/j.neunet.2025.108496_bib0004","first-page":"1","article-title":"Laplace neural operator for solving differential equations","author":"Cao","year":"2024","journal-title":"Nature Machine Intelligence"},{"key":"10.1016\/j.neunet.2025.108496_bib0005","first-page":"24924","article-title":"Choose a Transformer: Fourier or Galerkin","volume":"34","author":"Cao","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"issue":"4","key":"10.1016\/j.neunet.2025.108496_bib0006","doi-asserted-by":"crossref","first-page":"911","DOI":"10.1109\/72.392253","article-title":"Universal approximation to nonlinear operators by neural networks with arbitrary activation functions and its application to dynamical systems","volume":"6","author":"Chen","year":"1995","journal-title":"IEEE Transactions on Neural Networks"},{"key":"10.1016\/j.neunet.2025.108496_bib0007","series-title":"Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)","first-page":"1724","article-title":"Learning phrase representations using RNN encoder\u2013decoder for statistical machine translation","author":"Cho","year":"2014"},{"key":"10.1016\/j.neunet.2025.108496_bib0008","unstructured":"Dao, T., & Gu, A. (2024). Transformers are SSMs: Generalized models and efficient algorithms through structured state space duality. arXiv: 2405.21060."},{"key":"10.1016\/j.neunet.2025.108496_bib0009","article-title":"Cminns: Compartment model informed neural networks\u2013unlocking drug dynamics","volume":"184","author":"Daryakenari","year":"2025","journal-title":"Computers in Biology and Medicine"},{"key":"10.1016\/j.neunet.2025.108496_bib0010","doi-asserted-by":"crossref","first-page":"411","DOI":"10.1016\/j.neunet.2022.06.019","article-title":"Approximation rates of DeepONets for learning operators arising from advection\u2013diffusion equations","volume":"153","author":"Deng","year":"2022","journal-title":"Neural Networks"},{"key":"10.1016\/j.neunet.2025.108496_bib0011","unstructured":"Devlin, J., Chang, M.W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv: 1810.04805."},{"key":"10.1016\/j.neunet.2025.108496_bib0012","doi-asserted-by":"crossref","first-page":"129","DOI":"10.1016\/j.neunet.2023.01.029","article-title":"Approximation bounds for convolutional neural networks in operator learning","volume":"161","author":"Franco","year":"2023","journal-title":"Neural Networks"},{"key":"10.1016\/j.neunet.2025.108496_bib0013","unstructured":"Fu, D.Y., Dao, T., Saab, K.K., Thomas, A.W., Rudra, A., & R\u00e9, C. (2022). Hungry hungry hippos: Towards language modeling with state space models. arXiv: 2212.14052."},{"issue":"6","key":"10.1016\/j.neunet.2025.108496_bib0014","doi-asserted-by":"crossref","first-page":"801","DOI":"10.1016\/S0893-6080(05)80125-X","article-title":"Approximation of dynamical systems by continuous time recurrent neural networks","volume":"6","author":"Funahashi","year":"1993","journal-title":"Neural Networks"},{"key":"10.1016\/j.neunet.2025.108496_bib0015","series-title":"Machine learning in modeling and simulation: Methods and applications","first-page":"219","article-title":"Physics-informed deep neural operator networks","author":"Goswami","year":"2023"},{"key":"10.1016\/j.neunet.2025.108496_bib0016","doi-asserted-by":"crossref","DOI":"10.1016\/j.cma.2023.116674","article-title":"Learning stiff chemical kinetics using extended deep neural operators","volume":"419","author":"Goswami","year":"2024","journal-title":"Computer Methods in Applied Mechanics and Engineering"},{"key":"10.1016\/j.neunet.2025.108496_bib0017","doi-asserted-by":"crossref","DOI":"10.1016\/j.cma.2022.114587","article-title":"A physics-informed variational DeepONet for predicting crack path in quasi-brittle materials","volume":"391","author":"Goswami","year":"2022","journal-title":"Computer Methods in Applied Mechanics and Engineering"},{"key":"10.1016\/j.neunet.2025.108496_bib0018","unstructured":"Gu, A., & Dao, T. (2023). Mamba: Linear-time sequence modeling with selective state spaces. arXiv: 2312.00752."},{"key":"10.1016\/j.neunet.2025.108496_bib0019","unstructured":"Gu, A., Goel, K., & R\u00e9, C. (2021). Efficiently modeling long sequences with structured state spaces. arXiv: 2111.00396."},{"key":"10.1016\/j.neunet.2025.108496_bib0020","series-title":"International conference on machine learning","first-page":"12556","article-title":"GNOT: A general neural operator transformer for operator learning","author":"Hao","year":"2023"},{"key":"10.1016\/j.neunet.2025.108496_bib0021","series-title":"Proceedings of the IEEE conference on computer vision and pattern recognition","first-page":"770","article-title":"Deep residual learning for image recognition","author":"He","year":"2016"},{"issue":"8","key":"10.1016\/j.neunet.2025.108496_bib0022","doi-asserted-by":"crossref","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long short-term memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Computation"},{"key":"10.1016\/j.neunet.2025.108496_bib0023","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2023.107183","article-title":"Augmented physics-informed neural networks (APINNs): A gating network-based soft domain decomposition methodology","volume":"126","author":"Hu","year":"2023","journal-title":"Engineering Applications of Artificial Intelligence"},{"key":"10.1016\/j.neunet.2025.108496_bib0024","doi-asserted-by":"crossref","DOI":"10.1016\/j.neunet.2024.106369","article-title":"Tackling the curse of dimensionality with physics-informed neural networks","author":"Hu","year":"2024","journal-title":"Neural Networks"},{"issue":"6","key":"10.1016\/j.neunet.2025.108496_bib0025","doi-asserted-by":"crossref","first-page":"422","DOI":"10.1038\/s42254-021-00314-5","article-title":"Physics-informed machine learning","volume":"3","author":"Karniadakis","year":"2021","journal-title":"Nature Reviews Physics"},{"key":"10.1016\/j.neunet.2025.108496_bib0026","series-title":"International conference on machine learning","first-page":"5156","article-title":"Transformers are RNNs: Fast autoregressive transformers with linear attention","author":"Katharopoulos","year":"2020"},{"key":"10.1016\/j.neunet.2025.108496_bib0027","series-title":"International Conference on Learning Representations","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2015"},{"key":"10.1016\/j.neunet.2025.108496_bib0028","series-title":"International conference on learning representations","article-title":"Fourier neural operator for parametric partial differential equations","author":"Li","year":"2021"},{"key":"10.1016\/j.neunet.2025.108496_bib0029","unstructured":"Li, Z., Liu-Schiaffini, M., Kovachki, N., Liu, B., Azizzadenesheli, K., Bhattacharya, K., Stuart, A., & Anandkumar, A. (2021b). Learning dissipative dynamics in chaotic systems. arXiv: 2106.06898."},{"key":"10.1016\/j.neunet.2025.108496_bib0030","article-title":"Transformer for partial differential equations\u2019 operator learning","author":"Li","year":"2023","journal-title":"Transactions on Machine Learning Research"},{"key":"10.1016\/j.neunet.2025.108496_bib0031","series-title":"Proceedings of the IEEE\/CVF international conference on computer vision","first-page":"10012","article-title":"Swin transformer: Hierarchical vision transformer using shifted windows","author":"Liu","year":"2021"},{"issue":"2","key":"10.1016\/j.neunet.2025.108496_bib0032","doi-asserted-by":"crossref","first-page":"130","DOI":"10.1175\/1520-0469(1963)020<0130:DNF>2.0.CO;2","article-title":"Deterministic nonperiodic flow","volume":"20","author":"Lorenz","year":"1963","journal-title":"Journal of Atmospheric Sciences"},{"issue":"3","key":"10.1016\/j.neunet.2025.108496_bib0033","doi-asserted-by":"crossref","first-page":"218","DOI":"10.1038\/s42256-021-00302-5","article-title":"Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators","volume":"3","author":"Lu","year":"2021","journal-title":"Nature Machine Intelligence"},{"key":"10.1016\/j.neunet.2025.108496_bib0034","doi-asserted-by":"crossref","DOI":"10.1016\/j.cma.2022.114778","article-title":"A comprehensive and fair comparison of two neural operators (with practical extensions) based on FAIR data","volume":"393","author":"Lu","year":"2022","journal-title":"Computer Methods in Applied Mechanics and Engineering"},{"key":"10.1016\/j.neunet.2025.108496_bib0035","unstructured":"Mehta, H., Gupta, A., Cutkosky, A., & Neyshabur, B. (2022). Long range language modeling via gated state spaces. arXiv: 2206.13947."},{"key":"10.1016\/j.neunet.2025.108496_bib0036","doi-asserted-by":"crossref","unstructured":"Micha\u0142owska, K., Goswami, S., Karniadakis, G.E., & Riemer-S\u00f8rensen, S. (2023). Neural operator learning for long-time integration in dynamical systems with recurrent neural networks. arXiv: 2303.02243.","DOI":"10.1109\/IJCNN60899.2024.10650331"},{"key":"10.1016\/j.neunet.2025.108496_bib0037","first-page":"43177","article-title":"Hyenadna: Long-range genomic sequence modeling at single nucleotide resolution","volume":"36","author":"Nguyen","year":"2023","journal-title":"Advances in Neural Information Processing Systems"},{"key":"10.1016\/j.neunet.2025.108496_bib0038","unstructured":"Ovadia, O., Turkel, E., Kahana, A., & Karniadakis, G.E. (2023). Ditto: Diffusion-inspired temporal transformer operator. arXiv: 2307.09072."},{"key":"10.1016\/j.neunet.2025.108496_bib0039","series-title":"International conference on machine learning","first-page":"1310","article-title":"On the difficulty of training recurrent neural networks","author":"Pascanu","year":"2013"},{"key":"10.1016\/j.neunet.2025.108496_bib0040","doi-asserted-by":"crossref","unstructured":"Patil, S., Li, Z., & Farimani, A.B. (2023). HNO: Hyena neural operator for solving PDEs. arXiv: 2306.16524.","DOI":"10.2139\/ssrn.4509361"},{"key":"10.1016\/j.neunet.2025.108496_bib0041","doi-asserted-by":"crossref","unstructured":"Peng, B., Alcaide, E., Anthony, Q., Albalak, A., Arcadinho, S., Biderman, S., Cao, H., Cheng, X., Chung, M., Grella, M. et al. (2023). RWKV: Reinventing RNNs for the transformer era. arXiv: 2305.13048.","DOI":"10.18653\/v1\/2023.findings-emnlp.936"},{"key":"10.1016\/j.neunet.2025.108496_bib0042","series-title":"Proceedings of the 40th international conference on machine learning","first-page":"28043","article-title":"Hyena hierarchy: Towards larger convolutional language models","volume":"vol. 202","author":"Poli","year":"2023"},{"key":"10.1016\/j.neunet.2025.108496_bib0043","doi-asserted-by":"crossref","first-page":"686","DOI":"10.1016\/j.jcp.2018.10.045","article-title":"Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations","volume":"378","author":"Raissi","year":"2019","journal-title":"Journal of Computational Physics"},{"key":"10.1016\/j.neunet.2025.108496_bib0044","doi-asserted-by":"crossref","first-page":"333","DOI":"10.1016\/j.neunet.2022.07.023","article-title":"Physics guided neural networks for modelling of non-linear dynamics","volume":"154","author":"Robinson","year":"2022","journal-title":"Neural Networks"},{"key":"10.1016\/j.neunet.2025.108496_bib0045","doi-asserted-by":"crossref","first-page":"359","DOI":"10.1016\/j.neunet.2021.08.033","article-title":"Physics-incorporated convolutional recurrent neural networks for source identification and forecasting of dynamical systems","volume":"144","author":"Saha","year":"2021","journal-title":"Neural Networks"},{"key":"10.1016\/j.neunet.2025.108496_bib0046","doi-asserted-by":"crossref","unstructured":"Shih, B., Peyvan, A., Zhang, Z., & Karniadakis, G.E. (2024). Transformers as neural operators for solutions of differential equations with finite regularity. arXiv: 2405.19166.","DOI":"10.1016\/j.cma.2024.117560"},{"key":"10.1016\/j.neunet.2025.108496_bib0047","doi-asserted-by":"crossref","unstructured":"Shukla, K., Oommen, V., Peyvan, A., Penwarden, M., Bravo, L., Ghoshal, A., Kirby, R.M., & Karniadakis, G. E. (2023). Deep neural operators can serve as accurate surrogates for shape optimization: A case study for airfoils. arXiv: 2302.00807.","DOI":"10.1016\/j.engappai.2023.107615"},{"issue":"3","key":"10.1016\/j.neunet.2025.108496_bib0048","doi-asserted-by":"crossref","first-page":"1094","DOI":"10.1158\/0008-5472.CAN-03-2524","article-title":"Predictive pharmacokinetic-pharmacodynamic modeling of tumor growth kinetics in xenograft models after administration of anticancer agents","volume":"64","author":"Simeoni","year":"2004","journal-title":"Cancer Research"},{"key":"10.1016\/j.neunet.2025.108496_bib0049","unstructured":"Sun, Y., Dong, L., Huang, S., Ma, S., Xia, Y., Xue, J., Wang, J., & Wei, F. (2023). Retentive network: A successor to transformer for large language models. arXiv: 2307.08621."},{"key":"10.1016\/j.neunet.2025.108496_bib0050","article-title":"Sequence to sequence learning with neural networks","volume":"27","author":"Sutskever","year":"2014","journal-title":"Advances in Neural Information Processing Systems"},{"key":"10.1016\/j.neunet.2025.108496_bib0051","unstructured":"Tallec, C., & Ollivier, Y. (2018). Can recurrent neural networks warp time?arXiv: 1804.11188."},{"key":"10.1016\/j.neunet.2025.108496_bib0052","article-title":"Attention is all you need","volume":"30","author":"Vaswani","year":"2017"},{"key":"10.1016\/j.neunet.2025.108496_bib0053","doi-asserted-by":"crossref","first-page":"191","DOI":"10.1016\/j.neunet.2020.02.016","article-title":"Backpropagation algorithms and reservoir computing in recurrent neural networks for the forecasting of complex spatiotemporal dynamics","volume":"126","author":"Vlachas","year":"2020","journal-title":"Neural Networks"},{"issue":"40","key":"10.1016\/j.neunet.2025.108496_bib0054","doi-asserted-by":"crossref","DOI":"10.1126\/sciadv.abi8605","article-title":"Learning the solution operator of parametric partial differential equations with physics-informed DeepONets","volume":"7","author":"Wang","year":"2021","journal-title":"Science Advances"},{"issue":"3","key":"10.1016\/j.neunet.2025.108496_bib0055","doi-asserted-by":"crossref","first-page":"285","DOI":"10.1016\/0167-2789(85)90011-9","article-title":"Determining Lyapunov exponents from a time series","volume":"16","author":"Wolf","year":"1985","journal-title":"Physica D: Nonlinear Phenomena"},{"key":"10.1016\/j.neunet.2025.108496_bib0056","series-title":"Proceedings of the European conference on computer vision (ECCV)","first-page":"3","article-title":"Group normalization","author":"Wu","year":"2018"},{"key":"10.1016\/j.neunet.2025.108496_bib0057","doi-asserted-by":"crossref","DOI":"10.1016\/j.cma.2023.116064","article-title":"Reliable extrapolation of deep neural operators informed by physics or sparse observations","volume":"412","author":"Zhu","year":"2023","journal-title":"Computer Methods in Applied Mechanics and Engineering"}],"container-title":["Neural Networks"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0893608025013772?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0893608025013772?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T20:01:47Z","timestamp":1771876907000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0893608025013772"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,5]]},"references-count":57,"alternative-id":["S0893608025013772"],"URL":"https:\/\/doi.org\/10.1016\/j.neunet.2025.108496","relation":{},"ISSN":["0893-6080"],"issn-type":[{"value":"0893-6080","type":"print"}],"subject":[],"published":{"date-parts":[[2026,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"State-space models are accurate and efficient neural operators for dynamical systems","name":"articletitle","label":"Article Title"},{"value":"Neural Networks","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.neunet.2025.108496","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"108496"}}