{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T07:12:36Z","timestamp":1776150756779,"version":"3.50.1"},"reference-count":63,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62350710797"],"award-info":[{"award-number":["62350710797"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neural Networks"],"published-print":{"date-parts":[[2026,7]]},"DOI":"10.1016\/j.neunet.2026.108733","type":"journal-article","created":{"date-parts":[[2026,2,18]],"date-time":"2026-02-18T16:18:12Z","timestamp":1771431492000},"page":"108733","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Enhancing progressive ensemble learning via normalized extra-Gradient initialization"],"prefix":"10.1016","volume":"199","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4099-552X","authenticated-orcid":false,"given":"Zheshun","family":"Wu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7515-8492","authenticated-orcid":false,"given":"Yu","family":"Pan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4517-5379","authenticated-orcid":false,"given":"Dun","family":"Zeng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7570-5756","authenticated-orcid":false,"given":"Qifan","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5550-6461","authenticated-orcid":false,"given":"Zenglin","family":"Xu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6209-6886","authenticated-orcid":false,"given":"Jie","family":"Liu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neunet.2026.108733_bib0001","unstructured":"Agarwal, N., Awasthi, P., Kale, S., & Zhao, E. (2024). Stacking as accelerated gradient descent. arXiv: 2403.04978."},{"key":"10.1016\/j.neunet.2026.108733_bib0002","unstructured":"Allingham, J. U., Wenzel, F., Mariet, Z. E., Mustafa, B., Puigcerver, J., Houlsby, N., Jerfel, G., Fortuin, V., Lakshminarayanan, B., Snoek, J. et al. (2021). Sparse moes meet efficient ensembles. arXiv: 2110.03360."},{"key":"10.1016\/j.neunet.2026.108733_bib0003","series-title":"International conference on machine learning, ICML 2022, 17\u201323 july 2022, baltimore, maryland, USA","first-page":"948","article-title":"Understanding gradient descent on the edge of stability in deep learning","volume":"vol. 162","author":"Arora","year":"2022"},{"issue":"3","key":"10.1016\/j.neunet.2026.108733_bib0004","doi-asserted-by":"crossref","first-page":"167","DOI":"10.1016\/S0167-6377(02)00231-6","article-title":"Mirror descent and nonlinear projected subgradient methods for convex optimization","volume":"31","author":"Beck","year":"2003","journal-title":"Operations Research Letters"},{"key":"10.1016\/j.neunet.2026.108733_bib0005","series-title":"Advances in neural information processing systems 19, proceedings of the twentieth annual conference on neural information processing systems, vancouver, british columbia, canada, december 4\u20137, 2006","first-page":"153","article-title":"Greedy layer-wise training of deep networks","author":"Bengio","year":"2006"},{"key":"10.1016\/j.neunet.2026.108733_bib0006","unstructured":"Caldas, S., Duddu, S. M. K., Wu, P., Li, T., Kone\u010dn\u1ef3, J., McMahan, H. B., Smith, V., & Talwalkar, A. (2018). Leaf: A benchmark for federated settings. arXiv: 1812.01097."},{"key":"10.1016\/j.neunet.2026.108733_bib0007","unstructured":"Chen, T., Goodfellow, I., & Shlens, J. (2015). Net2net: Accelerating learning via knowledge transfer. arXiv: 1511.05641."},{"key":"10.1016\/j.neunet.2026.108733_bib0008","series-title":"Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, san francisco, ca, usa, august 13\u201317, 2016","first-page":"785","article-title":"Xgboost: A scalable tree boosting system","author":"Chen","year":"2016"},{"key":"10.1016\/j.neunet.2026.108733_bib0009","series-title":"International conference on learning representations","article-title":"Gradient descent on neural networks typically occurs at the edge of stability","author":"Cohen","year":"2021"},{"key":"10.1016\/j.neunet.2026.108733_bib0010","series-title":"Advances in neural information processing systems 32: Annual conference on neural information processing systems 2019, neurIPS 2019, december 8\u201314, 2019, vancouver, BC, canada","first-page":"12624","article-title":"Metainit: Initializing learning by learning to initialize","author":"Dauphin","year":"2019"},{"key":"10.1016\/j.neunet.2026.108733_bib0011","series-title":"Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: human language technologies, NAACL-HLT 2019, minneapolis, mn, usa, june 2\u20137, 2019, volume 1 (long and short papers)","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2019"},{"key":"10.1016\/j.neunet.2026.108733_bib0012","doi-asserted-by":"crossref","DOI":"10.1016\/j.neunet.2025.107558","article-title":"Progressive fine-to-coarse reconstruction for accurate low-bit post-training quantization in vision transformers","volume":"189","author":"Ding","year":"2025","journal-title":"Neural Networks"},{"key":"10.1016\/j.neunet.2026.108733_bib0013","series-title":"Advances in neural information processing systems 38: Annual conference on neural information processing systems 2024, neurIPS 2024, vancouver, BC, canada, december 10, - 15, 2024","article-title":"Stacking your transformers: A closer look at model growth for efficient LLM pre-training","author":"Du","year":"2024"},{"key":"10.1016\/j.neunet.2026.108733_bib0014","series-title":"International conference on learning representations","article-title":"Gradmax: Growing neural networks using gradient information","author":"Evci","year":"2022"},{"key":"10.1016\/j.neunet.2026.108733_bib0015","first-page":"1189","article-title":"Greedy function approximation: A gradient boosting machine","author":"Friedman","year":"2001","journal-title":"Annals of Statistics"},{"key":"10.1016\/j.neunet.2026.108733_bib0016","series-title":"34Th british machine vision conference 2023, BMVC 2023, aberdeen, uk, november 20\u201324, 2023","first-page":"871","article-title":"On the lipschitz constant of deep networks and double descent","author":"Gamba","year":"2023"},{"key":"10.1016\/j.neunet.2026.108733_bib0017","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2022.105151","article-title":"Ensemble deep learning: A review","volume":"115","author":"Ganaie","year":"2022","journal-title":"Engineering Applications of Artificial Intelligence"},{"key":"10.1016\/j.neunet.2026.108733_bib0018","series-title":"Proceedings of the 29th international conference on computational linguistics, COLING 2022, gyeongju, republic of korea, october 12\u201317, 2022","first-page":"3263","article-title":"Parameter-efficient mixture-of-experts architecture for pre-trained language models","author":"Gao","year":"2022"},{"key":"10.1016\/j.neunet.2026.108733_bib0019","series-title":"Proceedings of the 36th international conference on machine learning, ICML 2019, 9\u201315 june 2019, long beach, california, USA","first-page":"2337","article-title":"Efficient training of BERT by progressively stacking","volume":"vol. 97","author":"Gong","year":"2019"},{"key":"10.1016\/j.neunet.2026.108733_bib0020","series-title":"Deep learning","volume":"vol. 1","author":"Goodfellow","year":"2016"},{"key":"10.1016\/j.neunet.2026.108733_bib0021","unstructured":"Goodfellow, I. J., Vinyals, O., & Saxe, A. M. (2015). Qualitatively characterizing neural network optimization problems."},{"key":"10.1016\/j.neunet.2026.108733_bib0022","series-title":"International conference on artificial intelligence and statistics, AISTATS 2022, 28\u201330 march 2022, virtual event","first-page":"366","article-title":"Extragradient method: o(1\/k) last-iterate convergence for monotone variational inequalities and connections with cocoercivity","volume":"vol. 151","author":"Gorbunov","year":"2022"},{"key":"10.1016\/j.neunet.2026.108733_bib0023","unstructured":"Gu, X., Liu, L., Yu, H., Li, J., Chen, C., & Han, J. (2020). On the transformer growth for progressive bert training. arXiv: 2010.12562."},{"issue":"10","key":"10.1016\/j.neunet.2026.108733_bib0024","doi-asserted-by":"crossref","first-page":"306","DOI":"10.1007\/s10462-025-11309-4","article-title":"Advances in artificial intelligence for olfaction and gustation: A comprehensive review","volume":"58","author":"Hao","year":"2025","journal-title":"Artificial Intelligence Review"},{"issue":"8","key":"10.1016\/j.neunet.2026.108733_bib0025","doi-asserted-by":"crossref","first-page":"1772","DOI":"10.1109\/JAS.2023.123360","article-title":"An isomerism learning model to solve time-varying problems through intelligent collaboration","volume":"10","author":"Hao","year":"2023","journal-title":"IEEE\/CAA Journal of Automatica Sinica"},{"issue":"1","key":"10.1016\/j.neunet.2026.108733_bib0026","doi-asserted-by":"crossref","first-page":"249","DOI":"10.1109\/TNNLS.2023.3274912","article-title":"A novel public sentiment analysis method based on an isomerism learning model via multiphase processing","volume":"36","author":"Hao","year":"2023","journal-title":"IEEE Transactions on Neural Networks and Learning Systems"},{"key":"10.1016\/j.neunet.2026.108733_bib0027","series-title":"Aaai-25, sponsored by the association for the advancement of artificial intelligence, february 25, - march 4, 2025, philadelphia, pa, USA","first-page":"17077","article-title":"DCHM: Dynamic collaboration of heterogeneous models through isomerism learning in a blockchain-powered federated learning framework","author":"Hao","year":"2025"},{"key":"10.1016\/j.neunet.2026.108733_bib0028","series-title":"2025\u202fIEEE International conference on acoustics, speech and signal processing, ICASSP 2025, hyderabad, india, april 6\u201311, 2025","first-page":"1","article-title":"Transforming classification with federated learning on blockchain: A unique model integration approach","author":"Hao","year":"2025"},{"key":"10.1016\/j.neunet.2026.108733_bib0029","series-title":"International conference on learning representations","article-title":"Identity matters in deep learning","author":"Hardt","year":"2017"},{"issue":"7","key":"10.1016\/j.neunet.2026.108733_bib0030","doi-asserted-by":"crossref","first-page":"1527","DOI":"10.1162\/neco.2006.18.7.1527","article-title":"A fast learning algorithm for deep belief nets","volume":"18","author":"Hinton","year":"2006","journal-title":"Neural Computation"},{"key":"10.1016\/j.neunet.2026.108733_bib0031","doi-asserted-by":"crossref","DOI":"10.1016\/j.neunet.2024.106106","article-title":"Self-supervised anomaly detection in computer vision and beyond: A survey and outlook","volume":"172","author":"Hojjati","year":"2024","journal-title":"Neural Networks"},{"key":"10.1016\/j.neunet.2026.108733_bib0032","series-title":"ICLR","article-title":"Progressive growing of GANs for improved quality, stability, and variation","author":"Karras","year":"2018"},{"key":"10.1016\/j.neunet.2026.108733_bib0033","series-title":"Advances in neural information processing systems 29: Annual conference on neural information processing systems 2016, december 5\u201310, 2016, barcelona, spain","first-page":"586","article-title":"Deep learning without poor local minima","author":"Kawaguchi","year":"2016"},{"issue":"4","key":"10.1016\/j.neunet.2026.108733_bib0034","first-page":"35","article-title":"Extragradient method for finding saddle points and other problems","volume":"13","author":"Korpelevich","year":"1977","journal-title":"Matekon: translations of Russian & East European mathematical Economics"},{"issue":"2","key":"10.1016\/j.neunet.2026.108733_bib0035","doi-asserted-by":"crossref","first-page":"425","DOI":"10.1007\/s10107-011-0442-6","article-title":"Validation analysis of mirror descent stochastic approximation method","volume":"134","author":"Lan","year":"2012","journal-title":"Mathematical Programming"},{"key":"10.1016\/j.neunet.2026.108733_bib0036","unstructured":"Li, H., Xu, Z., Taylor, G., Studer, C., & Goldstein, T. (2018). Visualizing the loss landscape of neural nets."},{"key":"10.1016\/j.neunet.2026.108733_bib0037","series-title":"ICLR","article-title":"Progressive learning and disentanglement of hierarchical representations","author":"Li","year":"2020"},{"key":"10.1016\/j.neunet.2026.108733_bib0038","series-title":"Advances in neural information processing systems 12, <upper-case>[NIPS conference, denver, colorado, USA, november 29, - december 4, 1999]<\/upper-case>","first-page":"512","article-title":"Boosting algorithms as gradient descent","author":"Mason","year":"1999"},{"key":"10.1016\/j.neunet.2026.108733_bib0039","series-title":"Dokl. akad. nauk. SSSR","first-page":"543","article-title":"A method for unconstrained convex minimization problem with the rate of convergence o (1\/k2)","volume":"vol. 269","author":"Nesterov","year":"1983"},{"key":"10.1016\/j.neunet.2026.108733_bib0040","series-title":"Advances in neural information processing systems 35: Annual conference on neural information processing systems 2022, neurIPS 2022, new orleans, LA, USA, november 28, - december 9, 2022","article-title":"Signal propagation in transformers: Theoretical perspectives and the role of rank collapse","author":"Noci","year":"2022"},{"key":"10.1016\/j.neunet.2026.108733_bib0041","first-page":"18860","article-title":"Preparing lessons for progressive training on language models","author":"Pan","year":"2024"},{"key":"10.1016\/j.neunet.2026.108733_bib0042","series-title":"Advances in neural information processing systems 36: Annual conference on neural information processing systems 2023, neurIPS 2023, new orleans, LA, USA, december 10, - 16, 2023","article-title":"Reusing pretrained models by multi-linear operators for efficient training","author":"Pan","year":"2023"},{"key":"10.1016\/j.neunet.2026.108733_bib0043","series-title":"The thirteenth international conference on learning representations","article-title":"Efficient stagewise pretraining via progressive subnetworks","author":"Panigrahi","year":"2025"},{"key":"10.1016\/j.neunet.2026.108733_bib0044","series-title":"IEEE\/CVF Winter conference on applications of computer vision, WACV 2024, waikoloa, hi, usa, january 3\u20138, 2024","first-page":"2788","article-title":"Mixturegrowth: Growing neural networks by recombining learned parameters","author":"Pham","year":"2024"},{"issue":"8","key":"10.1016\/j.neunet.2026.108733_bib0045","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"key":"10.1016\/j.neunet.2026.108733_bib0046","first-page":"140:1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"Journal of Machine Learning Research : JMLR"},{"key":"10.1016\/j.neunet.2026.108733_bib0047","series-title":"International conference on machine learning, ICML 2023, 23\u201329 july 2023, honolulu, hawaii, USA","first-page":"14553","article-title":"Efficient training of language models using few-shot learning","volume":"vol. 202","author":"Reddi","year":"2023"},{"key":"10.1016\/j.neunet.2026.108733_bib0048","series-title":"Advances in neural information processing systems 34: Annual conference on neural information processing systems 2021, neurIPS 2021, december 6\u201314, 2021, virtual","first-page":"8583","article-title":"Scaling vision with sparse mixture of experts","author":"Riquelme","year":"2021"},{"key":"10.1016\/j.neunet.2026.108733_bib0049","series-title":"International conference on machine learning, ICML 2022, 17\u201323 july 2022, baltimore, maryland, USA","first-page":"19310","article-title":"Convergence rates of non-convex stochastic gradient descent under a generic lojasiewicz condition and local smoothness","volume":"vol. 162","author":"Scaman","year":"2022"},{"issue":"1","key":"10.1016\/j.neunet.2026.108733_bib0050","doi-asserted-by":"crossref","first-page":"164","DOI":"10.1108\/03684921311295547","article-title":"Boosting: Foundations and algorithms","volume":"42","author":"Schapire","year":"2013","journal-title":"Kybernetes"},{"issue":"3","key":"10.1016\/j.neunet.2026.108733_bib0051","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3700439","article-title":"On efficient training of large-scale deep learning models","volume":"57","author":"Shen","year":"2024","journal-title":"ACM Computing Surveys"},{"key":"10.1016\/j.neunet.2026.108733_bib0052","series-title":"International conference on machine learning, ICML 2022, 17\u201323 july 2022, baltimore, maryland, USA","first-page":"19893","article-title":"Staged training for transformer language models","volume":"vol. 162","author":"Shen","year":"2022"},{"key":"10.1016\/j.neunet.2026.108733_bib0053","doi-asserted-by":"crossref","DOI":"10.1016\/j.neunet.2024.106630","article-title":"Snn-bert: Training-efficient spiking neural networks for energy-efficient bert","volume":"180","author":"Su","year":"2024","journal-title":"Neural Networks"},{"key":"10.1016\/j.neunet.2026.108733_bib0054","series-title":"International conference on machine learning, ICML 2022, 17\u201323 july 2022, baltimore, maryland, USA","first-page":"23034","article-title":"Progfed: Effective, communication, and computation efficient federated learning by progressive training","volume":"vol. 162","author":"Wang","year":"2022"},{"key":"10.1016\/j.neunet.2026.108733_bib0055","series-title":"Proceedings of the 23rd ACM SIGPLAN symposium on principles and practice of parallel programming, ppopp 2018, vienna, austria, february 24\u201328, 2018","first-page":"41","article-title":"Superneurons: Dynamic GPU memory management for training deep neural networks","author":"Wang","year":"2018"},{"key":"10.1016\/j.neunet.2026.108733_bib0056","series-title":"CVPR Workshops","first-page":"864","article-title":"A fully progressive approach to single-image super-resolution","author":"Wang","year":"2018"},{"key":"10.1016\/j.neunet.2026.108733_bib0057","series-title":"IEEE\/CVF Conference on computer vision and pattern recognition, CVPR 2022, new orleans, la, usa, june 18\u201324, 2022","first-page":"1204","article-title":"Scaling vision transformers","author":"Zhai","year":"2022"},{"issue":"3","key":"10.1016\/j.neunet.2026.108733_bib0058","doi-asserted-by":"crossref","first-page":"329","DOI":"10.1016\/j.orl.2022.04.001","article-title":"Extragradient and extrapolation methods with generalized bregman distances for saddle point problems","volume":"50","author":"Zhang","year":"2022","journal-title":"Operations Research Letters"},{"key":"10.1016\/j.neunet.2026.108733_bib0059","series-title":"2019 IEEE\/CVF International conference on computer vision, ICCV 2019, seoul, korea (south), october 27, - november 2, 2019","first-page":"3712","article-title":"Be your own teacher: Improve the performance of convolutional neural networks via self distillation","author":"Zhang","year":"2019"},{"key":"10.1016\/j.neunet.2026.108733_bib0060","series-title":"Advances in neural information processing systems 38: Annual conference on neural information processing systems 2024, neurIPS 2024, vancouver, BC, canada, december 10, - 15, 2024","article-title":"Why transformers need adam: A hessian perspective","author":"Zhang","year":"2024"},{"key":"10.1016\/j.neunet.2026.108733_bib0061","article-title":"On the convergence of adaptive gradient methods for nonconvex optimization","volume":"2024","author":"Zhou","year":"2024","journal-title":"Transactions on Machine Learning Research"},{"key":"10.1016\/j.neunet.2026.108733_bib0062","series-title":"Advances in neural information processing systems 34: Annual conference on neural information processing systems 2021, neurIPS 2021, december 6\u201314, 2021, virtual","first-page":"16410","article-title":"Gradinit: Learning to initialize neural networks for stable and efficient training","author":"Zhu","year":"2021"},{"key":"10.1016\/j.neunet.2026.108733_bib0063","doi-asserted-by":"crossref","unstructured":"Zhuang, B., Liu, J., Pan, Z., He, H., Weng, Y., & Shen, C. (2023). A survey on efficient training of transformers. arXiv: 2302.01107.","DOI":"10.24963\/ijcai.2023\/764"}],"container-title":["Neural Networks"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0893608026001954?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0893608026001954?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T06:38:10Z","timestamp":1776148690000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0893608026001954"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,7]]},"references-count":63,"alternative-id":["S0893608026001954"],"URL":"https:\/\/doi.org\/10.1016\/j.neunet.2026.108733","relation":{},"ISSN":["0893-6080"],"issn-type":[{"value":"0893-6080","type":"print"}],"subject":[],"published":{"date-parts":[[2026,7]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Enhancing progressive ensemble learning via normalized extra-Gradient initialization","name":"articletitle","label":"Article Title"},{"value":"Neural Networks","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.neunet.2026.108733","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"108733"}}