{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T12:08:32Z","timestamp":1772626112487,"version":"3.50.1"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2024,8,21]],"date-time":"2024-08-21T00:00:00Z","timestamp":1724198400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2024,8,21]],"date-time":"2024-08-21T00:00:00Z","timestamp":1724198400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100015752","name":"Northeast Electric Power University","doi-asserted-by":"publisher","award":["BSZT07202107"],"award-info":[{"award-number":["BSZT07202107"]}],"id":[{"id":"10.13039\/501100015752","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s40747-024-01587-w","type":"journal-article","created":{"date-parts":[[2024,8,22]],"date-time":"2024-08-22T15:56:11Z","timestamp":1724342171000},"page":"8319-8339","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["A DQN based approach for large-scale EVs charging scheduling"],"prefix":"10.1007","volume":"10","author":[{"given":"Yingnan","family":"Han","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0001-5545-8700","authenticated-orcid":false,"given":"Tianyang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Qingzhu","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,21]]},"reference":[{"issue":"9","key":"1587_CR1","doi-asserted-by":"publisher","first-page":"9150","DOI":"10.1109\/TITS.2023.3274608","volume":"24","author":"S Hussain","year":"2023","unstructured":"Hussain S et al (2023) Enhancing the efficiency of electric vehicles charging stations based on novel fuzzy integer linear programming. IEEE Trans Intell Transp Syst 24(9):9150\u20139164. https:\/\/doi.org\/10.1109\/TITS.2023.3274608","journal-title":"IEEE Trans Intell Transp Syst"},{"issue":"5","key":"1587_CR2","doi-asserted-by":"publisher","first-page":"4029","DOI":"10.1109\/TSG.2021.3078445","volume":"12","author":"T Long","year":"2021","unstructured":"Long T, Jia QS, Wang G, Yang Y (2021) Efficient real-time EV charging scheduling via ordinal optimization. IEEE Trans Smart Grid 12(5):4029\u20134038. https:\/\/doi.org\/10.1109\/TSG.2021.3078445","journal-title":"IEEE Trans Smart Grid"},{"key":"1587_CR3","unstructured":"V. Global (2021) \"The global electric vehicle market overview In 2021. Statistics & Forecasts 1:2022"},{"key":"1587_CR4","doi-asserted-by":"publisher","unstructured":"Li T, Li X, He T, Zhang Y (2022) \"An EDA-based Genetic Algorithm for EV Charging Scheduling under Surge Demand,\" In: 2022 IEEE International Conference on Services Computing (SCC), 10\u201316 July 2022, pp. 231\u2013238, https:\/\/doi.org\/10.1109\/SCC55611.2022.00042","DOI":"10.1109\/SCC55611.2022.00042"},{"key":"1587_CR5","doi-asserted-by":"publisher","first-page":"106673","DOI":"10.1016\/j.compeleceng.2020.106673","volume":"85","author":"MM Rahman","year":"2020","unstructured":"Rahman MM, Al-Ammar EA, Das HS, Ko WS (2020) Comprehensive impact analysis of electric vehicle charging scheduling on load-duration curve. Comput Electr Eng 85:106673. https:\/\/doi.org\/10.1016\/j.compeleceng.2020.106673","journal-title":"Comput Electr Eng"},{"key":"1587_CR6","doi-asserted-by":"publisher","first-page":"41506","DOI":"10.1109\/ACCESS.2021.3064354","volume":"9","author":"HM Abdullah","year":"2021","unstructured":"Abdullah HM, Gastli A, Ben-Brahim L (2021) Reinforcement learning based EV charging management systems-a review. IEEE Access, Rev 9:41506\u201341531. https:\/\/doi.org\/10.1109\/ACCESS.2021.3064354","journal-title":"IEEE Access, Rev"},{"key":"1587_CR7","doi-asserted-by":"publisher","first-page":"127806","DOI":"10.1016\/j.energy.2023.127806","volume":"278","author":"J Wu","year":"2023","unstructured":"Wu J, Su H, Meng JH, Lin MQ (2023) Electric vehicle charging scheduling considering infrastructure constraints. Energy 278:127806. https:\/\/doi.org\/10.1016\/j.energy.2023.127806","journal-title":"Energy"},{"issue":"5","key":"1587_CR8","doi-asserted-by":"publisher","first-page":"5702","DOI":"10.1109\/tvt.2022.3231901","volume":"72","author":"S Mishra","year":"2023","unstructured":"Mishra S, Mondal A, Mondal S (2023) A multi-objective optimization framework for electric vehicle charge scheduling with adaptable charging ports. IEEE Trans Veh Technol 72(5):5702\u20135714. https:\/\/doi.org\/10.1109\/tvt.2022.3231901","journal-title":"IEEE Trans Veh Technol"},{"issue":"6","key":"1587_CR9","doi-asserted-by":"publisher","first-page":"2925","DOI":"10.3390\/s23062925","volume":"23","author":"A Amin","year":"2023","unstructured":"Amin A, Mahmood A, Khan AR, Arshad K, Assaleh K, Zoha A (2023) A two-stage multi-agent EV charging coordination scheme for maximizing grid performance and customer satisfaction. Sensors 23(6):2925. https:\/\/doi.org\/10.3390\/s23062925","journal-title":"Sensors"},{"issue":"9","key":"1587_CR10","doi-asserted-by":"publisher","first-page":"3023","DOI":"10.3390\/en15093023","volume":"15","author":"D Falabretti","year":"2022","unstructured":"Falabretti D, Gulotta F (2022) A nature-inspired algorithm to enable the E-mobility participation in the ancillary service market. Energies 15(9):3023. https:\/\/doi.org\/10.3390\/en15093023","journal-title":"Energies"},{"key":"1587_CR11","doi-asserted-by":"publisher","first-page":"113742","DOI":"10.1016\/j.applthermaleng.2019.113742","volume":"157","author":"W Cai","year":"2019","unstructured":"Cai W, Vosoogh M, Reinders B, Toshin DS, Ebadi AG (2019) Application of quantum artificial bee colony for energy management by considering the heat and cooling storages. Appl Thermal Eng 157:113742. https:\/\/doi.org\/10.1016\/j.applthermaleng.2019.113742","journal-title":"Appl Thermal Eng"},{"key":"1587_CR12","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2023.106375","volume":"123","author":"SE Comert","year":"2023","unstructured":"Comert SE, Yazgan HR (2023) A new approach based on hybrid ant colony optimization-artificial bee colony algorithm for multi-objective electric vehicle routing problems. Eng Appl Artif Intell 123:106375","journal-title":"Eng Appl Artif Intell"},{"issue":"13","key":"1587_CR13","doi-asserted-by":"publisher","first-page":"5070","DOI":"10.3390\/en16135070","volume":"16","author":"Q Yang","year":"2023","unstructured":"Yang Q, Huang Y, Zhang Q, Zhang J (2023) A bi-level optimization and scheduling strategy for charging stations considering battery degradation. Energies 16(13):5070. https:\/\/doi.org\/10.3390\/en16135070","journal-title":"Energies"},{"key":"1587_CR14","doi-asserted-by":"publisher","DOI":"10.1080\/03772063.2023.2187891","author":"P Das","year":"2023","unstructured":"Das P, Samantaray S, Kayal P (2023) Evaluation of distinct EV scheduling at residential charging points in an unbalanced power distribution system. IETE J Res. https:\/\/doi.org\/10.1080\/03772063.2023.2187891","journal-title":"IETE J Res"},{"key":"1587_CR15","doi-asserted-by":"publisher","DOI":"10.1080\/03772063.2023.2177756","author":"B Sukumar","year":"2023","unstructured":"Sukumar B, Aslam S, Karthikeyan N, Rajesh P (2023) A hybrid BCMPO technique for optimal scheduling of electric vehicle aggregators under market price uncertainty. IETE J Res. https:\/\/doi.org\/10.1080\/03772063.2023.2177756","journal-title":"IETE J Res"},{"key":"1587_CR16","doi-asserted-by":"publisher","first-page":"109348","DOI":"10.1016\/j.ijepes.2023.109348","volume":"153","author":"L Fu","year":"2023","unstructured":"Fu L, Wang T, Song M, Zhou Y, Gao S (2023) Electric vehicle charging scheduling control strategy for the large-scale scenario with non-cooperative game-based multi-agent reinforcement learning. Int J Electr Power Energy Syst 153:109348","journal-title":"Int J Electr Power Energy Syst"},{"key":"1587_CR17","doi-asserted-by":"publisher","first-page":"101162","DOI":"10.1016\/j.segan.2023.101162","volume":"36","author":"A Poddubnyy","year":"2023","unstructured":"Poddubnyy A, Nguyen P, Slootweg H (2023) \"Online EV charging controlled by reinforcement learning with experience replay. Sustain Energy Grids Netw 36:101162","journal-title":"Sustain Energy Grids Netw"},{"key":"1587_CR18","doi-asserted-by":"publisher","first-page":"101124","DOI":"10.1016\/j.segan.2023.101124","volume":"35","author":"S Sykiotis","year":"2023","unstructured":"Sykiotis S, Menos-Aikateriniadis C, Doulamis A, Doulamis N, Georgilakis PS (2023) A self-sustained EV charging framework with N-step deep reinforcement learning. Sustain Energy, Grids Netw 35:101124","journal-title":"Sustain Energy, Grids Netw"},{"key":"1587_CR19","doi-asserted-by":"publisher","first-page":"1135","DOI":"10.1016\/j.egyr.2023.07.054","volume":"10","author":"S Lee","year":"2023","unstructured":"Lee S, Choi D-H (2023) Two-stage scheduling of smart electric vehicle charging stations and inverter-based Volt-VAR control using a prediction error-integrated deep reinforcement learning method. Energy Rep 10:1135\u20131150","journal-title":"Energy Rep"},{"key":"1587_CR20","doi-asserted-by":"publisher","first-page":"106987","DOI":"10.1016\/j.est.2023.106987","volume":"64","author":"SJ Sultanuddin","year":"2023","unstructured":"Sultanuddin SJ, Vibin R, Rajesh Kumar A, Behera NR, Pasha MJ, Baseer KK (2023) Development of improved reinforcement learning smart charging strategy for electric vehicle fleet. J Energy Storage 64:106987","journal-title":"J Energy Storage"},{"key":"1587_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.apenergy.2022.120111","volume":"328","author":"K Park","year":"2022","unstructured":"Park K, Moon I (2022) Multi-agent deep reinforcement learning approach for EV charging scheduling in a smart grid. Appl Energy 328:120111","journal-title":"Appl Energy"},{"issue":"2","key":"1587_CR22","doi-asserted-by":"publisher","first-page":"849","DOI":"10.1109\/TII.2019.2950809","volume":"17","author":"S Wang","year":"2021","unstructured":"Wang S, Bi S, Zhang YA (2021) Reinforcement learning for real-time pricing and scheduling control in EV charging stations. IEEE Trans Industr Inf 17(2):849\u2013859. https:\/\/doi.org\/10.1109\/TII.2019.2950809","journal-title":"IEEE Trans Industr Inf"},{"key":"1587_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2022.105557","volume":"118","author":"M Tan","year":"2023","unstructured":"Tan M, Dai Z, Su Y, Chen C, Wang L, Chen J (2023) Bi-level optimization of charging scheduling of a battery swap station based on deep reinforcement learning. Eng Appl Artif Intell 118:105557","journal-title":"Eng Appl Artif Intell"},{"key":"1587_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijepes.2022.108603","volume":"144","author":"H Li","year":"2023","unstructured":"Li H et al (2023) Constrained large-scale real-time EV scheduling based on recurrent deep reinforcement learning. Int J Electr Power Energy Syst 144:108603","journal-title":"Int J Electr Power Energy Syst"},{"key":"1587_CR25","doi-asserted-by":"publisher","first-page":"108180","DOI":"10.1016\/j.cie.2022.108180","volume":"169","author":"M Alqahtani","year":"2022","unstructured":"Alqahtani M, Scott MJ, Hu M (2022) Dynamic energy scheduling and routing of a large fleet of electric vehicles using multi-agent reinforcement learning. Comput Ind Eng 169:108180","journal-title":"Comput Ind Eng"},{"key":"1587_CR26","unstructured":"Wang Z, Schaul T, Hessel M, Hasselt H, Lanctot M, Freitas N (2016) \"Dueling network architectures for deep reinforcement learning,\" presented at the Proceedings of The 33rd International Conference on Machine Learning, Proceedings of Machine Learning Research"},{"key":"1587_CR27","unstructured":"Fortunato M et al. (2017) \"Noisy networks for exploration,\" arXiv preprint arXiv:1706.10295"},{"issue":"1","key":"1587_CR28","doi-asserted-by":"publisher","first-page":"146","DOI":"10.3390\/en16010146","volume":"16","author":"Q Chen","year":"2023","unstructured":"Chen Q, Folly KA (2023) Application of artificial intelligence for EV charging and discharging scheduling and dynamic pricing: a review. Energies, Rev 16(1):146. https:\/\/doi.org\/10.3390\/en16010146","journal-title":"Energies, Rev"},{"issue":"1","key":"1587_CR29","doi-asserted-by":"publisher","first-page":"7","DOI":"10.3390\/en16010007","volume":"16","author":"PP Singh","year":"2023","unstructured":"Singh PP, Wen F, Palu I, Sachan S, Deb S (2023) electric vehicles charging infrastructure demand and deployment: challenges and solutions. Energ, Rev 16(1):7. https:\/\/doi.org\/10.3390\/en16010007","journal-title":"Energ, Rev"},{"key":"1587_CR30","doi-asserted-by":"publisher","DOI":"10.1016\/j.scs.2021.103207","volume":"76","author":"M Ren","year":"2022","unstructured":"Ren M, Liu X, Yang Z, Zhang J, Guo Y, Jia Y (2022) A novel forecasting based scheduling method for household energy management system based on deep reinforcement learning. Sustain Cities Soc 76:103207","journal-title":"Sustain Cities Soc"},{"key":"1587_CR31","doi-asserted-by":"publisher","first-page":"118127","DOI":"10.1016\/j.apenergy.2021.118127","volume":"307","author":"B Svetozarevic","year":"2022","unstructured":"Svetozarevic B, Baumann C, Muntwiler S, Di Natale L, Zeilinger MN, Heer P (2022) Data-driven control of room temperature and bidirectional EV charging using deep reinforcement learning: simulations and experiments. Appl Energy 307:118127","journal-title":"Appl Energy"},{"key":"1587_CR32","doi-asserted-by":"publisher","DOI":"10.1016\/j.apenergy.2022.120140","volume":"328","author":"R Jin","year":"2022","unstructured":"Jin R, Zhou Y, Lu C, Song J (2022) Deep reinforcement learning-based strategy for charging station participating in demand response. Appl Energy 328:120140","journal-title":"Appl Energy"},{"key":"1587_CR33","doi-asserted-by":"publisher","first-page":"100280","DOI":"10.1016\/j.etran.2023.100280","volume":"18","author":"A Hussain","year":"2023","unstructured":"Hussain A, Bui V-H, Musilek P (2023) Local demand management of charging stations using vehicle-to-vehicle service: a welfare maximization-based soft actor-critic model. Etransportation 18:100280","journal-title":"Etransportation"},{"issue":"5","key":"1587_CR34","doi-asserted-by":"publisher","first-page":"5901","DOI":"10.1109\/TIA.2020.2984614","volume":"56","author":"D Qiu","year":"2020","unstructured":"Qiu D, Ye Y, Papadaskalopoulos D, Strbac G (2020) A deep reinforcement learning method for pricing electric vehicles with discrete charging levels. IEEE Trans Ind Appl 56(5):5901\u20135912","journal-title":"IEEE Trans Ind Appl"},{"key":"1587_CR35","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2022.105642","volume":"118","author":"Y Zhang","year":"2023","unstructured":"Zhang Y, Rao X, Liu C, Zhang X, Zhou Y (2023) A cooperative EV charging scheduling strategy based on double deep Q-network and Prioritized experience replay. Eng Appl Artif Intell 118:105642","journal-title":"Eng Appl Artif Intell"},{"issue":"2","key":"1587_CR36","doi-asserted-by":"publisher","first-page":"1714","DOI":"10.1109\/TSG.2019.2942593","volume":"11","author":"T Qian","year":"2019","unstructured":"Qian T, Shao C, Wang X, Shahidehpour M (2019) Deep reinforcement learning for EV charging navigation by coordinating smart grid and intelligent transportation system. IEEE Trans Smart Grid 11(2):1714\u20131723","journal-title":"IEEE Trans Smart Grid"},{"key":"1587_CR37","doi-asserted-by":"crossref","unstructured":"Ur Rehman U, Riaz M (2018) \"Real time controlling algorithm for vehicle to grid system under price uncertainties,\" In: 2018 1st International Conference on Power, Energy and Smart Grid (ICPESG), IEEE, pp. 1\u20137","DOI":"10.1109\/ICPESG.2018.8384522"},{"key":"1587_CR38","doi-asserted-by":"publisher","first-page":"41506","DOI":"10.1109\/ACCESS.2021.3064354","volume":"9","author":"HM Abdullah","year":"2021","unstructured":"Abdullah HM, Gastli A, Ben-Brahim L (2021) Reinforcement learning based EV charging management systems\u2013a review. IEEE Access 9:41506\u201341531","journal-title":"IEEE Access"},{"issue":"01","key":"1587_CR39","first-page":"922","volume":"34","author":"J Sun","year":"2020","unstructured":"Sun J, Zheng Y, Hao J, Meng Z, Liu Y (2020) Continuous multiagent control using collective behavior entropy for large-scale home energy management. Proceed AAAI Conf Artif Intell 34(01):922\u2013929","journal-title":"Proceed AAAI Conf Artif Intell"},{"issue":"5","key":"1587_CR40","doi-asserted-by":"publisher","first-page":"3075","DOI":"10.1109\/JIOT.2020.3015204","volume":"8","author":"F Zhang","year":"2021","unstructured":"Zhang F, Yang Q, An D (2021) CDDPG: a deep-reinforcement-learning-based approach for electric vehicle charging control. IEEE Internet Things J 8(5):3075\u20133087. https:\/\/doi.org\/10.1109\/JIOT.2020.3015204","journal-title":"IEEE Internet Things J"},{"issue":"5","key":"1587_CR41","doi-asserted-by":"publisher","first-page":"3493","DOI":"10.1109\/TII.2019.2944183","volume":"16","author":"M Shin","year":"2019","unstructured":"Shin M, Choi D-H, Kim J (2019) Cooperative management for PV\/ESS-enabled electric vehicle charging stations: a multiagent deep reinforcement learning approach. IEEE Trans Industr Inf 16(5):3493\u20133503","journal-title":"IEEE Trans Industr Inf"},{"issue":"1","key":"1587_CR42","doi-asserted-by":"publisher","first-page":"217","DOI":"10.1109\/JSAC.2019.2951966","volume":"38","author":"J Liu","year":"2019","unstructured":"Liu J, Guo H, Xiong J, Kato N, Zhang J, Zhang Y (2019) Smart and resilient EV charging in SDN-enhanced vehicular edge computing networks. IEEE J Sel Areas Commun 38(1):217\u2013228","journal-title":"IEEE J Sel Areas Commun"},{"issue":"5","key":"1587_CR43","doi-asserted-by":"publisher","first-page":"2312","DOI":"10.1109\/TSG.2015.2396993","volume":"6","author":"Z Wen","year":"2015","unstructured":"Wen Z, O\u2019Neill D, Maei H (2015) Optimal demand response using device-based reinforcement learning. IEEE Transactions on Smart Grid 6(5):2312\u20132324","journal-title":"IEEE Transactions on Smart Grid"},{"issue":"3","key":"1587_CR44","doi-asserted-by":"publisher","first-page":"2427","DOI":"10.1109\/TSG.2019.2955437","volume":"11","author":"H Li","year":"2019","unstructured":"Li H, Wan Z, He H (2019) Constrained EV charging scheduling based on safe deep reinforcement learning. IEEE Trans Smart Grid 11(3):2427\u20132439","journal-title":"IEEE Trans Smart Grid"},{"key":"1587_CR45","doi-asserted-by":"crossref","unstructured":"Dusparic I, Harris C, Marinescu A, Cahill V, Clarke S (2013) \"Multi-agent residential demand response based on load forecasting,\" In: 2013 1st IEEE conference on technologies for sustainability (SusTech), IEEE, pp. 90\u201396","DOI":"10.1109\/SusTech.2013.6617303"},{"key":"1587_CR46","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-024-09530-3","author":"A Zhang","year":"2024","unstructured":"Zhang A, Liu Q, Liu J, Cheng L (2024) CASA: cost-effective EV charging scheduling based on deep reinforcement learning. Neural Comput Appl. https:\/\/doi.org\/10.1007\/s00521-024-09530-3","journal-title":"Neural Comput Appl"},{"issue":"3","key":"1587_CR47","doi-asserted-by":"publisher","first-page":"1618","DOI":"10.3390\/s23031618","volume":"23","author":"D Liu","year":"2023","unstructured":"Liu D, Zeng P, Cui S, Song C (2023) Deep reinforcement learning for charging scheduling of electric vehicles considering distribution network voltage stability. Sensors 23(3):1618","journal-title":"Sensors"},{"issue":"22","key":"1587_CR48","doi-asserted-by":"publisher","first-page":"22571","DOI":"10.1109\/JIOT.2022.3181613","volume":"9","author":"J Jin","year":"2022","unstructured":"Jin J, Xu Y (2022) Shortest-path-based deep reinforcement learning for EV charging routing under stochastic traffic condition and electricity prices. IEEE Internet Things J 9(22):22571\u201322581. https:\/\/doi.org\/10.1109\/JIOT.2022.3181613","journal-title":"IEEE Internet Things J"},{"key":"1587_CR49","doi-asserted-by":"publisher","DOI":"10.1016\/j.ress.2022.108920","volume":"230","author":"S Wang","year":"2023","unstructured":"Wang S, Fan Y, Jin S, Takyi-Aninakwa P, Fernandez C (2023) Improved anti-noise adaptive long short-term memory neural network modeling for the robust remaining useful life prediction of lithium-ion batteries. Reliab Eng Syst Saf 230:108920","journal-title":"Reliab Eng Syst Saf"},{"key":"1587_CR50","doi-asserted-by":"publisher","DOI":"10.1016\/j.energy.2023.128677","volume":"284","author":"S Wang","year":"2023","unstructured":"Wang S, Wu F, Takyi-Aninakwa P, Fernandez C, Stroe D-I, Huang Q (2023) Improved singular filtering-Gaussian process regression-long short-term memory model for whole-life-cycle remaining capacity estimation of lithium-ion batteries adaptive to fast aging and multi-current variations. Energy 284:128677","journal-title":"Energy"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01587-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-024-01587-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01587-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,16]],"date-time":"2024-10-16T22:21:04Z","timestamp":1729117264000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-024-01587-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,21]]},"references-count":50,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["1587"],"URL":"https:\/\/doi.org\/10.1007\/s40747-024-01587-w","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,8,21]]},"assertion":[{"value":"16 December 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 July 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 August 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"There are no financial or non-financial interests that are directly or indirectly related to the work submitted for publication.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}